Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

SO and Google are still a problem especially they are one of the sources of LLMs.

So now we can get wrong code but written in a more confident language.

If you are lucky it’s a hallucination and the error is obvious.



> So now we can get wrong code but written in a more confident language.

The lesson to learn is rather that also in "real life", you shouldn't trust confident people.


Mean’s don’t trust AI




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: