Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I get wrong answers for basic things like how to fill out a government form or the relationship between two distant historical figures, things I'm actually working on directly and not some "trick" to get the machine to screw up. They get a lot right a lot of the time, but they're inherently untrustworthy because they sometimes get things subtly or catastrophically wrong and without some kind of consistent confidence scoring, there's no way to tell the difference without further research, and almost necessarily on some other tool because LLMs like to hold onto their lies and it's very difficult to convince them to discard a hallucination.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: