Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I tried the 27B QAT model and it hallucinates like crazy. When I ask it for information about some made up person, restaurant, place name, etc., it never says "I don't know about that" and instead seems eager to just make up details. The larger local models like the older Llama 3.3 70B seem better at this, but are also too big to fit on a 24GB GPU.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: