Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Small models retain much less of the knowledge they were trained on, especially when quantized.

One good use case for 32gb Mac is being able to run 8b models at full precision, something that is not possible with 8-16gb macs



Or better run quantized 14B or even 32B models...




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: