Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

as someone who uses deepseek, glm and kimi models exclusively, an llm telling me what to do is just off the wall

glm and kimi in particular, they can't stop writing... seriously very eager to please. always finishing with fireworks emoji and saying how pleased it is with the test working.

i have to say to write less documentation and simplify their code.



LLMs are next token predictors. Outputting tokens is what they do, and the natural steady-state for them is an infinite loop of endlessly generated tokens.

You need to train them on a special "stop token" to get them to act more human. (Whether explicitly in post-training or with system prompt hacks.)

This isn't a general solution to the problem and likely there will never be one.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: