IIRC the codestral fim tokens aren't properly implemented in llama.cpp/ollama, what backend are you using to run them? id probably have to drop down to iq2_xxs or something for the full fat deepseek but I'll definitely look into codestral, I'm a big fan of mixtral, hopefully a MoE code model with FIM comes along soon.
All of those are FIM capable, but especially deepseek-v2-lite is very picky with its prompt template so make sure you use it correctly...
Depending on your hardware codestral-22B might be fast enough for everything, but for me it's a bit to slow...
If you can run it deepseek v2 non-light is amazing, but it requires loads of VRAM