Streaming a response from a chatbot is only one use-case of LLMs.
I would argue the most interesting applications do not fall into this category.
…not yet anyway. Fast moving area, lots of blue water outside the chat interface.
Groq model shines at latency, not at the other two.
Streaming a response from a chatbot is only one use-case of LLMs.
I would argue the most interesting applications do not fall into this category.