Just how fast is 857 tokens/sec?
"Groq is on a mission to set the standard for GenAI inference speed, helping real-time AI applications come to life today. An LPU Inference Engine, with LPU standing for Language Processing Unit?, is a new type of end-to-end processing unit system that provides the fastest inference for computationally intensive applications with a sequential component to them, such as AI language applications (LLMs)."