Product Demo - Groq: Groq LPU™ Offers Best-in-class Inference Performance on Large Language Models
September 20, 2023
Booth 316, Hall 1


- Groq is leading in ultra-low latency with >240 tokens per second per user inference performance for Llama-2 70B
- The Language Processing Unit™ system provides the fastest AI solution for LLM inference deployment
- Groq solutions combine easy-to use software, programmability, and scalable hardware
- Groq enables new language-based AI applications with record-breaking speed, revolutionizing use cases with real-time analysis and response