Chip Acceleration: Groq Chatbot Zooms Past ChatGPT

posted  19 Feb 2024
Photo - Chip Acceleration: Groq Chatbot Zooms Past ChatGPT
The Groq chatbot, harnessing the power of the Mixtral AI model, has impressively outpaced OpenAI's ChatGPT in terms of query processing and response generation speeds.

The chatbot's primary advantage lies in its capability to process 500 tokens per second—eightfold faster than ChatGPT 3.5, which manages around 40 tokens in the same timeframe.

The cornerstone of Groq's success lies in its innovative Tensor Processing Unit, specifically designed for large language models (LLM), aptly named the Language Processing Unit (LPU). This specialized chip is akin to a video card or ASIC miner, but exclusively engineered for AI operations.

Meanwhile, OpenAI is also delving into the realm of developing its own specialized AI processor and is currently seeking the necessary funding.

Interestingly, Groq predates Elon Musk's chatbot Grok by seven years, having been established in 2016. Reflecting on the name resemblance, Groq's team humorously suggested Musk rename his chatbot to 'Slartibartfast', a quirky reference to Douglas Adams' famous "Hitchhiker's Guide to the Galaxy."