Notice: Function _load_textdomain_just_in_time was called incorrectly. Translation loading for the thegem domain was triggered too early. This is usually an indicator for some code in the plugin or theme running too early. Translations should be loaded at the init action or later. Please see Debugging in WordPress for more information. (This message was added in version 6.7.0.) in /home/j8p72agj2cgw/fanaticalfuturist.com/wp-includes/functions.php on line 6121

Notice: Function _load_textdomain_just_in_time was called incorrectly. Translation loading for the wp-2fa domain was triggered too early. This is usually an indicator for some code in the plugin or theme running too early. Translations should be loaded at the init action or later. Please see Debugging in WordPress for more information. (This message was added in version 6.7.0.) in /home/j8p72agj2cgw/fanaticalfuturist.com/wp-includes/functions.php on line 6121
Groq's ultrafast LPU accelerator smashes AI LLM benchmarks – Matthew Griffin | Keynote Speaker & Master Futurist
Scroll Top

Groq’s ultrafast LPU accelerator smashes AI LLM benchmarks

WHY THIS MATTERS

When it comes to AI in some cases fast is best and a new rival GPU manufacturer has just one upped Nvidia.

 

Love the Exponential Future? Join our XPotential Community, future proof yourself with courses from XPotential University, read about exponential tech and trendsconnect, watch a keynote, or browse my blog.

Groq, led by ex-Google engineer and CEO Jonathan Ross, claims to have created the first ever Language Processing Unit (LPU) which it says can deliver the fastest speeds for AI applications. It’s a bold claim, but one that the latest demos more than back up, suggesting it could well become an absolute game-changer for AI.

 

RELATED
Researchers turn plants into biological computers to grow better crops

 

Ross, who previously designed Google’s Tensor Processing Unit (TPU), launched Groq in 2016 to create a chip capable of executing deep learning inference tasks more efficiently than existing CPUs and GPUs.

The company’s Tensor Stream Processor (TSP) is likened to an assembly line, processing data tasks in a sequential, organized manner. In contrast, a GPU is akin to a static workstation, where workers come and go to apply processing steps. The TSP’s efficiency became evident with the rise of Generative AI, leading Ross to rebrand the TSP as the Language Processing Unit (LPU) to increase its recognizability.

 

See it in action.

 

Unlike GPUs, LPUs utilize a streamlined approach, eliminating the need for complex scheduling hardware, ensuring consistent latency and throughput. LPUs are also energy efficient, reducing the overhead of managing multiple threads and avoiding underutilization of cores. Groq’s scalable chip design allows multiple TSPs to be linked without traditional bottlenecks, simplifying hardware requirements for large-scale AI models.

 

RELATED
For the first time a flying yacht has made green hydrogen at sea

 

The first public demo of Groq was a lightning-fast AI answers engine that generated answers with hundreds of words in less that a second. Matt Shumer posted the test on X and says more than 3/4 of the time was spent searching not generating.

If you want to try Groq for yourself, to get an idea of just how fast it can be for AI, go to this chat page. Use the drop down on the left to switch between the different available models.

Related Posts

Leave a comment

Pin It on Pinterest

Share This