15.3 C
New York
Saturday, May 18, 2024

Groq’s breakthrough AI chip achieves blistering 800 tokens per second on Meta’s LLaMA 3

In a surprising benchmark result that could shake up the competitive landscape for AI inference, startup chip company Groq appears to have confirmed through a series of retweets that its system is serving Meta’s newly released LLaMA 3 large language model at over 800 tokens per second. “We’ve been …

This post was originally published on this site

Subscribe
Notify of
0 Comments
Inline Feedbacks
View all comments

Stay Connected

157,202FansLike
396,312FollowersFollow
2,300SubscribersSubscribe

Latest Articles

0
Would love your thoughts, please comment.x
()
x