15.1 C
New York
Saturday, November 2, 2024

Groq’s breakthrough AI chip achieves blistering 800 tokens per second on Meta’s LLaMA 3

In a surprising benchmark result that could shake up the competitive landscape for AI inference, startup chip company Groq appears to have confirmed through a series of retweets that its system is serving Meta’s newly released LLaMA 3 large language model at over 800 tokens per second. “We’ve been …

This post was originally published on this site

Subscribe
Notify of
0 Comments
Inline Feedbacks
View all comments

Stay Connected

156,534FansLike
396,312FollowersFollow
2,320SubscribersSubscribe

Latest Articles

0
Would love your thoughts, please comment.x
()
x