Llama and Groq are going to be huge!
CEO @ Groq, the Most Popular API for Fast Inference | Creator of the TPU and LPU, Two of the World’s Most Important AI Chips | On a Mission to Double the World's AI Inference Compute by 2027
"The reports of the LLM scaling laws' demise have been greatly exaggerated." Today, our partner Meta released its latest version of Llama-3.3-70B-Instruct. And to all those who speculated that the industry had hit the wall - maybe some have, but Meta hasn’t yet. 😉 This is a big deal. Though ~1/5th the size of Llama 3.1-405B, our benchmarking showed Llama-3.3-70B to be performing neck and neck in quality, and in many crucial cases substantially outperforming the larger model (Instruction Following, Coding, Math, etc.), making it a suitable replacement for a majority of workloads. It's also significantly less expensive and faster than the larger model. Meta continues to push the lead in open weight innovations, and is keeping the pressure high for proprietary model providers to attempt to keep ahead of the giant wave of open. The new Llama-3.3-70B model launched and is now available to all 645,000 GroqCloud™ developers as of this morning. Go cook, and don't forget to share what you build here. Thank you for making GroqCloud™ the #1 API for fast inference! And remember, this is only just the beginning. You can read the blog for the details of how to upgrade to the refreshed model (link in comments).