- Meta's latest Llama 3.3 70B is a text-only model and it's optimized for cost efficiency.
- Llama 3.3 70B model comes very close to beating the much larger, Llama 3.1 405B model.
- Llama 3.3 70B scores 86.0 on MMLU and 88.4 on HumanEval.
In September 2024, Meta released a family of Llama 3.2 models, and now the company has launched a new Llama 3.3 70B model that is optimized for efficiency. While the Llama 3.2 90B is a multimodal model with vision capability, the new Llama 3.3 70B model is a text-only model. But what makes it stand out?
Well, according to Meta, the new Llama 3.3 70B model nearly matches the performance of the larger Llama 3.1 405B model. That’s a huge improvement since its size is much smaller and can be served at a much lower cost. But it doesn’t outrightly beat the larger 405B model in all benchmarks.
The Llama 3.3 70B model scores 86.0 and 88.4 in MMLU and HumanEval benchmarks, respectively. The 405B model does slightly better and achieves 88.6 and 89.0 in the same set of tests. That said, the Llama 3.3 70B model scores better in MATH and GPQA Diamond.
Basically, Meta is saying that if you have text-only applications, you should use the new Llama 3.3 70B model rather than the 405B model. Due to its smaller size, it costs just $0.1 / $0.4 for 1 million input/output tokens. The larger 405B model costs $1.0 / $1.8 for 1 million input/input tokens.
As for language support, the Llama 3.3 70B model supports English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai. Its knowledge cutoff date is December 2023 and the context length is up to 128K tokens. You can chat with the new Llama 3.3 70B model on HuggingChat for free.