I’m excited to announce the availability of Meta’s Llama 3 — the next generation of Meta’s open large language model — on our watsonx.ai platform.
Starting today, the Llama 3 8B and Llama 3 70B models are generally available on watsonx.ai. They join previous models in the Llama family, Llama 2 7B, Llama 2 70B, and CodeLlama 34B. A self-deployed release of the Llama 3 series is coming soon that will allow you to deploy these models on the platform of your choice (multi-cloud or on-premises, with no vendor lock-in).
💡 Quick facts about Llama 3 models:
- Llama3 8B and 70B are available today on watsonx.ai. We are excited to have these models available the same day they were released!
- Llama 3 8B is ideal for environments with limited computational power or resources (e.g. edge devices)
- Llama 3 70B is suited for content creation, conversational AI, language understanding, and enterprise applications
- Impressive performance: MMLU (5-shot) for instruction-tuned Llama3 70B is 82.0 and for Llama3 8B is 68.4
- Both models are licensed for commercial use
- Stats: 15T+ training tokens, 8K context length, new efficiencies in tokenization that improves performance on English and multilingual benchmarks
- IBM offers competitive pricing on Llama 3 models: Llama 3 8B is $0.60 per 1M tokens, Llama 3 70B is $1.81 per 1M tokens.
![](https://dw1.s81c.com//IMWUC/MessageImages/820995d3aadf441ca06db8286e7c451f.png)
<pricing as of April 18, 2024>
Over the past few months, Llama 2 models have been extensively adopted by IBM customers for summarization, classification, information extraction, content-grounded question answering, and content generation. As an example, the Recording Academy — the non-profit that hosts the GRAMMYs — tuned Llama 2 to produce digital content consistent with their brand’s standards and tone of voice.
Read the full article.
\