Llama-3 (70B): Meta's most powerful open-source language model for developers to date.
Llama-3 (70B) is a state-of-the-art open-source language model developed by Meta AI. This 70-billion parameter model is designed to excel in reasoning, coding, and broad application across multiple languages and domains.
Llama-3 (70B) is intended for a wide range of natural language processing tasks, including:
Llama-3 (70B) supports over 30 languages, with a focus on high-quality non-English data, which accounts for more than 5% of the pretraining dataset.
Llama-3 (70B) uses an optimized transformer architecture with grouped-query attention for improved inference scalability. The model's architecture is designed to efficiently process and generate text while maintaining high performance.
Llama-3 (70B) was pretrained on over 15 trillion tokens from publicly available sources, including web pages, books, and other text corpora. The dataset includes a significant amount of code, with 4x more code than Llama-2.
The training data for Llama-3 (70B) comes from a variety of publicly available sources, such as web pages, books, and other text corpora. The total size of the pretraining dataset is over 15 trillion tokens.
The knowledge cutoff for Llama-3 (70B) is December 2023, meaning that the model's knowledge is current up to that date.
Llama-3 (70B) was trained on a diverse dataset that includes content from various sources and perspectives. However, as with any large language model, there may be biases present in the training data that could be reflected in the model's outputs.
Llama-3 (70B) has demonstrated strong performance on various benchmarks and tasks. It performs on par with GPT-4o on certain tasks while being 15x cheaper.
Meta has invested in tools to enhance the safety of Llama-3 (70B) and reduce the risk of harmful outputs. The model's usage is subject to Meta's ethical guidelines and principles.
Llama-3 (70B) is available under a custom commercial license from Meta.