Skip to content
Thoughts on GitHub Models?

Meta-Llama-3-70B-Instruct

A powerful 70-billion parameter model excelling in reasoning, coding, and broad language applications.
Context
8k input · 4k output
Training date
Dec 2023
Rate limit tier
Provider support
Try Meta-Llama-3-70B-Instruct
Get early access to our playground for modelsJoin our limited beta waiting list today and be among the first to try out an easy way to test models

Model navigation navigation

Meta

Meta developed and released the Meta Llama 3 family of large language models (LLMs), a collection of pretrained and instruction tuned generative text models in 8 and 70B sizes. The Llama 3 instruction tuned models are optimized for dialogue use cases and outperform many of the available open source chat models on common industry benchmarks. Further, in developing these models, we took great care to optimize helpfulness and safety.

Model Architecture

Llama 3 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.

Training Datasets

Overview Llama 3 was pretrained on over 15 trillion tokens of data from publicly available sources. The fine-tuning data includes publicly available instruction datasets, as well as over 10M human-annotated examples. Neither the pretraining nor the fine-tuning datasets include Meta user data.

Data Freshness The pretraining data has a cutoff of March 2023 for the 8B and December 2023 for the 70B models respectively.

Languages

 (1)
English

About

A powerful 70-billion parameter model excelling in reasoning, coding, and broad language applications.
Context
8k input · 4k output
Training date
Dec 2023
Rate limit tier
Provider support

Languages

 (1)
English