Chinchilla Description
Chinchilla has a large language. Chinchilla has the same compute budget of Gopher, but 70B more parameters and 4x as much data. Chinchilla consistently and significantly outperforms Gopher 280B, GPT-3 175B, Jurassic-1 178B, and Megatron-Turing (530B) in a wide range of downstream evaluation tasks. Chinchilla also uses less compute to perform fine-tuning, inference and other tasks. This makes it easier for downstream users to use. Chinchilla reaches a high-level average accuracy of 67.5% for the MMLU benchmark. This is a greater than 7% improvement compared to Gopher.
Integrations
Company Details
Company:
Google DeepMind
Headquarters:
United States
Website:
arxiv.org/abs/2203.15556
Media
Recommended Products
Product Details
Platforms
SaaS
Windows
Mac
Linux
On-Premises
Type of Training
Documentation