Reka Flash 3
Reka Flash 3 is an AI model with 21 billion parameters developed by Reka AI. It excels in general chat, programming, instruction following and function calling. It can process and reason with text, audio, video, or image inputs. This makes it a general-purpose, compact solution for a variety of applications. Reka Flash 3, which was trained from scratch using diverse datasets including publically accessible and synthetic data for optimization, underwent instruction tuning with curated, high quality data. The final stage of training involved reinforcement learning with REINFORCE Let One Out (RLOO), which included both model-based as well as rule-based rewards. This enhanced its reasoning abilities. Reka Flash 3, with a context length 32,000 tokens is competitive with proprietary models such as OpenAI's O1-mini and is suitable for low-latency deployments or on-device deployments. Full precision of the model requires 39GB (fp16), however, it can be compressed down to 11GB by using 4-bit quantization.
Learn more
Llama 2
The next generation of the large language model. This release includes modelweights and starting code to pretrained and fine tuned Llama languages models, ranging from 7B-70B parameters.
Llama 1 models have a context length of 2 trillion tokens. Llama 2 models have a context length double that of Llama 1. The fine-tuned Llama 2 models have been trained using over 1,000,000 human annotations.
Llama 2, a new open-source language model, outperforms many other open-source language models in external benchmarks. These include tests of reasoning, coding and proficiency, as well as knowledge tests.
Llama 2 has been pre-trained using publicly available online data sources. Llama-2 chat, a fine-tuned version of the model, is based on publicly available instruction datasets, and more than 1 million human annotations.
We have a wide range of supporters in the world who are committed to our open approach for today's AI. These companies have provided early feedback and have expressed excitement to build with Llama 2
Learn more
Yi-Lightning
Yi-Lightning is the latest large language model developed by 01.AI, under the leadership Kai-Fu Lee. It focuses on high performance, cost-efficiency, and a wide range of languages. It has a maximum context of 16K tokens, and costs $0.14 per million tokens both for input and output. This makes it very competitive. Yi-Lightning uses an enhanced Mixture-of-Experts architecture that incorporates fine-grained expert segments and advanced routing strategies to improve its efficiency. This model has excelled across a variety of domains. It achieved top rankings in categories such as Chinese, math, coding and hard prompts in the chatbot arena where it secured the sixth position overall and ninth in style control. Its development included pre-training, supervised tuning, and reinforcement learning based on human feedback. This ensured both performance and safety with optimizations for memory usage and inference speeds.
Learn more
Baichuan-13B
Baichuan-13B is a large language model with 13 billion parameters, developed by Baichuan Intelligent as an open-source and commercially available product, succeeding the Baichuan-7B model. It has excelled in authoritative benchmarks for both Chinese and English, outperforming other models of comparable size. This release features two distinct pre-training versions: Baichuan-13B-Base and Baichuan-13B-Chat.
With an impressive expansion in size, Baichuan-13B enhances its parameter count to 13 billion, building upon the foundation laid by Baichuan-7B, while training on a remarkable 1.4 trillion tokens derived from high-quality datasets, representing a 40% increase compared to LLaMA-13B. It is recognized as the open-source model in the 13B category with the most extensive training data available. Additionally, it supports bilingual capabilities in Chinese and English, utilizes ALiBi positional encoding, and boasts a context window length of 4096 tokens, making it a versatile choice for various applications in natural language processing.
Learn more