Average Ratings 0 Ratings

Total
ease
features
design
support

No User Reviews. Be the first to provide a review:

Write a Review

Average Ratings 0 Ratings

Total
ease
features
design
support

No User Reviews. Be the first to provide a review:

Write a Review

Description

Liquid AI's LFM2.5 represents an advanced iteration of on-device AI foundation models, engineered to provide high-efficiency and performance for AI inference on edge devices like smartphones, laptops, vehicles, IoT systems, and embedded hardware without the need for cloud computing resources. This new version builds upon the earlier LFM2 framework by greatly enhancing the scale of pretraining and the stages of reinforcement learning, resulting in a suite of hybrid models that boast around 1.2 billion parameters while effectively balancing instruction adherence, reasoning skills, and multimodal functionalities for practical applications. The LFM2.5 series comprises various models including Base (for fine-tuning and personalization), Instruct (designed for general-purpose instruction), Japanese-optimized, Vision-Language, and Audio-Language variants, all meticulously crafted for rapid on-device inference even with stringent memory limitations. These models are also made available as open-weight options, facilitating deployment through platforms such as llama.cpp, MLX, vLLM, and ONNX, thus ensuring versatility for developers. With these enhancements, LFM2.5 positions itself as a robust solution for diverse AI-driven tasks in real-world environments.

Description

The TinyLlama initiative seeks to pretrain a Llama model with 1.1 billion parameters using a dataset of 3 trillion tokens. With the right optimizations, this ambitious task can be completed in a mere 90 days, utilizing 16 A100-40G GPUs. We have maintained the same architecture and tokenizer as Llama 2, ensuring that TinyLlama is compatible with various open-source projects that are based on Llama. Additionally, the model's compact design, consisting of just 1.1 billion parameters, makes it suitable for numerous applications that require limited computational resources and memory. This versatility enables developers to integrate TinyLlama seamlessly into their existing frameworks and workflows.

API Access

Has API

API Access

Has API

Screenshots View All

Screenshots View All

No images available

Integrations

Amazon Bedrock
ElevenLabs
Gemma 3
Gemma 4
Hugging Face
LEAP
Llama
Llama 3.2
Qwen3
RunPod

Integrations

Amazon Bedrock
ElevenLabs
Gemma 3
Gemma 4
Hugging Face
LEAP
Llama
Llama 3.2
Qwen3
RunPod

Pricing Details

Free
Free Trial
Free Version

Pricing Details

Free
Free Trial
Free Version

Deployment

Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook

Deployment

Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook

Customer Support

Business Hours
Live Rep (24/7)
Online Support

Customer Support

Business Hours
Live Rep (24/7)
Online Support

Types of Training

Training Docs
Webinars
Live Training (Online)
In Person

Types of Training

Training Docs
Webinars
Live Training (Online)
In Person

Vendor Details

Company Name

Liquid AI

Founded

2023

Country

United States

Website

www.liquid.ai/blog/introducing-lfm2-5-the-next-generation-of-on-device-ai

Vendor Details

Company Name

TinyLlama

Website

github.com/jzhang38/TinyLlama

Product Features

Product Features

Alternatives

Alternatives

Llama 2 Reviews

Llama 2

Meta
MedGemma Reviews

MedGemma

Google DeepMind
HunyuanOCR Reviews

HunyuanOCR

Tencent