LongLLaMA Description

This repository showcases the research preview of LongLLaMA, an advanced large language model that can manage extensive contexts of up to 256,000 tokens or potentially more. LongLLaMA is developed on the OpenLLaMA framework and has been fine-tuned utilizing the Focused Transformer (FoT) technique. The underlying code for LongLLaMA is derived from Code Llama. We are releasing a smaller 3B base variant of the LongLLaMA model, which is not instruction-tuned, under an open license (Apache 2.0), along with inference code that accommodates longer contexts available on Hugging Face. This model's weights can seamlessly replace LLaMA in existing systems designed for shorter contexts, specifically those handling up to 2048 tokens. Furthermore, we include evaluation results along with comparisons to the original OpenLLaMA models, thereby providing a comprehensive overview of LongLLaMA's capabilities in the realm of long-context processing.

Pricing

Pricing Starts At:
Free
Free Version:
Yes

Integrations

No Integrations at this time

Reviews

Total
ease
features
design
support

No User Reviews. Be the first to provide a review:

Write a Review

Company Details

Company:
LongLLaMA
Website:
github.com/CStanKonrad/long_llama

Media

LongLLaMA Screenshot 1
Recommended Products
Auth0 for AI Agents now in GA Icon
Auth0 for AI Agents now in GA

Ready to implement AI with confidence (without sacrificing security)?

Connect your AI agents to apps and data more securely, give users control over the actions AI agents can perform and the data they can access, and enable human confirmation for critical agent actions.
Start building today

Product Details

Platforms
Web-Based
On-Premises
Types of Training
Training Docs
Customer Support
Online Support

LongLLaMA Features and Options

LongLLaMA Lists

LongLLaMA User Reviews

Write a Review
  • Previous
  • Next