Average Ratings 0 Ratings

Total
ease
features
design
support

No User Reviews. Be the first to provide a review:

Write a Review

Average Ratings 0 Ratings

Total
ease
features
design
support

No User Reviews. Be the first to provide a review:

Write a Review

Description

DeepCoder, an entirely open-source model for code reasoning and generation, has been developed through a partnership between Agentica Project and Together AI. Leveraging the foundation of DeepSeek-R1-Distilled-Qwen-14B, it has undergone fine-tuning via distributed reinforcement learning, achieving a notable accuracy of 60.6% on LiveCodeBench, which marks an 8% enhancement over its predecessor. This level of performance rivals that of proprietary models like o3-mini (2025-01-031 Low) and o1, all while operating with only 14 billion parameters. The training process spanned 2.5 weeks on 32 H100 GPUs, utilizing a carefully curated dataset of approximately 24,000 coding challenges sourced from validated platforms, including TACO-Verified, PrimeIntellect SYNTHETIC-1, and submissions to LiveCodeBench. Each problem mandated a legitimate solution along with a minimum of five unit tests to guarantee reliability during reinforcement learning training. Furthermore, to effectively manage long-range context, DeepCoder incorporates strategies such as iterative context lengthening and overlong filtering, ensuring it remains adept at handling complex coding tasks. This innovative approach allows DeepCoder to maintain high standards of accuracy and reliability in its code generation capabilities.

Description

EXAONE Deep represents a collection of advanced language models that are enhanced for reasoning, created by LG AI Research, and come in sizes of 2.4 billion, 7.8 billion, and 32 billion parameters. These models excel in a variety of reasoning challenges, particularly in areas such as mathematics and coding assessments. Significantly, the EXAONE Deep 2.4B model outshines other models of its size, while the 7.8B variant outperforms both open-weight models of similar dimensions and the proprietary reasoning model known as OpenAI o1-mini. Furthermore, the EXAONE Deep 32B model competes effectively with top-tier open-weight models in the field. The accompanying repository offers extensive documentation that includes performance assessments, quick-start guides for leveraging EXAONE Deep models with the Transformers library, detailed explanations of quantized EXAONE Deep weights formatted in AWQ and GGUF, as well as guidance on how to run these models locally through platforms like llama.cpp and Ollama. Additionally, this resource serves to enhance user understanding and accessibility to the capabilities of EXAONE Deep models.

API Access

Has API

API Access

Has API

Screenshots View All

Screenshots View All

Integrations

Hugging Face
Ollama
OpenAI
OpenAI o1-mini
Together AI

Integrations

Hugging Face
Ollama
OpenAI
OpenAI o1-mini
Together AI

Pricing Details

Free
Free Trial
Free Version

Pricing Details

Free
Free Trial
Free Version

Deployment

Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook

Deployment

Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook

Customer Support

Business Hours
Live Rep (24/7)
Online Support

Customer Support

Business Hours
Live Rep (24/7)
Online Support

Types of Training

Training Docs
Webinars
Live Training (Online)
In Person

Types of Training

Training Docs
Webinars
Live Training (Online)
In Person

Vendor Details

Company Name

Agentica Project

Founded

2025

Country

United States

Website

agentica-project.com

Vendor Details

Company Name

LG

Founded

1947

Country

South Korea

Website

github.com/LG-AI-EXAONE/EXAONE-Deep

Product Features

Product Features

Alternatives

Devstral 2 Reviews

Devstral 2

Mistral AI

Alternatives

Phi-4-reasoning Reviews

Phi-4-reasoning

Microsoft
DeepSWE Reviews

DeepSWE

Agentica Project
Qwen2 Reviews

Qwen2

Alibaba
Devstral Small 2 Reviews

Devstral Small 2

Mistral AI
DeepSeek R1 Reviews

DeepSeek R1

DeepSeek