Average Ratings 0 Ratings

Total
ease
features
design
support

No User Reviews. Be the first to provide a review:

Write a Review

Average Ratings 0 Ratings

Total
ease
features
design
support

No User Reviews. Be the first to provide a review:

Write a Review

Description

CodeT5 is an innovative pre-trained encoder-decoder model specifically designed for understanding and generating code. This model is identifier-aware and serves as a unified framework for various coding tasks. The official PyTorch implementation originates from a research paper presented at EMNLP 2021 by Salesforce Research. A notable variant, CodeT5-large-ntp-py, has been fine-tuned to excel in Python code generation, forming the core of our CodeRL approach and achieving groundbreaking results in the APPS Python competition-level program synthesis benchmark. This repository includes the necessary code for replicating the experiments conducted with CodeT5. Pre-trained on an extensive dataset of 8.35 million functions across eight programming languages—namely Python, Java, JavaScript, PHP, Ruby, Go, C, and C#—CodeT5 has demonstrated exceptional performance, attaining state-of-the-art results across 14 different sub-tasks in the code intelligence benchmark known as CodeXGLUE. Furthermore, it is capable of generating code directly from natural language descriptions, showcasing its versatility and effectiveness in coding applications.

Description

Pre-trained language models have made significant strides, achieving top-tier performance across multiple Natural Language Processing (NLP) applications. The impressive capabilities of GPT-3 highlight how increasing the scale of these models can unlock their vast potential. Recently, a comprehensive framework known as ERNIE 3.0 was introduced to pre-train large-scale models enriched with knowledge, culminating in a model boasting 10 billion parameters. This iteration of ERNIE 3.0 has surpassed the performance of existing leading models in a variety of NLP tasks. To further assess the effects of scaling, we have developed an even larger model called ERNIE 3.0 Titan, which consists of up to 260 billion parameters and is built on the PaddlePaddle platform. Additionally, we have implemented a self-supervised adversarial loss alongside a controllable language modeling loss, enabling ERNIE 3.0 Titan to produce texts that are both reliable and modifiable, thus pushing the boundaries of what these models can achieve. This approach not only enhances the model's capabilities but also opens new avenues for research in text generation and control.

API Access

Has API

API Access

Has API

Screenshots View All

Screenshots View All

Integrations

C
C#
ERNIE Bot
Go
Java
JavaScript
PHP
Python
Ruby

Integrations

C
C#
ERNIE Bot
Go
Java
JavaScript
PHP
Python
Ruby

Pricing Details

No price information available.
Free Trial
Free Version

Pricing Details

No price information available.
Free Trial
Free Version

Deployment

Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook

Deployment

Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook

Customer Support

Business Hours
Live Rep (24/7)
Online Support

Customer Support

Business Hours
Live Rep (24/7)
Online Support

Types of Training

Training Docs
Webinars
Live Training (Online)
In Person

Types of Training

Training Docs
Webinars
Live Training (Online)
In Person

Vendor Details

Company Name

Salesforce

Website

github.com/salesforce/CodeT5

Vendor Details

Company Name

Baidu

Founded

2000

Country

China

Website

research.baidu.com/Public/uploads/61c4362c79ee8.pdf

Product Features

Product Features

Alternatives

Mu Reviews

Mu

Microsoft

Alternatives

GLM-OCR Reviews

GLM-OCR

Z.ai
PanGu-α Reviews

PanGu-α

Huawei
ERNIE X1 Reviews

ERNIE X1

Baidu
ERNIE 4.5 Reviews

ERNIE 4.5

Baidu