Average Ratings 0 Ratings
Average Ratings 0 Ratings
Description
Pre-trained language models have made significant strides, achieving top-tier performance across multiple Natural Language Processing (NLP) applications. The impressive capabilities of GPT-3 highlight how increasing the scale of these models can unlock their vast potential. Recently, a comprehensive framework known as ERNIE 3.0 was introduced to pre-train large-scale models enriched with knowledge, culminating in a model boasting 10 billion parameters. This iteration of ERNIE 3.0 has surpassed the performance of existing leading models in a variety of NLP tasks. To further assess the effects of scaling, we have developed an even larger model called ERNIE 3.0 Titan, which consists of up to 260 billion parameters and is built on the PaddlePaddle platform. Additionally, we have implemented a self-supervised adversarial loss alongside a controllable language modeling loss, enabling ERNIE 3.0 Titan to produce texts that are both reliable and modifiable, thus pushing the boundaries of what these models can achieve. This approach not only enhances the model's capabilities but also opens new avenues for research in text generation and control.
Description
Recent breakthroughs in natural language processing, comprehension, and generation have been greatly influenced by the development of large language models. This research presents a system that employs Ascend 910 AI processors and the MindSpore framework to train a language model exceeding one trillion parameters, specifically 1.085 trillion, referred to as PanGu-{\Sigma}. This model enhances the groundwork established by PanGu-{\alpha} by converting the conventional dense Transformer model into a sparse format through a method known as Random Routed Experts (RRE). Utilizing a substantial dataset of 329 billion tokens, the model was effectively trained using a strategy called Expert Computation and Storage Separation (ECSS), which resulted in a remarkable 6.3-fold improvement in training throughput through the use of heterogeneous computing. Through various experiments, it was found that PanGu-{\Sigma} achieves a new benchmark in zero-shot learning across multiple downstream tasks in Chinese NLP, showcasing its potential in advancing the field. This advancement signifies a major leap forward in the capabilities of language models, illustrating the impact of innovative training techniques and architectural modifications.
API Access
Has API
API Access
Has API
Screenshots View All
No images available
Pricing Details
No price information available.
Free Trial
Free Version
Pricing Details
No price information available.
Free Trial
Free Version
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Vendor Details
Company Name
Baidu
Founded
2000
Country
China
Website
research.baidu.com/Public/uploads/61c4362c79ee8.pdf
Vendor Details
Company Name
Huawei
Founded
1987
Country
China
Website
huawei.com