Average Ratings 0 Ratings
Average Ratings 0 Ratings
Description
AudioCraft serves as a comprehensive codebase tailored for all your generative audio requirements, including music, sound effects, and compression, following its training on raw audio signals. By utilizing AudioCraft, we enhance the design of generative audio models significantly compared to earlier methodologies. Both MusicGen and AudioGen rely on a unified autoregressive Language Model (LM) that functions across streams of compressed discrete music representations known as tokens. We propose a straightforward technique to exploit the intrinsic structure of the parallel token streams, demonstrating that with a single model and a refined interleaving pattern, we can effectively model audio sequences while capturing long-term dependencies, resulting in the generation of high-quality audio outputs. Our models utilize the EnCodec neural audio codec to derive discrete audio tokens from the raw waveform, with EnCodec transforming the audio signal into multiple parallel streams of discrete tokens. This innovative approach not only streamlines audio generation but also enhances the overall efficiency and quality of the output.
Description
PanGu-α has been created using the MindSpore framework and utilizes a powerful setup of 2048 Ascend 910 AI processors for its training. The training process employs an advanced parallelism strategy that leverages MindSpore Auto-parallel, which integrates five different parallelism dimensions—data parallelism, operation-level model parallelism, pipeline model parallelism, optimizer model parallelism, and rematerialization—to effectively distribute tasks across the 2048 processors. To improve the model's generalization, we gathered 1.1TB of high-quality Chinese language data from diverse fields for pretraining. We conduct extensive tests on PanGu-α's generation capabilities across multiple situations, such as text summarization, question answering, and dialogue generation. Additionally, we examine how varying model scales influence few-shot performance across a wide array of Chinese NLP tasks. The results from our experiments highlight the exceptional performance of PanGu-α, demonstrating its strengths in handling numerous tasks even in few-shot or zero-shot contexts, thus showcasing its versatility and robustness. This comprehensive evaluation reinforces the potential applications of PanGu-α in real-world scenarios.
API Access
Has API
API Access
Has API
Screenshots View All
No images available
Integrations
No details available.
Integrations
No details available.
Pricing Details
No price information available.
Free Trial
Free Version
Pricing Details
No price information available.
Free Trial
Free Version
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Vendor Details
Company Name
Meta AI
Founded
2004
Country
United States
Website
audiocraft.metademolab.com
Vendor Details
Company Name
Huawei
Founded
1987
Country
China
Website
arxiv.org/abs/2104.12369