Average Ratings 0 Ratings
Average Ratings 0 Ratings
Description
This innovative tool is designed for quantizing convolutional neural networks (CNNs). It allows for the transformation of both weights/biases and activations from 32-bit floating-point (FP32) to 8-bit integer (INT8) format, or even other bit depths. Utilizing this tool can greatly enhance inference performance and efficiency, all while preserving accuracy levels. It is compatible with various common layer types found in neural networks, such as convolution, pooling, fully-connected layers, and batch normalization, among others. Remarkably, the quantization process does not require the network to be retrained or the use of labeled datasets; only a single batch of images is sufficient. Depending on the neural network's size, the quantization can be completed in a matter of seconds to several minutes, facilitating quick updates to the model. Furthermore, this tool is specifically optimized for collaboration with DeePhi DPU and can generate the INT8 format model files necessary for DNNC integration. By streamlining the quantization process, developers can ensure their models remain efficient and robust in various applications.
Description
FriendliAI serves as an advanced generative AI infrastructure platform that delivers rapid, efficient, and dependable inference solutions tailored for production settings. The platform is equipped with an array of tools and services aimed at refining the deployment and operation of large language models (LLMs) alongside various generative AI tasks on a large scale. Among its key features is Friendli Endpoints, which empowers users to create and implement custom generative AI models, thereby reducing GPU expenses and hastening AI inference processes. Additionally, it facilitates smooth integration with well-known open-source models available on the Hugging Face Hub, ensuring exceptionally fast and high-performance inference capabilities. FriendliAI incorporates state-of-the-art technologies, including Iteration Batching, the Friendli DNN Library, Friendli TCache, and Native Quantization, all of which lead to impressive cost reductions (ranging from 50% to 90%), a significant decrease in GPU demands (up to 6 times fewer GPUs), enhanced throughput (up to 10.7 times), and a marked decrease in latency (up to 6.2 times). With its innovative approach, FriendliAI positions itself as a key player in the evolving landscape of generative AI solutions.
API Access
Has API
API Access
Has API
Integrations
Amazon Web Services (AWS)
DeepSeek
Gemma 3
Gemma 4
Grafana Cloud
Hugging Face
Kubernetes
LangChain
LiteLLM
Llama 3.3
Integrations
Amazon Web Services (AWS)
DeepSeek
Gemma 3
Gemma 4
Grafana Cloud
Hugging Face
Kubernetes
LangChain
LiteLLM
Llama 3.3
Pricing Details
$0.90 per hour
Free Trial
Free Version
Pricing Details
$5.9 per hour
Free Trial
Free Version
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Vendor Details
Company Name
DeePhi Quantization Tool
Website
aws.amazon.com/marketplace/pp/prodview-bwtx6kzwg3gva
Vendor Details
Company Name
FriendliAI
Founded
2021
Country
United States
Website
friendli.ai/