Average Ratings 0 Ratings
Average Ratings 0 Ratings
Description
This innovative tool is designed for quantizing convolutional neural networks (CNNs). It allows for the transformation of both weights/biases and activations from 32-bit floating-point (FP32) to 8-bit integer (INT8) format, or even other bit depths. Utilizing this tool can greatly enhance inference performance and efficiency, all while preserving accuracy levels. It is compatible with various common layer types found in neural networks, such as convolution, pooling, fully-connected layers, and batch normalization, among others. Remarkably, the quantization process does not require the network to be retrained or the use of labeled datasets; only a single batch of images is sufficient. Depending on the neural network's size, the quantization can be completed in a matter of seconds to several minutes, facilitating quick updates to the model. Furthermore, this tool is specifically optimized for collaboration with DeePhi DPU and can generate the INT8 format model files necessary for DNNC integration. By streamlining the quantization process, developers can ensure their models remain efficient and robust in various applications.
Description
Mirai is an advanced platform tailored for developers that focuses on on-device AI infrastructure, enabling the conversion, optimization, and execution of machine learning models directly on Apple devices with a strong emphasis on performance and user privacy. This platform offers a cohesive workflow that allows teams to efficiently convert and quantize models, assess their performance, distribute them, and conduct local inference seamlessly. Specifically designed for Apple Silicon, Mirai strives to achieve near-zero latency and zero inference cost, while ensuring that sensitive data processing remains securely on the user's device. Through its comprehensive SDK and inference engine, developers can swiftly integrate AI functionalities into their applications, leveraging hardware-aware optimizations to maximize the capabilities of the GPU and Neural Engine. Additionally, Mirai features dynamic routing abilities that intelligently determine the best execution path for requests, whether that be locally on the device or utilizing cloud resources, taking into account factors such as latency, privacy, and workload demands. This flexibility not only enhances the user experience but also allows developers to create more responsive and efficient applications tailored to their users' needs.
API Access
Has API
API Access
Has API
Integrations
DeepSeek R1
Gemma 3
Gemma 4
LFM-3B
Llama
Polaris
Qwen3
SmolLM2
gpt-oss-120b
Integrations
DeepSeek R1
Gemma 3
Gemma 4
LFM-3B
Llama
Polaris
Qwen3
SmolLM2
gpt-oss-120b
Pricing Details
$0.90 per hour
Free Trial
Free Version
Pricing Details
No price information available.
Free Trial
Free Version
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Vendor Details
Company Name
DeePhi Quantization Tool
Website
aws.amazon.com/marketplace/pp/prodview-bwtx6kzwg3gva
Vendor Details
Company Name
Mirai
Founded
2024
Country
United States
Website
trymirai.com