Average Ratings 0 Ratings

Total
ease
features
design
support

No User Reviews. Be the first to provide a review:

Write a Review

Average Ratings 0 Ratings

Total
ease
features
design
support

No User Reviews. Be the first to provide a review:

Write a Review

Description

This innovative tool is designed for quantizing convolutional neural networks (CNNs). It allows for the transformation of both weights/biases and activations from 32-bit floating-point (FP32) to 8-bit integer (INT8) format, or even other bit depths. Utilizing this tool can greatly enhance inference performance and efficiency, all while preserving accuracy levels. It is compatible with various common layer types found in neural networks, such as convolution, pooling, fully-connected layers, and batch normalization, among others. Remarkably, the quantization process does not require the network to be retrained or the use of labeled datasets; only a single batch of images is sufficient. Depending on the neural network's size, the quantization can be completed in a matter of seconds to several minutes, facilitating quick updates to the model. Furthermore, this tool is specifically optimized for collaboration with DeePhi DPU and can generate the INT8 format model files necessary for DNNC integration. By streamlining the quantization process, developers can ensure their models remain efficient and robust in various applications.

Description

Voyage AI has unveiled voyage-code-3, an advanced embedding model specifically designed to enhance code retrieval capabilities. This innovative model achieves superior performance, surpassing OpenAI-v3-large and CodeSage-large by averages of 13.80% and 16.81% across a diverse selection of 32 code retrieval datasets. It accommodates embeddings of various dimensions, including 2048, 1024, 512, and 256, and provides an array of embedding quantization options such as float (32-bit), int8 (8-bit signed integer), uint8 (8-bit unsigned integer), binary (bit-packed int8), and ubinary (bit-packed uint8). With a context length of 32 K tokens, voyage-code-3 exceeds the limitations of OpenAI's 8K and CodeSage Large's 1K context lengths, offering users greater flexibility. Utilizing an innovative approach known as Matryoshka learning, it generates embeddings that feature a layered structure of varying lengths within a single vector. This unique capability enables users to transform documents into a 2048-dimensional vector and subsequently access shorter dimensional representations (such as 256, 512, or 1024 dimensions) without the need to re-run the embedding model, thus enhancing efficiency in code retrieval tasks. Additionally, voyage-code-3 positions itself as a robust solution for developers seeking to improve their coding workflow.

API Access

Has API

API Access

Has API

Screenshots View All

Screenshots View All

Integrations

Elasticsearch
Milvus
Qdrant
Vespa
Weaviate

Integrations

Elasticsearch
Milvus
Qdrant
Vespa
Weaviate

Pricing Details

$0.90 per hour
Free Trial
Free Version

Pricing Details

No price information available.
Free Trial
Free Version

Deployment

Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook

Deployment

Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook

Customer Support

Business Hours
Live Rep (24/7)
Online Support

Customer Support

Business Hours
Live Rep (24/7)
Online Support

Types of Training

Training Docs
Webinars
Live Training (Online)
In Person

Types of Training

Training Docs
Webinars
Live Training (Online)
In Person

Vendor Details

Company Name

DeePhi Quantization Tool

Website

aws.amazon.com/marketplace/pp/prodview-bwtx6kzwg3gva

Vendor Details

Company Name

MongoDB

Founded

2007

Country

United States

Website

blog.voyageai.com/2024/12/04/voyage-code-3/

Product Features

Product Features

Alternatives

Deci Reviews

Deci

Deci AI

Alternatives

voyage-4-large Reviews

voyage-4-large

Voyage AI
Voyage AI Reviews

Voyage AI

MongoDB