Average Ratings 0 Ratings
Average Ratings 0 Ratings
Description
Software transforms a vehicle into a smart machine, and the NVIDIA DRIVE™ Software stack serves as an open platform that enables developers to effectively create and implement a wide range of advanced autonomous vehicle applications, such as perception, localization and mapping, planning and control, driver monitoring, and natural language processing. At the core of this software ecosystem lies DRIVE OS, recognized as the first operating system designed for safe accelerated computing. This system incorporates NvMedia for processing sensor inputs, NVIDIA CUDA® libraries to facilitate efficient parallel computing, and NVIDIA TensorRT™ for real-time artificial intelligence inference, alongside numerous tools and modules that provide access to hardware capabilities. The NVIDIA DriveWorks® SDK builds on DRIVE OS, offering essential middleware functions that are critical for the development of autonomous vehicles. These functions include a sensor abstraction layer (SAL) and various sensor plugins, a data recorder, vehicle I/O support, and a framework for deep neural networks (DNN), all of which are vital for enhancing the performance and reliability of autonomous systems. With these powerful resources, developers are better equipped to innovate and push the boundaries of what's possible in automated transportation.
Description
VLLM is an advanced library tailored for the efficient inference and deployment of Large Language Models (LLMs). Initially created at the Sky Computing Lab at UC Berkeley, it has grown into a collaborative initiative enriched by contributions from both academic and industry sectors. The library excels in providing exceptional serving throughput by effectively handling attention key and value memory through its innovative PagedAttention mechanism. It accommodates continuous batching of incoming requests and employs optimized CUDA kernels, integrating technologies like FlashAttention and FlashInfer to significantly improve the speed of model execution. Furthermore, VLLM supports various quantization methods, including GPTQ, AWQ, INT4, INT8, and FP8, and incorporates speculative decoding features. Users enjoy a seamless experience by integrating easily with popular Hugging Face models and benefit from a variety of decoding algorithms, such as parallel sampling and beam search. Additionally, VLLM is designed to be compatible with a wide range of hardware, including NVIDIA GPUs, AMD CPUs and GPUs, and Intel CPUs, ensuring flexibility and accessibility for developers across different platforms. This broad compatibility makes VLLM a versatile choice for those looking to implement LLMs efficiently in diverse environments.
API Access
Has API
API Access
Has API
Integrations
KServe
BlinqIO
Brytlyt
Chainer
Cisco Kinetic
Cognata
DataMotto
Google Cloud GPUs
GridMatrix
Hugging Face
Integrations
KServe
BlinqIO
Brytlyt
Chainer
Cisco Kinetic
Cognata
DataMotto
Google Cloud GPUs
GridMatrix
Hugging Face
Pricing Details
No price information available.
Free Trial
Free Version
Pricing Details
No price information available.
Free Trial
Free Version
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Vendor Details
Company Name
NVIDIA
Founded
1993
Country
United States
Website
www.nvidia.com/en-us/self-driving-cars/drive-platform/software/
Vendor Details
Company Name
VLLM
Country
United States
Website
docs.vllm.ai/en/latest/