Categories
Misc

NVIDIA Triton Inference Server Achieves Outstanding Performance in MLPerf Inference 4.1 Benchmarks

Decorative image.Six years ago, we embarked on a journey to develop an AI inference serving solution specifically designed for high-throughput and time-sensitive production use…Decorative image.

Six years ago, we embarked on a journey to develop an AI inference serving solution specifically designed for high-throughput and time-sensitive production use cases from the ground up. At that time, ML developers were deploying bespoke, framework-specific AI solutions, which were driving up their operational costs and not meeting their latency and throughput service level agreements.

Source

Categories
Misc

New Foundational Models and Training Capabilities with NVIDIA TAO 5.5

GIF shows multiple photos and images selected within the photos according to a prompt, such as "person with glasses" or "tallest cat."NVIDIA TAO is a framework designed to simplify and accelerate the development and deployment of AI models. It enables you to use pretrained models, fine-tune…GIF shows multiple photos and images selected within the photos according to a prompt, such as

NVIDIA TAO is a framework designed to simplify and accelerate the development and deployment of AI models. It enables you to use pretrained models, fine-tune them with your own data, and optimize the models for specific use cases without needing deep AI expertise. TAO integrates seamlessly with the NVIDIA hardware and software ecosystem, providing tools for efficient AI model training…

Source

Categories
Misc

NVIDIA Blackwell Sets New Standard for Generative AI in MLPerf Inference Debut

As enterprises race to adopt generative AI and bring new services to market, the demands on data center infrastructure have never been greater. Training large language models is one challenge, but delivering LLM-powered real-time services is another. In the latest round of MLPerf industry benchmarks, Inference v4.1, NVIDIA platforms delivered leading performance across all data
Read Article

Categories
Misc

Deploy Diverse AI Apps with Multi-LoRA Support on RTX AI PCs and Workstations

An illustration depicting AI model deployment steps.Today’s large language models (LLMs) achieve unprecedented results across many use cases. Yet, application developers often need to customize and tune these…An illustration depicting AI model deployment steps.

Today’s large language models (LLMs) achieve unprecedented results across many use cases. Yet, application developers often need to customize and tune these models to work specifically for their use cases, due to the general nature of foundation models. Full fine-tuning requires a large amount of data and compute infrastructure, resulting in model weights being updated.

Source

Categories
Misc

Low Latency Inference Chapter 1: Up to 1.9X Higher Llama 3.1 Performance with Medusa on NVIDIA HGX H200 with NVLink Switch

Image of an HGX H200As large language models (LLMs) continue to grow in size and complexity, multi-GPU compute is a must-have to deliver the low latency and high throughput that…Image of an HGX H200

As large language models (LLMs) continue to grow in size and complexity, multi-GPU compute is a must-have to deliver the low latency and high throughput that real-time generative AI applications demand. Performance depends both on the ability for the combined GPUs to process requests as “one mighty GPU” with ultra-fast GPU-to-GPU communication and advanced software able to take full…

Source

Categories
Misc

More Than Fine: Multi-LoRA Support Now Available in NVIDIA RTX AI Toolkit

Large language models are driving some of the most exciting developments in AI with their ability to quickly understand, summarize and generate text-based content.

Categories
Misc

Simplifying Camera Calibration to Enhance AI-Powered Multi-Camera Tracking

This post is the third in a series on building multi-camera tracking vision AI applications. We introduce the overall end-to-end workflow and fine-tuning…

This post is the third in a series on building multi-camera tracking vision AI applications. We introduce the overall end-to-end workflow and fine-tuning process to enhance system accuracy in the first part and second part. NVIDIA Metropolis is an application framework and set of developer tools that leverages AI for visual data analysis across industries. Its multi-camera tracking reference…

Source

Categories
Misc

Scaling robotics datasets with video encoding

Categories
Misc

Enhancing RAG Applications with NVIDIA NIM

Workflow diagram on a black backgound.The advent of large language models (LLMs) has significantly benefited the AI industry, offering versatile tools capable of generating human-like text and…Workflow diagram on a black backgound.

The advent of large language models (LLMs) has significantly benefited the AI industry, offering versatile tools capable of generating human-like text and handling a wide range of tasks. However, while LLMs demonstrate impressive general knowledge, their performance in specialized fields, such as veterinary science, is limited when used out of the box. To enhance their utility in specific areas…

Source

Categories
Misc

Optimize Large-Scale AI Workloads with NVIDIA Spectrum-X

Decorative image of networking switches.In today’s rapidly evolving technological landscape, staying ahead of the curve is not just a goal—it’s a necessity. The surge of innovations, particularly…Decorative image of networking switches.

In today’s rapidly evolving technological landscape, staying ahead of the curve is not just a goal—it’s a necessity. The surge of innovations, particularly in AI, is driving dramatic changes across the technology stack. One area witnessing profound transformation is Ethernet networking, a cornerstone of digital communication that has been foundational to enterprise and data center…

Source