Agents have been the primary drivers of applying large language models (LLMs) to solve complex problems. Since AutoGPT in 2023, various techniques have been…
Agents have been the primary drivers of applying large language models (LLMs) to solve complex problems. Since AutoGPT in 2023, various techniques have been developed to build reliable agents across industries. The discourse around agentic reasoning and AI reasoning models further adds a layer of nuance when designing these applications. The rapid pace of this development also makes it hard for…
NVIDIA DALI, a portable, open source software library for decoding and augmenting images, videos, and speech, recently introduced several features that improve…
NVIDIA DALI, a portable, open source software library for decoding and augmenting images, videos, and speech, recently introduced several features that improve performance and enable DALI with new use cases. These updates aim at simplifying the integration of DALI into existing PyTorch data processing logic, improving flexibility in building data processing pipelines by enabling CPU-to-GPU flows…
LLM Streaming sends a model’s response incrementally in real time, token by token, as it’s being generated. The output streaming capability has evolved…
LLM Streaming sends a model’s response incrementally in real time, token by token, as it’s being generated. The output streaming capability has evolved from a nice-to-have feature to an essential component of modern LLM applications. The traditional approach of waiting several seconds for full LLM responses creates delays, especially in complex applications with multiple model calls.
Researchers, using AI to analyze routine brain scans, have discovered a promising new method to reliably identify a common but hard-to-detect precursor of many…
Researchers, using AI to analyze routine brain scans, have discovered a promising new method to reliably identify a common but hard-to-detect precursor of many strokes. In a study published in the journal Cerebrovascular Diseases, scientists from the Royal Melbourne Hospital described a new AI model that could one day prevent at-risk patients from becoming stroke victims.
NVIDIA has achieved a world-record large language model (LLM) inference speed. A single NVIDIA DGX B200 node with eight NVIDIA Blackwell GPUs can achieve over…
NVIDIA has achieved a world-record large language model (LLM) inference speed. A single NVIDIA DGX B200 node with eight NVIDIA Blackwell GPUs can achieve over 1,000 tokens per second (TPS) per user on the 400-billion-parameter Llama 4 Maverick model, the largest and most powerful model available in the Llama 4 collection. This speed was independently measured by the AI benchmarking service…
Computing is an essential tool for the modern financial services industry. Profits are won and lost based on the speed and accuracy of algorithms guiding…
Computing is an essential tool for the modern financial services industry. Profits are won and lost based on the speed and accuracy of algorithms guiding financial decision making. Accelerated quantum computing has the potential to impact the financial services industry with new algorithms able to speed-up or enhance existing tools, such as portfolio optimization techniques.
What does it take to win a Kaggle competition in 2025? In the April Playground challenge, the goal was to predict how long users would listen to a podcast—and…
What does it take to win a Kaggle competition in 2025? In the April Playground challenge, the goal was to predict how long users would listen to a podcast—and the top solution wasn’t just accurate, it was fast. In this post, Kaggle Grandmaster Chris Deotte will break down the exact stacking strategy that powered his first-place finish using GPU-accelerated modeling with cuML. You’ll learn a…