Categories
Misc

Video Processing and Streaming – Top Resources from GTC 21

This year at GTC we announced the release of NVIDIA Maxine, a GPU-accelerated SDK for building innovative virtual collaboration and content creation applications such as video conferencing and live streaming.

AI has been instrumental in providing exciting features and improving quality and operational efficiency for conferencing, media delivery and content creation. 

This year at GTC we announced the release of NVIDIA Maxine, a GPU-accelerated SDK for building innovative virtual collaboration and content creation applications such as video conferencing and live streaming. 

Check out some of the most popular sessions, demos, and videos from GTC showcasing Maxine’s latest advancements:

The developer resources listed below are exclusively available to NVIDIA Developer Program members. Join today for free in order to get access to the tools and training necessary to build on NVIDIA’s technology platform here

SDK

NVIDIA Maxine Now Available
With Maxine’s AI SDKs—Video Effects, Audio Effects, and Augmented Reality (AR)—developers can now create real-time, video-based experiences easily deployed to PCs, data centers, and the cloud. Maxine can also leverage NVIDIA Jarvis to access conversational AI capabilities such as transcription, translation, and virtual assistants.

On-Demand

How NVIDIA’s Maxine Changed the Way We Communicate
Hear from Avaya’s Mike Kuch, Sr. Director of Solutions Marketing, and Paul Relf, Sr. Director of Product Management about Avaya Spaces built on CPaaS. Avaya is making capabilities associated with meetings available in contact centers. With AI noise elimination, agents and customers can hear each other in noisy environments. We’re combining components to realize the art of the possible for unique experiences by Avaya with NVIDIA AI.

Real-time AI for Video-Conferencing with Maxine
Learn from Andrew Rabinovich, Co-Founder and CTO, Julian Green, Co-Founder and CEO, and Tarrence van As, Co-Founder and Principal Engineer, from Headroom about applying the latest AI research on real-time video and audio streams for a more-human video-conferencing application. Explore employing generative models for super-resolution, giving order-of-magnitude reduced bandwidth. See new solutions for saliency segmentation delivering contextual virtual backgrounds of stuff that matters. 

Demo

Building AI-Powered Virtual Collaboration and Content Creation Solutions with NVIDIA Maxine
With new state-of the-art AI features for video, audio, and augmented reality—including AI face codec, eye contact, super resolution, noise removal, and more—NVIDIA Maxine is reinventing virtual collaboration on PCs, in the data center, and in the cloud. 

Reinvent Video Conferencing, Content Creation & Streaming with AI Using NVIDIA Maxine
Developers from video conferencing, content creation and streaming providers such as Notch, Headroom, Be.Live, and Touchcast are using the Maxine SDK to create real-time video-based experiences easily deployed to PCs, data centers or in the cloud.

Categories
Misc

Data Center Networking – Top Resources from GTC 21

NVIDIA is enabling these organizations to easily develop accelerated applications and implement cybersecurity frameworks in order to deliver breakthrough networking, security, and storage performance with a comprehensive, open development platform.

As organizations embrace cloud and edge computing models, they are looking for more efficient, modern computing architectures that create a secure, accelerated, virtual private cloud (SA-VPC), able to support multi-tenancy and deliver applications at data center scale with all the necessary levels of performance and cyber protection. NVIDIA is enabling these organizations to easily develop accelerated applications and implement cybersecurity frameworks in order to deliver breakthrough networking, security, and storage performance with a comprehensive, open development platform.

The developer resources listed below are exclusively available to NVIDIA Developer Program members. Join today for free in order to get access to the tools and training necessary to build on NVIDIA’s technology platform here

 

On-Demand Session

Program Data Center Infrastructure Acceleration with the Release of DOCA and the Latest DPU Software
Speakers: Ariel Kit, Director of Product Marketing for Networking, NVIDIA  and Ami Badani, Vice President of Marketing NVIDIA

DPU experts Ami Badani and Ariel Kit discuss how NVIDIA DOCA is enabling new infrastructure acceleration and management features in BlueField all while simplifying programming and application integration.

Morpheus: AI Inferencing for Cybersecurity Pipelines
Speaker: Bartley Richardson, NVIDIA

What does NVIDIA Morpheus mean for the future of the data center and cloud security? Take a deep-dive into the newly announced AI cybersecurity framework with engineer manager, Bartley Richardson by watching this on demand GTC 21 session.

SDK

NVIDIA DOCA
Develop applications with breakthrough networking, security, and storage performance using NVIDIA DOCA — the newly released complete, open software platform.

NVIDIA Morpheus
Detect Cybersecurity threats in an Instant with NVIDIA Morpheus, a new AI framework for creating zero-trust data center security.

Click here to view all of the Data Center Networking sessions and demos on NVIDIA On-Demand.

Categories
Misc

Develop Robotics Applications – Top Resources from GTC 21

NVIDIA Isaac is a developer toolbox for accelerating the development and deployment of AI-powered robots. The SDK includes Isaac applications, GEMs (robot capabilities), a Robot Engine, and Isaac Sim.

Isaac SDK is the robotics platform for accelerating the development and deployment of robotics applications. The SDK is the toolkit which is GPU-optimized for AI and computer vision applications, including perception, navigation, and manipulation features enabled by AI.

Isaac Sim leverages the powerful NVIDIA Omniverse to build the next generation of robotics and AI simulator. Start building virtual robotic worlds and experiments, supporting navigation and manipulation applications through the Isaac SDK with RGB-D, lidar and inertial measurement unit (IMU) sensors, domain randomization, ground truth labeling, segmentation, and bounding boxes.

Here are some resources to introduce you to the Isaac platform.

The developer resources listed below are exclusively available to NVIDIA Developer Program members. Join today for free in order to get access to the tools and training necessary to build on NVIDIA’s technology platform here.

On-Demand Sessions

Sim-to-Real in Isaac Sim
Speakers: Hai Loc Lu, Lead System Software Engineer, NVIDIA; Michael Gussert, Deep Learning Engineer, NVIDIA

Learn how to train and test robots in virtual environments with Isaac Sim on Omniverse, then transfer to physical Jetson powered robots.

Isaac Gym: End-to-End GPU-Accelerated Reinforcement Learning
Speakers: Gavriel State, Senior Director for Simulation and AI, NVIDIA; Lukasz Wawrzyniak, Senior Engineer, NVIDIA

Isaac Gym is NVIDIA’s environment for high-performance reinforcement learning on GPUs. We will review key API features, demonstrate examples of training agents, and provide updates on future integration of Isaac Gym functionality within the NVIDIA Omniverse platform. We will demonstrate how to create environments with thousands of agents to train in parallel, and how the Isaac Gym system allows developers to create tensor based views of physics state for all environments. We will also demonstrate the application of physics based domain randomization in Isaac Gym, which can help with sim2real transfer of learned policies to physical robots.

Bridging Sim2Real Gap: Simulation Tuning for Training Deep Learning Robotic Perception Models
Speaker: Peter Dykas, Solutions Architect, NVIDIA

Deep neural networks enable accurate perception for robots. Simulation offers a way to train deep learning robotic perception models that were previously not possible in scenarios where it is prohibitively expensive, time-consuming, or infeasible to collect large labeled datasets. We’ll dive into how NVIDIA is bridging the gap between simulation and reality with domain randomization, photorealistic simulation, and accurate physics imitation with Isaac Sim, and more.

Docs

NVIDIA Carter
Carter is a robot developed as a platform to demonstrate the capabilities of the Isaac SDK. It is based on a differential drive and uses lidar and a camera to perceive the world. This document walks you through hardware assembly and software setup for Carter.

Getting Started Tutorials and Sample Applications
Over 30 tutorials and samples provided with Isaac SDK to get you started.

Click here to view more Isaac SDK sessions on NVIDIA On-Demand.

Categories
Misc

Automotive – Top Resources from GTC 21

The annual DRIVE Developer Days was held during GTC 2021, featuring a series of specialized sessions on AV development led by NVIDIA experts. Learn about perception, mapping, simulation and more anytime with NVIDIA On-Demand.

The annual DRIVE Developer Days was held during GTC 2021, featuring a series of specialized sessions on autonomous vehicle hardware and software, including perception, mapping, simulation and more, all led by NVIDIA experts. These sessions are now available to view anytime with NVIDIA On-Demand.

The developer resources listed below are exclusively available to NVIDIA Developer Program members. Join today for free in order to get access to the tools and training necessary to build on NVIDIA’s technology platform here

 

On-Demand Sessions

DRIVE AGX Hardware Update with NVIDIA Orin

Speaker: Gary Hicok, Senior Vice President, Hardware and Systems, NVIDIA

This session will provide an early look at the next generation of DRIVE AGX hardware platforms based on the upcoming NVIDIA Orin SoC.

 

Turbocharge Autonomous Vehicle Development with DRIVE OS and DriveWorks

Speaker: Stephen Jones, Product Line Manager & Hope Allen, Product Manager, DriveWorks, NVIDIA

Learn how NVIDIA DRIVE OS and DriveWorks turbocharge autonomous vehicle development, delivering foundational autonomous tools and functional safety while simultaneously optimizing NVIDIA DRIVE AGX compute performance.

 

DRIVE AV Perception Overview

Speaker: Chongyu Wang, Product Manager

The ability to interpret a scene with 360° awareness is a critical function of an autonomous vehicle. In this session, we highlight the NVIDIA DRIVE AV Perception software stack, including an architecture overview and our latest algorithmic results.

 

Mapping and Localization with DRIVE AV

Speaker: Rambo Jacoby, Principal Product Manager, NVIDIA

The use of HD maps is a key part of ensuring a safe and comfortable journey. In this session, we’ll provide an overview of NVIDIA’s end-to-end solution for creating and maintaining crowdsourced HD maps, and how they’re used for vehicle localization.

 

A Modular Approach to AV Planning and Control

Speaker: Alexey Baranov, Senior Product Manager, NVIDIA

Planning and control executes maneuvers using input from perception, prediction, and mapping. In this session, we review the NVIDIA DRIVE AV modular approach to planning and control software and the variety of capabilities it enables.

 

Leveraging EGX and DGX for Developing AV Platforms and Supporting Connected Services

Speaker: Rambo Jacoby, Principal Product Manager, NVIDIA

In this session, we look at how NVIDIA DGX and NVIDIA EGX and are used to create the network of data centers and edge devices necessary for developing an AV platform and delivering functionality and connected services to vehicles of the future.

 

Automated Testing at Scale to Enable Deployment of Autonomous Vehicles

Speaker: Justyna Zander, Global Head of Verification and Validation, NVIDIA

In this session, we discuss the use of simulation and computing infrastructure for AV development. We also demonstrate a scalable and automated set of solutions for end-to-end testing to enable AV deployment on the road, according to safety standards.

 

NVIDIA DRIVE Sim and Omniverse

Speaker: Matt Cragun, Senior Product Manager, AV Simulation, NVIDIA

This session covers the use of simulation and computing infrastructure for AV development. We also demonstrate a scalable and automated set of solutions for end-to-end testing to enable AV deployment on the road, according to safety standards.

 

Click here to view all of the Automotive sessions and demos on NVIDIA On-Demand.

 

Categories
Misc

Healthcare – Top Resources from GTC 21

Here are the latest resources and news for healthcare developers from GTC 21, including demos and specialized sessions for building AI in drug discovery, medical imaging, genomics, and smart hospitals.

Here are the latest resources and news for healthcare developers from GTC 21, including demos and specialized sessions for building AI in drug discovery, medical imaging, genomics, and smart hospitals. Learn about new features now available in NVIDIA Clara Train 4.0, an application framework for medical imaging that includes pre-trained models, AI-assisted annotation, AutoML, and federated learning.

The developer resources listed below are exclusively available to NVIDIA Developer Program members. Join today for free in order to get access to the tools and training necessary to build on NVIDIA’s technology platform here.

On-Demand Sessions

Accelerating Drug Discovery with Advanced Computational Modeling
Speaker: Robert Abel, Executive Vice President, Chief Computational Scientist, Schrödinger

Learn about how integrated deployment and collaborative use of advanced computational modeling and next-generation machine learning can accelerate drug discovery from Robert Abel, Executive Vice President, Chief Computational Scientist at Schrödinger.

Using Ethernet to Stream Medical Sensor Data
Speaker: Mathias Blake, Platform Architect for Medical Devices, NVIDIA

Explore three technologies from NVIDIA that make streaming high-throughput medical sensor data over Ethernet easy and efficient—NVIDIA Networking ConnectX NICs, Rivermax SDK with GPUDirect, and Clara AGX. Learn about the capabilities of each of these technologies and explore examples of how these technologies can be leveraged by several different types of medical devices.

Automate 3D Medical Imaging Segmentation with AutoML and Neural Architecture Search
Speaker: Dong Yang, Applied Research Scientist, NVIDIA

Recently, neural architecture search (NAS) has been applied to automatically search high-performance networks for medical image segmentation. Hear from NVIDIA Applied Research Scientist, Dong Yang, to learn about AutoML and NAS techniques in the Clara Train SDK.

Deep Learning and Accelerated Computing for Single-Cell Genomic Data
Speaker: Avantika Lal, Sr. Scientist in Deep Learning and Genomics, NVIDIA

Learn about accelerating discovery of cell types in the human body with RAPIDS and AtacWorks, a deep learning toolkit to enhance ATAC-seq data and identify active regulatory DNA more accurately than existing state-of-the-art methods.

Blog

Creating Medical Imaging Models with Clara Train 4.0
Learn about the upcoming release of NVIDIA Clara Train 4.0, including infrastructure upgrades based on MONAI, expansion into digital pathology, and updates to DeepGrow for annotating organs effectively in 3D images.

Demos

Accelerating Drug Discovery with Clara Discovery’s MegaMolBart
See how NVIDIA Clara Discovery’s MegaMolBart, a transformer-based NLP model developed with AstraZeneca, trained on millions of molecules, can accelerate the drug discovery process.

NVIDIA Triton Inference Server: Generative Chemical Structures
Watch NVIDIA Triton Inference Server power deep learning models to propose thousands of molecules per second for drug design that can be further refined with physics-based simulations.

Visit NVIDIA On-Demand to explore the extensive catalog of sessions, podcasts, demos, research posters and more.

Categories
Misc

Conversational AI and NLP – Top Resources from GTC 21

NVIDIA announced several major breakthroughs in conversational AI for building and deploying ASR, NLP and TTS applications.

At GTC 21, NVIDIA announced several major breakthroughs in conversational AI for building and deploying automatic speech recognition (ASR), natural language processing (NLP), and text-to-speech (TTS) applications. The conference also hosted over 60 engaging sessions and workshops featuring the latest tools, technologies and research in conversational AI and NLP.

The developer resources listed below are exclusively available to NVIDIA Developer Program members. Join today for free in order to get access to the tools and training necessary to build on NVIDIA’s technology platform here

On-Demand Sessions

Conversational AI Demystified
Speaker: Meriem Bendris, Senior Solution Architect, NVIDIA

Conversational AI technologies are becoming ubiquitous, with countless products taking advantage of automatic speech recognition, natural language understanding, and speech synthesis coming to market. Thanks to new tools and technologies, developing conversational AI applications is easier than ever, enabling a much broader range of applications, such as virtual assistants, real-time transcription, and many more. We will give an overview of the conversational AI landscape and discuss how any organization can get started developing conversational AI applications today.

Building and Deploying a Custom Conversational AI App with NVIDIA Transfer Learning Toolkit and Jarvis
Speakers: Tripti Singhal, Solutions Architect, NVIDIA; Nikhil Srihari, Technical Marketing Engineer – Deep Learning, NVIDIA; Arun Venkatesan, Product Manager, NVIDIA

Tailoring the deep learning models in a conversational AI pipeline to your enterprise needs is time-consuming. Developing a domain-specific application typically requires several cycles of re-training, fine-tuning, and deploying the model until it satisfies the requirements. NVIDIA Jarvis helps you easily build production-ready conversational AI applications and provides tools for fine-tuning on your domain. In this session, we will walk you through the process of customizing automatic speech recognition and natural language processing pipelines to build a truly customized production-ready Conversational AI application.

Megatron GPT-3 Large Model Inference with Triton and ONNX Runtime
Speaker: Denis Timonin, AI Solutions Architect, NVIDIA

Huge NLP models like Megatron-LM GPT-3, Megatron-LM Bert require tens/hundreds of gigabytes of memory to store their weights or run inference. Frequently, one GPU is not enough for such a task. One way to run inference and maximize throughput of these models is to divide them into smaller sub-parts in the pipeline-parallelism (in-depth) style and run these subparts on multiple GPUs. This method will allow us to use bigger batch size and run inference through an ensemble of subparts in a conveyor manner. TRITON inference server is an open-source inference serving software that lets teams deploy trained AI models from any framework. And this is a perfect tool that allows us to run this ensemble. In this talk, we will take Megatron LM with billions of parameters, convert it in ONNX format, and will learn how to divide it into subparts with the new tool – ONNX-GraphSurgeon. Then, we will use TRITON ensemble API and ONNX runtime background and run this model inference on an NVIDIA DGX.

Blog

Announcing Megatron for Training Trillion Parameter Models and NVIDIA Jarvis Availability
NVIDIA announced Megatron for training giant transformer-based language models and major capabilities in NVIDIA Jarvis for building state-of-the-art interactive conversational AI applications.

Demo

World-Class ASR | Real-Time Machine Translation | Controllable Text-to-Speech
Watch this demo to see Jarvis’ automatic speech recognition (ASR) accuracy when fine-tuned on medical jargon, real-time neural machine translation from English to Spanish and Japanese, and powerful controllability of neural text-to-speech.

New pre-trained models, notebooks, and sample applications for conversational AI are all available to try from the NGC catalog. You can also find tutorials for building and deploying conversational AI applications at the NVIDIA Developer Blog.

Join the NVIDIA Developer Program for all of the latest tools and resources for building with NVIDIA technologies.

Categories
Misc

Omniverse – Top Resources from GTC 21

NVIDIA Omniverse is an open platform built for virtual collaboration and real-time physically accurate simulation. Explore the latest resources to learn and get started with Omniverse today.

At GTC 2021 we shared a glimpse of the immense power the NVIDIA Omniverse platform can bring to the world of architecture, manufacturing, product design, robotics, gaming development, and media and entertainment. This new open, cloud-native platform makes virtual collaboration easy for creators, researchers, and engineers on photorealistic rendering projects. For a deeper understanding of the platform and its capabilities, we curated a collection of the latest resources to help you get started on Omniverse.

The developer resources listed below are exclusively available to NVIDIA Developer Program members. Join today for free in order to get access to the tools and training necessary to build on NVIDIA’s technology platform here

On-Demand Sessions

Introduction to RTX Technology and the Omniverse Platform
Speaker: Vincent Brisebois, NVIDIA 

We introduce NVIDIA RTX technology at a very high level (Shaders, RT Cores and Tensor Cores), then introduce the Omniverse platform. We focus on the Omniverse technology stack and give a high-level overview of its components and how developers can leverage them. Our target audience is technical artists or developers who have little-to-no exposure to the platform or creatives, who are currently trying the open beta and looking for a deeper overview of the platform and components.

Panel: Plumbing the Metaverse with USD
Speakers: Dean Takahashi, VentureBeat; F. Sebastian Grassia, Pixar; Guido Quaroni, Adobe; Ivar Dahlberg, Embark Studios; Lori Hufford, Bentley Systems; Martha Tsigkari, Foster+Partners; Mattias Wikenmalm, Volvo; Perry Nightingale, WPP; Susanna Holt, Autodesk Forge 

Learn more about the Pixar USD file format and discover its benefits to digital creators in all verticals. The session will provide a brief overview of USD, followed by a panel of distinguished industry luminaries to discuss their experience and adoption of the format and its benefits.

Introduction to USD
Speaker: Dirk Van Gelder, NVIDIA

This session introduces Universal Scene Description (USD), open-source software from Pixar Animation Studios that’s used as the core representation of the 3D world within NVIDIA’s Omniverse platform. We’ll show what USD is and how assets are constructed with it. We’ll show why this standard and open representation enables interchange with 3D applications to more easily construct virtual worlds. We’ll include hands-on demo examples that illustrate USD scene construction in Python that you can try with a web browser at home, and show how to interact with USD files within Omniverse.

Making a Connector for Omniverse
Speaker: Brian Harrison, NVIDIA

Learn how to connect with the Omniverse platform and be able to send data to it, establish a live sync session, as well as a USD 101 overview to get you started. This is primarily targeted at developers who want to learn how to create a plugin for an application and push data to Omniverse. However, the topic also applies to those seeking to write command line or scripted converters, as well as connecting Omniverse and a data management system. We’ll start with an overview of USD structuring and some of the basics in geometry and materials. A tour of the Sample SDK, which is available on the Omniverse Launcher, will be discussed in detail. From there, we’ll look at how we implemented a Connector for an application, like SketchUp, to discuss design considerations and material mapping and handling.

For more on-demand content, check out the collection of developer sessions from GTC 2021 to learn how industry experts and our very own Omniverse engineers use and build on top of the platform.

Hands-on Demo

Marbles RTX Playable Sample Now Available in NVIDIA Omniverse
Download and discover the iconic RTX demo, and explore real-time physics, dynamic lights and rich, physically based materials in the virtual collaboration platform.

For additional resources, check out the Omniverse Forums, Discord, Twitch, and YouTube channels.

Categories
Misc

HPC – Top Resources from GTC 21

Get the latest resources and news about the NVIDIA technologies that are accelerating the latest innovations in HPC from industry leaders and developers.

Get the latest resources and news about the NVIDIA technologies that are accelerating the latest innovations in HPC from industry leaders and developers. Explore sessions and demos across a variety of HPC topics, ranging from weather forecasting and energy exploration to computational chemistry and molecular dynamics.

The developer resources listed below are exclusively available to NVIDIA Developer Program members. Join today for free in order to get access to the tools and training necessary to build on NVIDIA’s technology platform here

 

On-Demand Sessions

How GPU Computing Works
Speaker: Stephen Jones, CUDA Architect, NVIDIA

Get an introduction to GPU computing by the lead architect of CUDA. We’ll walk through the internals of how the GPU works and why CUDA is the way that it is, and connect the dots between physical hardware and parallel computing.

A Deep Dive into the Latest HPC Software
Speaker: Tim Costa, NVIDIA

Take a deep dive into the latest developments in NVIDIA software for HPC applications, including a comprehensive look at what’s new in programming models, compilers, libraries, and tools. We’ll cover topics of interest to HPC developers, targeting traditional HPC modeling and simulation, HPC+AI, scientific visualization, and high-performance data analytics.

Introducing Developer Tools for Arm and NVIDIA Systems
Speaker: David Owens, Product Director, Infrastructure Software, Arm

Explore the role of key tools and toolchains on Arm servers, from Arm, NVIDIA, and elsewhere — and show how each tool fits in the end-to-end journey to production science and simulation.

 

SDK

Accelerate Quantum Information Science with NVIDIA cuQuantum SDK
NVIDIA cuQuantum is an SDK of optimized libraries and tools for accelerating quantum computing workflows. Learn more about how NVIDIA cuQuantum speeds up quantum circuit simulations by orders of magnitude. 

 

Blog

NVIDIA Arm HPC Developer Kit for HPC, AI, and Scientific Computing Applications
The NVIDIA Arm HPC Developer Kit is an integrated hardware and software platform for creating, evaluating, and benchmarking HPC, AI, and scientific computing applications on a GPU- and CPU-accelerated computing system.

NVIDIA Nsight Visual Studio Code Edition: New Addition to the Nsight Developer Tools Suite
Read about Nsight Visual Studio Code Edition, an application for platforms that bring CUDA development for GPUs into Microsoft Visual Studio Code, including building and debugging GPU kernels and native CPU code.

 

Click here to view all of the HPC sessions and demos on NVIDIA On-Demand.

 

Categories
Misc

Game Developers – New Resources from GTC 21

NVIDIA RTX enables developers to create breathtaking, interactive worlds with performance that exceeds gamers expectations. Integrating RTX has never been easier – gain access through popular game engines such as Unreal Engine or through standalone SDKs made available at GTC.

Developers, engineers, artists and leaders from game studios across the world gathered virtually at this year’s virtual GTC to learn how the latest NVIDIA technologies are revolutionizing game development. 

NVIDIA RTX enables developers to create breathtaking, interactive worlds with performance that exceeds gamers expectations. Integrating RTX has never been easier – gain access through popular game engines such as Unreal Engine or through standalone SDKs.  

The developer resources listed below are exclusively available to NVIDIA Developer Program members. Join today for free in order to get access to the tools and training necessary to build on NVIDIA’s technology platform here

 

On-Demand

The State of RTX in Unreal Engine 4

Speaker: Richard Cowgill, NVIDIA

In this session, NVIDIA’s Richard Cowgill goes over where RTX is today in Unreal Engine 4, all the advancements in the past year, and a quick look at what’s coming up.

Collaborative Game Development using Omniverse

This session is a deep dive on how to leverage Omniverse, using new asset collaboration tools USD and MDL for game development. You’ll learn how to leverage Nucleus for collaboration, AI for asset tagging, and USD and MDL for ground truth content creation and lighting using ray tracing.

Demos

RTX Technology Showcase

Experience the latest NVIDIA RTX technologies available in Unreal Engine 4. Toggle ray tracing on and off between reflections, shadows and translucency to see the impact these features have on  NVIDIA’s Attic demo and what they could bring to your project. You’ll also learn how RTX Global Illumination adds dynamic range to the scene with multi-bounce indirect lighting. Maximize these ray-tracing settings with DLSS, which will boost the frame-rate of the Attic scene while maintaining a high resolution. With an RTX GPU, you can try this demo as a standalone build here.

RTXDI

RTXDI offers realistic lighting of dynamic scenes that require computing shadows from millions of area lights. Until now, this hasn’t been possible in video games in real-time. Traditionally, game developers have baked most lighting and supported a small number of “hero” lights that are computed at runtime. With RTXDI, lighting artists can render scenes with millions of dynamic area lights in real-time without complex computational overheads or disruptive changes to the artist’s workflow. In this scene, you can see neon billboards, brake lights, apartment windows, store displays, and wet roads—all acting as independent light sources. All this can now be captured in real-time with RTXDI. Learn more here.

NVIDIA Reflex

System Latency is the measure of PC responsiveness – a critical metric for gamers that is difficult to optimize for. In this demo, we will show you how NVIDIA Reflex optimizes system latency and helps developers give players a response experience.

In this talk available on NVIDIA On-Demand, Seth Schneider (NVIDIA) and Ryan Greene (Blizzard) will provide a crash course on system latency covering: the basics of system latency, NVIDIA Reflex, and the Overwatch team’s approach to system latency optimization. If you are interested in a deep dive into system level performance, you’ve come to the right place.

Categories
Offsites

Extending Contrastive Learning to the Supervised Setting

In recent years, self-supervised representation learning, which is used in a variety of image and video tasks, has significantly advanced due to the application of contrastive learning. These contrastive learning approaches typically teach a model to pull together the representations of a target image (a.k.a., the “anchor”) and a matching (“positive”) image in embedding space, while also pushing apart the anchor from many non-matching (“negative”) images. Because labels are assumed to be unavailable in self-supervised learning, the positive is often an augmentation of the anchor, and the negatives are chosen to be the other samples from the training minibatch. However, because of this random sampling, false negatives, i.e., negatives generated from samples of the same class as the anchor, can cause a degradation in the representation quality. Furthermore, determining the optimal method to generate positives is still an area of active research.

In contrast to the self-supervised approach, a fully-supervised approach could use labeled data to generate positives from existing same-class examples, providing more variability in pretraining than could typically be achieved by simply augmenting the anchor. However, very little work has been done to successfully apply contrastive learning in the fully-supervised domain.

In “Supervised Contrastive Learning”, presented at NeurIPS 2020, we propose a novel loss function, called SupCon, that bridges the gap between self-supervised learning and fully supervised learning and enables contrastive learning to be applied in the supervised setting. Leveraging labeled data, SupCon encourages normalized embeddings from the same class to be pulled closer together, while embeddings from different classes are pushed apart. This simplifies the process of positive selection, while avoiding potential false negatives. Because it accommodates multiple positives per anchor, this approach results in an improved selection of positive examples that are more varied, while still containing semantically relevant information. SupCon also allows label information to play an active role in representation learning rather than restricting it to be used only in downstream training, as is the case for conventional contrastive learning. To the best of our knowledge, this is the first contrastive loss to consistently perform better on large-scale image classification problems than the common approach of using cross-entropy loss to train the model directly. Importantly, SupCon is straightforward to implement and stable to train, provides consistent improvement to top-1 accuracy for a number of datasets and architectures (including Transformer architectures), and is robust to image corruptions and hyperparameter variations.

Self-supervised (left) vs supervised (right) contrastive losses: The self-supervised contrastive loss contrasts a single positive for each anchor (i.e., an augmented version of the same image) against a set of negatives consisting of the entire remainder of the minibatch. The supervised contrastive loss considered in this paper, however, contrasts the set of all samples from the same class as positives against the negatives from the remainder of the batch.

The Supervised Contrastive Learning Framework
SupCon can be seen as a generalization of both the SimCLR and N-pair losses — the former uses positives generated from the same sample as that of the anchor, and the latter uses positives generated from different samples by exploiting known class labels. The use of many positives and many negatives for each anchor allows SupCon to achieve state-of-the-art performance without the need for hard negative mining (i.e., searching for negatives similar to the anchor), which can be difficult to tune properly.

SupCon subsumes multiple losses from the literature and is a generalization of the SimCLR and N-Pair losses.

This method is structurally similar to those used in self-supervised contrastive learning, with modifications for supervised classification. Given an input batch of data, we first apply data augmentation twice to obtain two copies, or “views,” of each sample in the batch (though one could create and use any number of augmented views). Both copies are forward propagated through an encoder network, and the resulting embedding is then L2-normalized. Following standard practice, the representation is further propagated through an optional projection network to help identify meaningful features. The supervised contrastive loss is computed on the normalized outputs of the projection network. Positives for an anchor consist of the representations originating from the same batch instance as the anchor or from other instances with the same label as the anchor; the negatives are then all remaining instances. To measure performance on downstream tasks, we train a linear classifier on top of the frozen representations.

Cross-entropy, self-supervised contrastive loss and supervised contrastive loss Left: The cross-entropy loss uses labels and a softmax loss to train a classifier. Middle: The self-supervised contrastive loss uses a contrastive loss and data augmentations to learn representations. Right: The supervised contrastive loss also learns representations using a contrastive loss, but uses label information to sample positives in addition to augmentations of the same image.

Key Findings
SupCon consistently boosts top-1 accuracy compared to cross-entropy, margin classifiers (with use of labels), and self-supervised contrastive learning techniques on CIFAR-10 and CIFAR-100 and ImageNet datasets. With SupCon, we achieve excellent top-1 accuracy on the ImageNet dataset with the ResNet-50 and ResNet-200 architectures. On ResNet-200, we achieve a top-1 accuracy of 81.4%, which is a 0.8% improvement over the state-of-the-art cross-entropy loss using the same architecture (which represents a significant advance for ImageNet). We also compared cross-entropy and SupCon on a Transformer-based ViT-B/16 model and found a consistent improvement over cross-entropy (77.8% versus 76% for ImageNet; 92.6% versus 91.6% for CIFAR-10) under the same data augmentation regime (without any higher-resolution fine-tuning).

The SupCon loss consistently outperforms cross-entropy with standard data augmentation strategies (AutoAugment, RandAugment and CutMix). We show top-1 accuracy for ImageNet, on ResNet-50, ResNet-101 and ResNet200.

We also demonstrate analytically that the gradient of our loss function encourages learning from hard positives and hard negatives. The gradient contributions from hard positives/negatives are large while those for easy positives/negatives are small. This implicit property allows the contrastive loss to sidestep the need for explicit hard mining, which is a delicate but critical part of many losses, such as triplet loss. See the supplementary material of our paper for a full derivation.

SupCon is also more robust to natural corruptions, such as noise, blur and JPEG compression. The mean Corruption Error (mCE) measures the average degradation in performance compared to the benchmark ImageNet-C dataset. The SupCon models have lower mCE values across different corruptions compared to cross-entropy models, showing increased robustness.

We show empirically that the SupCon loss is less sensitive than cross-entropy to a range of hyperparameters. Across changes in augmentations, optimizers, and learning rates, we observe significantly lower variance in the output of the contrastive loss. Moreover, applying different batch sizes while holding all other hyperparameters constant results in consistently better top-1 accuracy of SupCon to that of cross-entropy at each batch size.

Accuracy of cross-entropy and supervised contrastive loss as a function of hyperparameters and training data size, measured on ImageNet with a ResNet-50 encoder. Left: Boxplot showing Top-1 accuracy vs changes in augmentation, optimizer and learning rates. SupCon yields more consistent results across variations in each, which is useful when the best strategies are unknown a priori. Right: Top-1 accuracy as a function of batch size shows both losses benefit from larger batch sizes while SupCon has higher Top-1 accuracy, even when trained with small batch sizes.
Accuracy of supervised contrastive loss as a function of training duration and the temperature hyperparameter, measured on ImageNet with a ResNet-50 encoder. Left: Top-1 accuracy as a function of SupCon pre-training epochs. Right: Top-1 accuracy as a function of temperature during the pre-training stage for SupCon. Temperature is an important hyperparameter in contrastive learning and reducing sensitivity to temperature is desirable.

Broader Impact and Next Steps
This work provides a technical advancement in the field of supervised classification. Supervised contrastive learning can improve both the accuracy and robustness of classifiers with minimal complexity. The classic cross-entropy loss can be seen as a special case of SupCon where the views correspond to the images and the learned embeddings in the final linear layer corresponding to the labels. We note that SupCon benefits from large batch sizes, and being able to train the models on smaller batches is an important topic for future research.

Our Github repository includes Tensorflow code to train the models in the paper. Our pre-trained models are also released on TF-Hub.

Acknowledgements
The NeurIPS paper was jointly co-authored with Prannay Khosla, Piotr Teterwak, Chen Wang, Aaron Sarna, Yonglong Tian, Phillip Isola, Aaron Maschinot, Ce Liu, and Dilip Krishnan. Special thanks to Jenny Huang for leading the writing process for this blogpost.