NVIDIA Brings AI Assistants to Life With GeForce RTX AI PCs (2024)

Project G-Assist, NVIDIA ACE NIMs for Digital Humans, and Generative AI Tools Deliver Advanced AI Experiences on RTX Laptops; Plus, RTX-Accelerated APIs for Small Language Models Coming to Windows Copilot Runtime

COMPUTEX—NVIDIA today announced new NVIDIA RTX™ technology to power AI assistants and digital humans running on new GeForce RTX™ AI laptops.

NVIDIA unveiled Project G-Assist — an RTX-powered AI assistant technology demo that provides context-aware help for PC games and apps. The Project G-Assist tech demo debuted with ARK: Survival Ascended from Studio Wildcard. NVIDIA also introduced the first PC-based NVIDIA NIM™ inference microservices for the NVIDIA ACE digital human platform.

These technologies are enabled by the NVIDIA RTX AI Toolkit, a new suite of tools and software development kits that aid developers in optimizing and deploying large generative AI models on Windows PCs. They join NVIDIA’s full-stack RTX AI innovations accelerating over 500 PC applications and games and 200 laptop designs from manufacturers.

In addition, newly announced RTX AI PC laptops from ASUS and MSI feature up to GeForce RTX 4070 GPUs and power-efficient systems-on-a-chip with Windows 11 AI PC capabilities. These Windows 11 AI PCs will receive a free update to Copilot+ PC experiences when available.

“NVIDIA launched the era of AI PCs in 2018 with the release of RTX Tensor Core GPUs and NVIDIA DLSS,” said Jason Paul, vice president of consumer AI at NVIDIA. “Now, with Project G-Assist and NVIDIA ACE, we’re unlocking the next generation of AI-powered experiences for over 100 million RTX AI PC users.”

Project G-Assist, a GeForce AI Assistant
AI assistants are set to transform gaming and in-app experiences — from offering gaming strategies and analyzing multiplayer replays to assisting with complex creative workflows. Project G-Assist is a glimpse into this future.

PC games offer vast universes to explore and intricate mechanics to master, which are challenging and time-consuming feats even for the most dedicated gamers. Project G-Assist aims to put game knowledge at players’ fingertips using generative AI.

Project G-Assist takes voice or text inputs from the player, along with contextual information from the game screen, and runs the data through AI vision models. These models enhance the contextual awareness and app-specific understanding of a large language model (LLM) linked to a game knowledge database, and then generate a tailored response delivered as text or speech.

NVIDIA partnered with Studio Wildcard to demo the technology with ARK: Survival Ascended. Project G-Assist can help answer questions about creatures, items, lore, objectives, difficult bosses and more. Because Project G-Assist is context-aware, it personalizes its responses to the player’s game session.

In addition, Project G-Assist can configure the player’s gaming system for optimal performance and efficiency. It can provide insights into performance metrics, optimize graphics settings depending on the user’s hardware, apply a safe overclock and even intelligently reduce power consumption while maintaining a performance target.

First ACE PC NIM Debuts
NVIDIA ACE technology for powering digital humans is now coming to RTX AI PCs and workstations with NVIDIA NIM — inference microservices that enable developers to reduce deployment times from weeks to minutes. ACE NIM microservices deliver high-quality inference running locally on devices for natural language understanding, speech synthesis, facial animation and more.

At COMPUTEX, the gaming debut of NVIDIA ACE NIM on the PC will be featured in the Covert Protocol tech demo, developed in collaboration with Inworld AI. It now showcases NVIDIA Audio2Face™ and NVIDIA Riva automatic speech recognition running locally on devices.

Windows Copilot Runtime to Add GPU Acceleration for Local PC SLMs
Microsoft and NVIDIA are collaborating to help developers bring new generative AI capabilities to their Windows native and web apps. This collaboration will provide application developers with easy application programming interface (API) access to GPU-accelerated small language models (SLMs) that enable retrieval-augmented generation (RAG) capabilities that run on-device as part of Windows Copilot Runtime.

SLMs provide tremendous possibilities for Windows developers, including content summarization, content generation and task automation. RAG capabilities augment SLMs by giving the AI models access to domain-specific information not well represented in ‌base models. RAG APIs enable developers to harness application-specific data sources and tune SLM behavior and capabilities to application needs.

These AI capabilities will be accelerated by NVIDIA RTX GPUs, as well as AI accelerators from other hardware vendors, providing end users with fast, responsive AI experiences across the breadth of the Windows ecosystem.

The API will be released in developer preview later this year.

4x Faster, 3x Smaller Models With the RTX AI Toolkit
The AI ecosystem has built hundreds of thousands of open-source models for app developers to leverage, but most models are pretrained for general purposes and built to run in a data center.

To help developers build application-specific AI models that run on PCs, NVIDIA is introducing RTX AI Toolkit — a suite of tools and SDKs for model customization, optimization and deployment on RTX AI PCs. RTX AI Toolkit will be available later this month for broader developer access.

Developers can customize a pretrained model with open-source QLoRa tools. Then, they can use the NVIDIA TensorRT™ model optimizer to quantize models to consume up to 3x less RAM. NVIDIA TensorRT Cloud then optimizes the model for peak performance across the RTX GPU lineups. The result is up to 4x faster performance compared with the pretrained model.

The newNVIDIA AI Inference ManagerSDK,now available in early access, simplifies the deployment of ACE to PCs. It preconfigures the PC with the necessary AI models, engines and dependencies while orchestrating AI inference seamlessly across PCs and the cloud.

Software partners such as Adobe, Blackmagic Design and Topaz are integrating components of the RTX AI Toolkit within their popular creative apps to accelerate AI performance on RTX PCs.

“Adobe and NVIDIA continue to collaborate to deliver breakthrough customer experiences across all creative workflows, from video to imaging, design, 3D and beyond,” said Deepa Subramaniam, vice president of product marketing, Creative Cloud at Adobe. “TensorRT 10.0 on RTX PCs delivers unprecedented performance and AI-powered capabilities for creators, designers and developers, unlocking new creative possibilities for content creation in industry-leading creative tools like Photoshop.”

Components of the RTX AI Toolkit, such as TensorRT-LLM, are integrated in popular developer frameworks and applications for generative AI, including Automatic1111, ComfyUI, Jan.AI, LangChain, LlamaIndex, Oobabooga and Sanctum.AI.

AI for Content Creation
NVIDIA is also integrating RTX AI acceleration into apps for creators, modders and video enthusiasts.

Last year, NVIDIA introduced RTX acceleration using TensorRT for one of the most popular Stable Diffusion user interfaces, Automatic1111. Starting this week, RTX will also accelerate the highly popular ComfyUI, delivering up to a 60% improvement in performance over the currently shipping version, and 7x faster performance compared with the MacBook Pro M3 Max.

NVIDIA RTX Remix is a modding platform for remastering classic DirectX 8 and DirectX 9 games with full ray tracing, NVIDIA DLSS 3.5 and physically accurate materials. RTX Remix includes a runtime renderer and the RTX Remix Toolkit app, which facilitates the modding of game assets and materials.

Last year, NVIDIA made RTX Remix Runtime open source, allowing modders to expand game compatibility and advance rendering capabilities.

Since RTX Remix Toolkit launched earlier this year, 20,000 modders have used it to mod classic games, resulting in over 100 RTX remasters in development on the RTX Remix Showcase Discord.

This month, NVIDIA will make the RTX Remix Toolkit open source, allowing modders to streamline how assets are replaced and scenes are relit, increase supported file formats for RTX Remix’s asset ingestor and bolster RTX Remix’s AI Texture Tools with new models.

In addition, NVIDIA is making the capabilities of RTX Remix Toolkit accessible via a REST API, allowing modders to livelink RTX Remix to digital content creation tools such as Blender, modding tools such as Hammer and generative AI apps such as ComfyUI. NVIDIA is also providing an SDK for RTX Remix Runtime to allow modders to deploy RTX Remix’s renderer into other applications and games beyond DirectX 8 and 9 classics.

With more of the RTX Remix platform being made open source, modders across the globe can build even more stunning RTX remasters.

NVIDIA RTX Video, the popular AI-powered super-resolution feature supported in the Google Chrome, Microsoft Edge and Mozilla Firefox browsers, is now available as an SDK to all developers, helping them natively integrate AI for upscaling, sharpening, compression artifact reduction and high-dynamic range (HDR) conversion.

Coming soon to video editing software Blackmagic Design’s DaVinci Resolve and Wondershare Filmora, RTX Video will enable video editors to upscale lower-quality video files to 4K resolution, as well as convert standard dynamic range source files into HDR. In addition, the free media player VLC media will soon add RTX Video HDR to its existing super-resolution capability.

Learn more about RTX AI PCs and technology by joining NVIDIA at COMPUTEX.

NVIDIA Brings AI Assistants to Life With GeForce RTX AI PCs (2024)

FAQs

NVIDIA Brings AI Assistants to Life With GeForce RTX AI PCs? ›

NVIDIA announced new NVIDIA RTX technology to power AI assistants and digital humans running on new GeForce RTX AI laptops. At Computex 2024, NVIDIA unveiled Project G-Assist — an RTX-powered AI assistant technology demo that provides context-aware help for PC games and apps.

Does the Nvidia RTX use AI? ›

That's because the same technology powering world-leading AI innovation is built into every RTX GPU, giving you the power to do the extraordinary. From class to work to entertainment, with RTX-powered AI, you're getting the most advanced AI experiences available on Windows PCs.

Why is NVIDIA needed for AI? ›

The Nvidia ecosystem, from its software to its sourcing of materials, allowed it to position itself as the go-to source for companies that needed massive computing power to handle their AI needs.

What is the NVIDIA AI platform? ›

NVIDIA AI Enterprise is an end-to-end, cloud-native software platform that accelerates data science pipelines and streamlines development and deployment of production-grade co-pilots and other generative AI applications.

What is the role of NVIDIA in generative AI? ›

NVIDIA AI is the world's most advanced platform for generative AI and is relied on by organizations at the forefront of innovation. Designed for the enterprise and continuously updated, the platform lets you confidently deploy generative AI applications into production, at scale, anywhere.

Is Nvidia AI free? ›

Kick-start your AI journey with access to NVIDIA AI workflows—for free.

Does ChatGPT use Nvidia? ›

The company's graphics processing units (GPUs) and other chips have experienced skyrocketing demand because those products are especially well suited to make demanding applications like ChatGPT run, and ChatGPT's own success has showcased the power of Nvidia's GPUs.

How much does the Nvidia AI chip cost? ›

NVIDIA CEO Jensen Huang, right after GTC 2024, confirmed that the company's cutting-edge AI chip, Blackwell B200, might cost anywhere between $30,000 and $40,000 per unit.

How does Nvidia make money from AI? ›

Key Data Points

Given that Nvidia relies on sales of chips that are being deployed in data centers for AI training and inference purposes, it can be easily concluded that this business segment can continue to be a major catalyst for the company.

Why did Apple stop using Nvidia? ›

Why did Apple stop using Nvidia graphics? Apple used to work closely with NVIDIA, until 2008. In 2007 and 2008, Apple made Macbook Pro models with the ill-fated NVIDIA Geforce 8400M and 9400M. These Geforce chips had a manufacturing defect in the chip die that caused the GPUs to fail prematurely.

What language does Nvidia AI use? ›

In conclusion, Nvidia supports several programming languages for AI development, including Python, C++, Java, and MATLAB. These languages are widely used in the computer science community and are known for their high performance and efficiency.

Who owns Nvidia? ›

Nvidia (NVDA) Ownership Overview

The ownership structure of Nvidia (NVDA) stock is a mix of institutional, retail and individual investors. Approximately 14.74% of the company's stock is owned by Institutional Investors, 3.76% is owned by Insiders and 81.50% is owned by Public Companies and Individual Investors.

Will Nvidia dominate AI? ›

Nvidia dominates the AI chip market, but there's more competition than ever. Nvidia's AI accelerators have between 70% and 95% of the market share for artificial intelligence chips. But there's more competition than ever as startups, cloud companies and other chipmakers ramp up development.

What is the main purpose of Nvidia? ›

NVIDIA engineers the most advanced chips, systems, and software for the AI factories of the future. We build new AI services that help companies create their own AI factories.

What is one thing current generative AI applications cannot do? ›

Generative AI can't generate new ideas or solutions

One of the key limitations of AI is its inability to generate new ideas or solutions.

Which Nvidia graphics card for AI? ›

For demanding tasks requiring high performance, the Nvidia A100 is the best choice. For medium-scale tasks, the RTX A6000 offers a good balance of performance and cost. The RTX 4090 is a suitable option for smaller-scale tasks or hobbyists.

Are GPUs used for AI? ›

GPU architecture offers unmatched computational speed and efficiency, making it the backbone of many AI advancements. The foundational support of GPU architecture allows AI to tackle complex algorithms and vast datasets, accelerating the pace of innovation and enabling more sophisticated, real-time applications.

Is NVDA involved in AI? ›

NVIDIA's Position in the AI Industry

The company also directs ample investment to the development of AI software products such as its Advanced AI platform for Enterprise that offers generative AI, data analytics, and inference.

How to get RTX AI? ›

All you need to do is run an installer, but the installer is prone to fail, and you'll need to satisfy some minimum system requirements. You need an RTX 40-series or 30-series GPU with at least 8GB of VRAM, along with 16GB of system RAM, 100GB of disk space, and Windows 11.

Top Articles
Latest Posts
Article information

Author: Tuan Roob DDS

Last Updated:

Views: 5828

Rating: 4.1 / 5 (62 voted)

Reviews: 85% of readers found this page helpful

Author information

Name: Tuan Roob DDS

Birthday: 1999-11-20

Address: Suite 592 642 Pfannerstill Island, South Keila, LA 74970-3076

Phone: +9617721773649

Job: Marketing Producer

Hobby: Skydiving, Flag Football, Knitting, Running, Lego building, Hunting, Juggling

Introduction: My name is Tuan Roob DDS, I am a friendly, good, energetic, faithful, fantastic, gentle, enchanting person who loves writing and wants to share my knowledge and understanding with you.