LAS VEGAS, Jan. 06, 2025 (GLOBE NEWSWIRE) — CES — NVIDIA made a groundbreaking announcement today at CES, unveiling foundation models that run locally on NVIDIA RTX™ AI PCs, significantly enhancing digital humans, content creation, productivity, and development. These models, known as NVIDIA NIM™ microservices, are powered by the latest GeForce RTX™ 50 Series GPUs, offering up to 3,352 trillion operations per second of AI performance and 32GB of VRAM. The GPUs are built on the innovative NVIDIA Blackwell architecture, making them the first consumer GPUs to support FP4 compute, which boosts AI inference performance by 2x and enables generative AI models to run locally in a smaller memory footprint compared to previous generations.
GeForce™ has historically been a pivotal platform for AI developers, with the first GPU-accelerated deep learning network, AlexNet, trained on the GeForce GTX™ 580 in 2012. Over 30% of published AI research papers last year cited the use of GeForce RTX, highlighting its importance in the AI community. Now, with the introduction of generative AI and RTX AI PCs, NVIDIA is democratizing AI development, allowing anyone to become a developer. Low-code and no-code tools such as AnythingLLM, ComfyUI, Langflow, and LM Studio enable enthusiasts to leverage AI models in complex workflows through intuitive graphical user interfaces.
NIM microservices connected to these GUIs make it easy to access and deploy the latest generative AI models. NVIDIA AI Blueprints, built on NIM microservices, provide preconfigured reference workflows for digital humans, content creation, and more, simplifying the development process.
To meet the increasing demand from AI developers and enthusiasts, every major PC manufacturer and system builder is launching NIM-ready RTX AI PCs featuring GeForce RTX 50 Series GPUs.
Jensen Huang, founder and CEO of NVIDIA, commented on the rapid advancement of AI, stating, “AI is advancing at light speed, from perception AI to generative AI and now agentic AI. NIM microservices and AI Blueprints give PC developers and enthusiasts the building blocks to explore the magic of AI.”
Foundation models, which are neural networks trained on vast amounts of raw data, are essential for generative AI. NVIDIA will release a series of NIM microservices for RTX AI PCs from top model developers like Black Forest Labs, Meta, Mistral, and Stability AI, covering a wide range of use cases such as large language models, vision language models, image generation, speech, and more.
NVIDIA also introduced the Llama Nemotron family of open models, offering high accuracy on various agentic tasks. The Llama Nemotron Nano model will be available as a NIM microservice for RTX AI PCs and excels at tasks like instruction following, function calling, chat, coding, and math.
NIM microservices include essential components for running AI on PCs and are optimized for deployment across NVIDIA GPUs, whether in RTX PCs, workstations, or the cloud. Developers and enthusiasts will be able to easily download, set up, and run these NIM microservices on Windows 11 PCs using Windows Subsystem for Linux (WSL).
Microsoft’s corporate vice president of Windows, Pavan Davuluri, highlighted the collaboration with NVIDIA, stating, “AI is driving Windows 11 PC innovation at a rapid rate, and Windows Subsystem for Linux (WSL) offers a great cross-platform environment for AI development on Windows 11 alongside Windows Copilot Runtime. NVIDIA NIM microservices, optimized for Windows PCs, give developers and enthusiasts ready-to-integrate AI models for their Windows apps, further accelerating deployment of AI capabilities to Windows users.”
The NIM microservices will be compatible with top AI development and agent frameworks, including AI Toolkit for VSCode, AnythingLLM, ComfyUI, CrewAI, Flowise AI, LangChain, Langflow, and LM Studio. Enthusiasts will have the opportunity to experience a range of NIM microservices using the upcoming NVIDIA ChatRTX tech demo.
In a demonstration of how NIM can be used to build AI agents and assistants, NVIDIA previewed Project R2X, a vision-enabled PC avatar capable of assisting users with various tasks including desktop applications, video conference calls, document reading and summarization, and more. The avatar is rendered using NVIDIA RTX Neural Faces and animated by a new diffusion-based NVIDIA Audio2Face™-3D model.
NVIDIA also showcased AI Blueprints, reference AI workflows that can run locally on RTX PCs, enabling developers to create podcasts from PDF documents, generate images guided by 3D scenes, and more. The blueprints leverage NIM microservices like Mistral-Nemo-12B-Instruct for language, NVIDIA Riva for text-to-speech, and the NeMo Retriever collection for PDF extraction.
NVIDIA NIM microservices and AI Blueprints will be available starting in February, initially supporting GeForce RTX 50 Series, GeForce RTX 4090 and 4080, and NVIDIA RTX 6000 and 5000 professional GPUs, with additional GPU support planned for the future. NIM-ready RTX AI PCs will be offered by leading manufacturers and system builders, providing users with access to cutting-edge AI capabilities.
In conclusion, NVIDIA’s latest innovations in AI technology are set to revolutionize the way developers and enthusiasts leverage AI models for a wide range of applications. The introduction of NIM microservices, AI Blueprints, and NIM-ready RTX AI PCs marks a significant step towards accelerating generative AI and making AI development more accessible to a broader audience.