Daily Tech News, Interviews, Reviews and Updates

NVIDIA officially introduces new NVIDIA In- Game Inferencing (NVIGI) SDK for developers; integrates AI inference directly into C++ games and applications

NVIDIA has today officially introduced NVIDIA In-Game Inferencing (NVIGI) SDK- a unified interface for developers to orchestrate the deployment of AI models from the cloud to local PC execution environments.

NVIGI SDK enables you to integrate AI inference directly into C++ games and applications for optimal performance and latency. Developers can bring NVIDIA ACE AI Characters to games with NVIGI SDK.

NVIDIA In-Game Inferencing SDK

This is a GPU-optimized, plugin-based inference manager designed to simplify the integration of ACE models into gaming and interactive applications. It offers the following functionality-

  • Plugin flexibility: Add, update, and manage AI plugins (ASR, language models, embeddings) with ease.
  • Windows-native DLLs: Streamline workflows for C++ games and applications.
  • GPU optimization: Use compute-in-graphics (CIG) technology for efficient AI inference alongside rendering tasks.

By combining NVIGI with ACE, you can create autonomous characters with advanced generative AI capabilities, such as real-time NPC dialogue, contextual memory, and lifelike animation. NVIGIA architecture enables flexible integration of various AI functionalities- Core plugins, helper plugins, Unified APIs, and local and cloud execution. It orchestrates an agentic workflow, enabling characters to listen, reason, speak, and animate in real time.

To get started with ACE on-device inference, users need to initialize NVIGI, load plugins, and models, create runtime configuration and inference, enable GPU scheduling and rendering integration, and run inference. To learn about the detailed process, click on this link.

List of available NVIGI plugins

NVIGI Plugin Supported Inference Hardware Supported Models
Speech – ASR Local GGML CUDA-Enabled GPU or CPU Whisper ASR
Speech – ASR Local TRT CUDA-Enabled GPU NVIDIA RIVA ASR (coming soon)
Language – GPT Local ONNX DML ONNX-supported GPU or CPU Mistral-7B-Instruct
Language- GPT Local GGML CUDA-Enabled GPU or CPU Llama-3.2-3b Instruct ,Nemotron-Mini-4B-Instruct, Mistral-Nemo-Minitron-2B-128k-Instruct , Mistral-Nemo-Minitron-4B-128k-Instruct, Mistral-Nemo-Minitron-8B-128k-Instruct, Nemovision-4B-Instruct
RAG – Embed Local GGML CUDA-Enabled GPU or CPU E5 Large Unsupervised

 

Get real time updates directly on you device, subscribe now.



You might also like