Home Gaming GTC 2026: NVIDIA Cuts VRAM From 6.5 GB to 970 MB Thanks...

GTC 2026: NVIDIA Cuts VRAM From 6.5 GB to 970 MB Thanks to Neural Texture Compression

7
0

Neural Rendering Takes Center Stage at GTC 2026

NVIDIA’s latest developments in neural rendering emphasize the integration of neural networks into the graphics pipeline, resulting in significant gains in VRAM and rendering time. The focus is on DLSS 5 message and optimizing assets and shading. Unlike DLSS 5, which raised concerns about neural rendering, the presentation at GDC is difficult to refute.

Pragmatic Neural Rendering: Beyond DLSS 5

NVIDIA is now placing neural rendering within the pipeline itself, moving away from post-processing alone. This approach relies on specialized neural networks for decoding textures, evaluating materials, and reducing memory usage, rather than a global final filter.

The GDC session highlighted three distinct approaches: ML post-processing, where DLSS 5 is a prime example, improves the final image output. ML integrated into the pipeline directly affects shaders, textures, and materials. The generative approach partially or fully generates the image using a network. None is superior to the others; they are combinable, as demonstrated by NVIDIA with the first two.

GTC 2026: NVIDIA Cuts VRAM From 6.5 GB to 970 MB Thanks to Neural Texture Compression

DLSS 5 is just one aspect, visible at the final image level. The GDC session detailed dedicated blocks for targeted tasks, with immediate benefits for memory footprint, bandwidth, and material evaluation speed, relevant for both gaming and real-time production.

Neural Texture Compression: From 6.5GB to 970MB

Comparison of Texture Compression

An illustrative example is Neural Texture Compression (NTC). On the Tuscan Wheels scene, NVIDIA announced a reduction in VRAM usage from about 6.5GB with traditional BCN textures to 970MB with NTC, maintaining perceived quality close to the original. In the same 970MB budget, NTC preserves more details than traditional block compression.

Explanation of NTC

NTC encodes a compact latent representation of the texture instead of directly storing texels. During rendering, a small neural network reconstructs the pixels on-the-fly on the GPU. The process is entirely deterministic, with no random generation or hallucination. It’s pure compression with learned reconstruction.

Beyond VRAM, the interest is logistical: lighter installations, smaller patches, less network bandwidth, and more room for detailed assets on the same GPU. For developers, this creates space to densify content without exceeding memory budgets. Alexi, a technical speaker at the session, sees NTC as already usable in production today.

Neural Materials Schema

Neural Materials: Fewer Channels, Faster 1080p Rendering

Neural Materials Schema

Another component, Neural Materials, compresses material behavior into a latent representation and decodes it through a small network. In the demo, a 19-channel configuration is reduced to eight channels, with renders in 1080p announced to be 1.4x to 7.7x faster on the test scene, with reduced BRDF complexity and streamlined storage.

What Changes in Practice

Consider a complex material: ceramic base, metallic sheen, varnish, dust. Traditionally, each layer involves separate texture and BRDF evaluations. With Neural Materials, all this is encoded into a single latent vector, decoded in a single inference at render time. The result: less memory, less bandwidth, and faster evaluation.

Realistic Material Details

The goal is not to alter artistic direction but to retain material information in a compact form and limit evaluation costs. Neural rendering acts as a shading accelerator and asset compressor, not a final image generator. NVIDIA acknowledges that Neural Materials is still an actively researched area, with production deployment still limited compared to NTC.

Realistic Statue Render

Slang and SlangPy: The Infrastructure Making It All Possible

Real-Time Graphics with Neural Rendering

These technologies rely on a common infrastructure. NVIDIA introduced Slang, a differentiable shader language that automatically generates the necessary backward pass for training and compiles to CUDA, HLSL, and GLSL. Additionally, SlangPy bridges Python and the GPU, unifying training and production deployment phases.

SlangPy

This addresses the historical challenge of real-time neural rendering: the mismatch between PyTorch training code and C++ or shader runtime code. With Slang, the same code serves both phases. For developers looking to integrate these neural blocks into their pipeline, this is the missing piece.

Neural Fixer and Asset Harvester: Neural Rendering Serving Simulation

The session also covered a less expected use case: simulation for autonomous vehicles via Omniverse Neurec. The Gaussian Splatting underpins a 3D reconstruction from real images, where the scene is represented by 3D Gaussian particles, each with position, rotation, opacity, and view-dependent color.

Neural Fixer

Fixer

The classic Gaussian Splatting issue leads to artifacts when deviating from the captured trajectory. The Neural Fixer is a network that corrects these artifacts in real-time by injecting real-world knowledge, making the simulation usable for scenarios never physically captured.

Asset Harvester

Asset Harvester

Another tool presented, the Asset Harvester, extracts partially visible objects in captured data, for example, a bus seen from a single angle, and generates a complete and usable 3D version. These assets can then be placed in various simulated scenarios, amplifying the value of the initial data collection.

A Less Frightening Neural Rendering

The DLSS 5 debate highlighted concerns about visual homogenization and AI hallucination over approximate rendering. The roadmap shown at GDC suggests a more consensual path to production: offloading VRAM, reducing bandwidth, and accelerating critical passes without distorting the image.

For studios, these targeted neural blocks can be iteratively integrated into the pipeline, improving the quality-cost ratio without complete overhauls. NTC is available today via the SDK RTX Neural Texture Compression. Neural Materials and simulation tools will follow the pace of research maturity. It’s neural rendering without the show, and that’s likely where it will succeed.

Source: