The new DLSS 4 Multi Frame Generation feature of the new RTX Blackwell cards has created a situation where one frame can be generated using traditional GPU computation, while the subsequent three frames can now be entirely generated by AI. That's a hell of an imbalance, so does one of the people responsible for making this AI voodoo a reality think we'll get to a point where there are no traditionally rendered frames? Jen-Hsun Huang, Nvidia's CEO, and one of the biggest proponents of AI in just about damned near everything, says: no. During a Q&A session today at CES 2025 Jen-Hsun was asked the question about whether we're likely to get purely AI generated game frames as the entirety of a game pipeline and he was unequivocal in his assertion to the contrary, stating that it's vital for AI to be given grounding, to be given context in order to build out its world.

In other words, AI still needs something to build from. In gaming terms Huang suggests that it works in the same way as we give context to ChatGPT. "The context could be a PDF, it could be a web search.

.. and the context in video games has to not only be relevant story-wise, but it has to be world and spatially relevant.

And the way you condition, the way you give it context is you give it early pieces of geometry, or early pieces of textures it could up-res from." He then brings it back around to DLSS 4 and Multi Frame Generation, and the example of one rendered 4K frame and three further 4K game frames. "Out of 33 .