Doom has been ported to dozens of devices, but it’s never been playable quite like this.
Google researchers have now generated an AI version of the retro first-person shooter classic entirely via neural network, based on ingested video clips of gameplay.
It’s a milestone, if a grim one, recorded in a paper published this week entitled “Diffusion models are real-time game engines” (thanks, VentureBeat). This documents how a small team from Google were able to “interactively simulate” a version of Doom, with only a “slightly better than random chance” of humans being able to tell the difference.
Humans are still (currently) required to play Doom first, to provide the video clips of gameplay that are then fed into GameNGen, the research team’s game engine which is “powered entirely by a neural model”. It’s the same principle as the now-commonplace ability for AI to learn from and then generate static images, based on ingesting huge amouts of dubiously-sourced data online.
GameNGen then produces sequential frames based on its learnings of ‘watching’ that gameplay, which are then output at 20fps with a visual quality “comparable to the original”. Here…