From Von Neumann to Neural Latent Stack: A Decade-Long Paradigm Shift
The Neural Computers paper represents the culmination of a research trajectory that began over a decade ago with Neural Turing Machines in 2014. Those early systems coupled neural networks with external memory banks — the network learned to read and write from a separate, structured memory using attention mechanisms. Differentiable Neural Computers in 2016 refined this approach with more sophisticated memory addressing. But both designs preserved the fundamental Von Neumann separation: the neural network was the processor, and memory remained an external resource.
Neural Computers break with this lineage in a radical way. Rather than giving a neural network access to separate memory and I/O systems, the entire computing stack collapses into a single learned latent state. As ArXivIQ described it, this is 'a fundamental shift from the traditional Von Neumann hardware/software stack to a unified neural latent stack.' The system does not execute instructions fetched from memory in the classical sense. Instead, built on a diffusion transformer architecture (Wan2.1), it generates successive screen frames — rolling out visual computation step by step. The implications are architectural: there is no operating system, no file system, no instruction set architecture. The model learns what all of those things should do from data.
This shift also reframes the relationship between agents and computers. 36Kr's analysis maps three eras: conventional Human-to-Computer interaction, the current Human-to-Agent-to-Computer paradigm where AI mediates between users and traditional software, and the proposed Human-to-Neural Computer relationship where the intermediary disappears because the AI is the machine itself. The paper identifies three converging trends making this plausible now: agents improving at real work (citing MetaGPT, Cursor, Claude Code), world models advancing in environment simulation (GameNGen, Genie 2/3, Waymo), and the structural friction of conventional computers when handling open-ended, long-horizon tasks.
