Kun Wang

Neural Computer: Toward a Learned Runtime

April 11, 2026

Much of the recent progress in AI systems can be described in fairly concrete terms: models have become better at using computers. They call tools, operate interfaces, write files, browse documents, and carry out increasingly long workflows. From that trajectory, it is natural to assume that the next step is simply a stronger agent.

The recent Neural Computer (NC) proposal points in a different direction.

What makes the concept interesting is that it does not focus on better task orchestration alone. Instead, it asks whether some responsibilities that normally belong to the machine’s own runtime can be shifted into a learned runtime state. In the paper’s framing, the goal is to unify computation, memory, and I/O inside that learned state, rather than treating the model as a layer that merely sits on top of an unchanged software stack.

That makes NC less about “using a computer well” and more about reconsidering what part of the computer remains external to the model in the first place.

1. Why This Question Is Appearing Now

The Neural Computer idea makes more sense when placed against the broader direction of AI systems research.

On one side, agents have become increasingly practical. Models are no longer confined to text generation; they operate terminals, browsers, desktop environments, and application workflows. They are moving closer to the execution layer of real work.

On another side, world models have become increasingly ambitious. They are used not only for representation, but also for rollout, prediction, simulation, planning, and control. The emphasis has shifted from static understanding to state evolution over time.

At the same time, the classical software stack can feel structurally awkward in AI-native settings. It remains far more reliable than any learned alternative, but for language-driven, weakly specified, long-horizon tasks, the separation between model, control logic, program, and runtime is not always the most natural abstraction.

Seen in that context, NC is best read as a systems question: can part of what is currently distributed across program logic, control layers, and runtime behavior be pulled back into a learned substrate?

2. What “Neural Computer” Actually Means

The defining phrase in the proposal is learned runtime state.

That phrase matters because it shifts attention away from single-task capability and toward the machine’s mode of operation over time. The issue is no longer just whether a model can complete a task once. The issue becomes whether the system can maintain a state that behaves more like a runtime: something persistent, updateable, interactive, and capable of retaining structure across repeated use.

This changes the set of questions that matter:

  • What state persists across interaction?
  • Which inputs genuinely modify the machine, rather than merely condition one response?
  • What counts as adaptation, and what counts as reprogramming?
  • Can capabilities remain inside the system rather than being reconstructed from prompts and tool calls each time?

In that sense, NC is not simply about model capability. It is about whether capability can become runtime-resident.

3. Not NTM/DNC, Not Just an Agent, Not Just a World Model

The term “Neural Computer” immediately invites comparison with older memory-augmented neural architectures such as Neural Turing Machines (NTM) and Differentiable Neural Computers (DNC). That comparison is useful up to a point, but it misses the center of the current proposal.

NTM and DNC were primarily concerned with giving neural networks differentiable access to external memory. The new NC framing is broader and more structural. The question is not simply how a model stores or retrieves information, but whether some part of machine operation itself can be internalized.

It also helps to distinguish NC from adjacent concepts.

Conventional Computer

A conventional computer is organized around an explicit program. The program is authored outside the machine state, and the system’s job is to execute it with stable semantics.

Agent

An agent is organized around a task. It assumes the surrounding computer already exists, then plans and acts within that environment to accomplish a goal.

World Model

A world model is organized around state dynamics. It learns how an environment changes over time and uses that understanding for prediction, simulation, or control.

Neural Computer

A neural computer is organized around the runtime itself. The proposal asks whether execution, state transition, capability retention, and controlled updating can live inside a single learned substrate.

A simple way to frame the difference is this:

  • with a conventional computer, software is installed
  • with an agent, a task is specified
  • with a neural computer, the ambition is to install capability into the machine itself

That last step is what makes the idea unusual.

4. What the Current Prototypes Actually Attempt

One of the more useful aspects of the current work is that it starts from a limited question rather than claiming to have solved the full problem.

The immediate question is: if a model only observes I/O traces, without access to the hidden internal state of the underlying program, can it still learn early runtime primitives?

To explore that, the proposal instantiates NC through interface environments and video-model-like rollouts.

CLIGen

CLIGen focuses on terminal and command-line environments. The model observes visible interface behavior: typed commands, echoed text, cursor motion, scrolling, TUI layout, and screen transitions.

This setting is more informative than it may look at first. A terminal is highly discrete, text-dense, and sensitive to timing and layout. Coherent behavior in such an environment suggests that the model is learning more than surface appearance; it is beginning to model some local structure of interaction.

GUIWorld

GUIWorld extends the same idea to desktop GUI environments. Here the challenge is not simply visual plausibility, but action-conditioned interface evolution: clicks causing windows to open, typing updating visible fields, pages changing after interaction, and so on.

The point is not that these systems already constitute real runtimes. The point is that they begin to test whether user action and interface state can be coupled inside a learned system in ways that resemble runtime behavior.

5. What These Results Seem to Establish

The current prototypes do not establish a new computational machine in the strong sense. What they do seem to establish is narrower and still meaningful: a learned system can begin to acquire early interface primitives from I/O traces alone.

The paper and project materials emphasize two early signs in particular:

  • I/O alignment, where user input and visible interface evolution line up in meaningful ways
  • short-horizon control, where brief action-feedback loops can be maintained with some coherence

That is a real result, but it should not be overstated. These are still local behaviors. The harder question is whether such behaviors can scale into something that deserves machine-level trust.

6. The Hard Problems Are Still the Main Problems

The authors are quite explicit about what remains unresolved, and those unresolved parts are not secondary.

Routine Reuse

A system may produce a convincing behavior once, yet fail to invoke the same procedure reliably later. Reuse is essential if capabilities are to become part of the machine rather than transient local responses.

Controlled Updates

If the machine changes through interaction, it becomes necessary to distinguish between ordinary updates and destructive drift. A runtime that cannot be updated without destabilizing prior behavior is not much of a runtime.

Symbolic Stability

This is arguably the core difficulty. Computation depends not only on responsiveness, but on discrete consistency over long horizons. The ability to preserve stable symbolic relations, repeat behavior, and support verification is what separates a machine from a convincing simulation of one.

These three problems together define the gap between runtime-like behavior and runtime reliability.

7. The Endgame: Completely Neural Computer

The proposal uses the term Completely Neural Computer (CNC) for the mature form of the idea.

That target implies more than stronger interactive behavior. It suggests a machine form with persistent operation, reusable capabilities, explicit reprogramming boundaries, and some kind of machine-native computational semantics. In other words, the endgame is not just a model that can imitate software well, but a system that can function as a programmable substrate in its own right.

This is why NC is conceptually larger than a new product category. It implicitly questions part of the classical boundary between the learned model and the surrounding computational stack.

8. Why the Idea Matters Even in an Early Form

The immediate value of NC is not that it already works as a replacement for conventional software. It does not. The value is that it sharpens a question that is otherwise easy to miss.

A great deal of current AI development assumes that the model becomes smarter while the machine around it stays conceptually fixed. Better reasoning, better tool use, better memory, better orchestration — but still the same basic division of labor between model and runtime.

NC introduces a different possibility: some of the structure currently living outside the model may eventually move inward.

If that happens, a number of familiar distinctions begin to shift:

  • software installation may partially become capability installation
  • prompting may begin to overlap with programming
  • the application layer may blur into the runtime layer
  • interaction may move from “an assistant using a computer” toward “a machine instantiated through the model itself”

Whether that transition will actually happen remains unclear. But NC offers a useful vocabulary for thinking about it.

Closing Thoughts

Neural Computer is still far from a settled architecture. At the moment, it is better understood as a serious research direction than as an established machine form.

Still, the proposal matters because it reframes the problem at a deeper level. The question is no longer only how to make models more capable within the existing stack. The question becomes whether part of the stack itself can become learned, persistent, and machine-like.

That is a more demanding question than most of the current language around agents or AI operating systems. It is also a more interesting one.

If the long-term trajectory of AI involves not only better models using computers, but systems in which some part of the computer has effectively become the model, then NC may turn out to be an early name for that shift.

References

You control you. Do the next right thing.