Back to Blog List
Back to Blog List

When AI learns physics: why the Marble world model and others need human editors

When AI learns physics: why the Marble world model and others need human editors

Minimal 3D render

WorldLabs just released the Marble world model this week. It's a prompt-to-3D model generator, and the AI community is paying attention.

The company is led by Fei-Fei Li, one of the pioneers who helped build the foundation for modern computer vision. So, this isn't just another startup releasing another tool. It's a signal that something fundamental is shifting in how we think about creating spatial content.

Because, for years, building 3D environments meant technical mastery. We had to know modeling software, understand lighting systems, and manually place every object. 

Marble suggests a different path: describe what you want, and the AI generates it. That's the promise of world model AI 3D generation.

But here's what's interesting about this moment. History shows us that when AI handles what used to require mastery, the question isn't whether humans matter. It's which human skills suddenly matter more.

We're entering the era of world models. The implications stretch far beyond content generation. They touch how we train people, how we communicate complex ideas, and how enterprises actually operate.

Beyond language: teaching AI to understand space

Large language models (LLMs) changed how we interact with information. They can write, summarize, translate, and reason through text. ChatGPT can explain quantum physics and then write poetry two minutes later. Claude can analyze complex documents and help with coding novel applications. 

But they all share a fundamental limitation: they don't understand physical space.

Ask an LLM why a ball rolls downhill, and it can explain gravity in words. It can even cite Newton's laws and describe acceleration. 

But it can't visualize the slope. It can’t predict the trajectory. It can’t understand how the ball would behave if you changed the angle. 

As we live in a physical world, this matters more than it seems. A lot of the work we humans do isn't about processing text. It's about navigating the space around us, manipulating objects, and understanding how things move and interact in three dimensions. 

This is where Spatial Intelligence plays a crucial role. Humans naturally develop spatial intelligence from the moment we are toddlers. We instinctually learn how objects move, how shadows shift with light, and how distance affects size. 

Like humans, the Marble world model and other image to 3D world AI models possess spatial intelligence.

Unlike humans, AI has to learn this differently. Multimodal world models do this through repeated exposure to visual data. Such as millions of images and videos showing how the physical world behaves.

They observe enough examples and begin to internalize the laws of physics and causal relationships in three dimensions. They learn that water flows downward, that solid objects don't pass through each other, and that light creates shadows in predictable patterns.

The results of this exposure are impressive. Multimodal world models can generate consistent, persistent 3D environments that obey real-world logic.  A chair stays solid. Light casts shadows in the right direction. Objects don't float unless they're supposed to. 

Spatial intelligence world models are learning what humans have always known: people are spatial, not textual. 

WorldLabs isn’t the only player in the field. Google is building world models of their own. So are Meta and Tencent

The race is on for world model AI 3D generation, and the applications are broader than most people realize: gaming, simulation, robotics, and yes, XR training. 

Spatial intelligence isn't just about generation; it's about application. Talk to our experts to learn how enterprises are already using AI-powered environments to train workers, guide real-time decisions, verify compliance, and communicate complex operational knowledge.

Learn more

The human layer still matters

Here's what gets lost in the excitement around world models. AI excels at massive generation and knowledge acquisition. But certain types of intelligence remain distinctly human. 

For immersive experiences, this gap becomes critical.

Because the suspension of disbelief in XR is higher than in any other medium. For instance, when someone watches a video, they're observers. Hence, small inconsistencies get forgiven. 

But in VR, users aren't just watching; they're present. Their brains are sending signals that they're standing on a factory floor or inside a pharmaceutical lab. 

So, when the brain notices when something doesn't match how the real world works, like a lighting inconsistency or an interaction that feels slightly off, it pulls you out of the experience. 

This isn't a technical problem that better world models will solve. It's a design problem. It's about understanding how humans perceive, learn, and make decisions in immersive environments.

AI can't yet understand why a specific training scenario needs to break those rules for pedagogical reasons.  

They don't know that a manufacturing floor should feel slightly more spacious than reality to reduce cognitive load. They don't know that certain safety hazards need exaggerated visual cues to register in a trainee's muscle memory. 

That's where the gap becomes clear. World models lack the nuance that subject matter experts bring. And SME knowledge is where the real value lives.

A pharmaceutical QA trainer knows that certain procedural steps require deliberate pacing. Because that's where mistakes actually happen under pressure. 

A safety instructor understands which near-miss scenarios trigger the right kind of attention without inducing panic. Too realistic, and you terrify people. Too abstract, and they don't take it seriously. 

This SME knowledge doesn't exist in the training data. It's not written down in PDFs or captured in videos. It lives in decades of on-the-ground experience, incident reports, and watching thousands of workers learn.

The solution isn't better generation. It's better editing. 

World models can create the raw material. But tools that can take those outputs and reshape them with human expertise become essential. 

And analytics systems that track not just user behavior but how the generated environments themselves perform become critical. Did workers who trained in this scenario make fewer mistakes on the floor? Did inspection rates improve? Did incident reports go down? 

This allows for course correction based on actual learning outcomes, not just technical accuracy. 

The convergence is the destination

The convergence is the destination

Spatial is coming. 2026 will be the year that sees numerous developments in the field of spatial intelligence.

Seismic moves within the market affirm this perspective that we've held for so long. 

Jeff Bezos' investment in Project Prometheus is one such move. The fact that Bezos is personally backing a spatial computing venture says something about where the smart money sees this going.

Samsung's Galaxy XR launched earlier this year. Valve announced Steam Frame. Apple released an updated Vision Pro.

But for the infrastructure for spatial computing to come together, it takes not just advances in hardware and software, but also in content. Especially in the case of enterprises.

For content to advance, you need the involvement of human experts. You need tools like VRseBuilder that have features like Excel integration, drag-and-drop creation, and pre-made templates, which enable on-site teams to configure training modules and other XR experiences based on their unique nuances.

When IBM's Deep Blue beat Kasparov in 1997, it felt like the end of human chess mastery. The machine had won. The best human player in the world couldn't compete with raw computational power. 

Later, Kasparov pioneered "advanced chess": humans paired with computers. Then, something unexpected happened. 

Turns out, a partnership of AI and humans beat not just humans, but also other AIs. 

The advantage shifts to who combines AI best with human expertise. 

Not AI replacing humans. Not humans working the same way they always have. But a new model where each does what it does best, and the combination creates something neither could achieve alone.

FAQs

1. What exactly are world models?

World models are AI systems that learn spatial intelligence by observing millions of images and videos. They internalize how the physical world works, learning about gravity, light, collision, and cause and effect. Hence, unlike language models that operate on text, world models comprehend three-dimensional space. As a result, they can generate consistent 3D environments that obey real-world physics.

1. What exactly are world models?

World models are AI systems that learn spatial intelligence by observing millions of images and videos. They internalize how the physical world works, learning about gravity, light, collision, and cause and effect. Hence, unlike language models that operate on text, world models comprehend three-dimensional space. As a result, they can generate consistent 3D environments that obey real-world physics.

1. What exactly are world models?

World models are AI systems that learn spatial intelligence by observing millions of images and videos. They internalize how the physical world works, learning about gravity, light, collision, and cause and effect. Hence, unlike language models that operate on text, world models comprehend three-dimensional space. As a result, they can generate consistent 3D environments that obey real-world physics.

1. What exactly are world models?

World models are AI systems that learn spatial intelligence by observing millions of images and videos. They internalize how the physical world works, learning about gravity, light, collision, and cause and effect. Hence, unlike language models that operate on text, world models comprehend three-dimensional space. As a result, they can generate consistent 3D environments that obey real-world physics.

2. How are multimodal world models different from traditional 3D modeling tools?

2. How are multimodal world models different from traditional 3D modeling tools?

2. How are multimodal world models different from traditional 3D modeling tools?

2. How are multimodal world models different from traditional 3D modeling tools?

3. Why can't the Marble world model and other world models just handle everything if they understand physics?

3. Why can't the Marble world model and other world models just handle everything if they understand physics?

3. Why can't the Marble world model and other world models just handle everything if they understand physics?

3. Why can't the Marble world model and other world models just handle everything if they understand physics?

Let’s talk about

Let’s talk about

Let’s talk about

your training

your training

your training

Talk to our team to learn how to implement VR training at scale

Talk to our team to learn how to implement VR training at scale

Talk to our team to learn how to implement VR training at scale