Close Menu
Scroll Tonic
  • Home
  • Smart Gadgets
  • AI & Daily Tools
  • Digital Well-Being
  • Home Office Setup
  • Productivity Apps

Subscribe to Updates

Stay updated with Smart Gadgets, AI tools, productivity apps, digital well-being tips, and smart home office ideas.

What's Hot

When product managers ship code: AI just broke the software org chart

This LG Smart Window AC with Heating and Smart Controls Is Just $400 Right Now

20+ Solved ML Projects to Boost Your Resume

Facebook X (Twitter) Instagram
Scroll Tonic
  • Home
  • Smart Gadgets
  • AI & Daily Tools
  • Digital Well-Being
  • Home Office Setup
  • Productivity Apps
Scroll Tonic
You are at:Home»Smart Gadgets»Three ways AI is learning to understand the physical world
Smart Gadgets

Three ways AI is learning to understand the physical world

team_scrolltonicBy team_scrolltonicMarch 22, 2026006 Mins Read
Share Facebook Twitter Pinterest LinkedIn Tumblr Email
Three ways AI is learning to understand the physical world
Share
Facebook Twitter LinkedIn Pinterest Email

Large language models are running into limits in domains that require an understanding of the physical world — from robotics to autonomous driving to manufacturing. That constraint is pushing investors toward world models, with AMI Labs raising a $1.03 billion seed round shortly after World Labs secured $1 billion.

Large language models (LLMs) excel at processing abstract knowledge through next-token prediction, but they fundamentally lack grounding in physical causality. They cannot reliably predict the physical consequences of real-world actions. 

AI researchers and thought leaders are increasingly vocal about these limitations as the industry tries to push AI out of web browsers and into physical spaces. In an interview with podcaster Dwarkesh Patel, Turing Award recipient Richard Sutton warned that LLMs just mimic what people say instead of modeling the world, which limits their capacity to learn from experience and adjust themselves to changes in the world.

This is why models based on LLMs, including vision-language models (VLMs), can show brittle behavior and break with very small changes to their inputs. 

Google DeepMind CEO Demis Hassabis echoed this sentiment in another interview, pointing out that today’s AI models suffer from “jagged intelligence.” They can solve complex math olympiads but fail at basic physics because they are missing critical capabilities regarding real-world dynamics. 

To solve this problem, researchers are shifting focus to building world models that act as internal simulators, allowing AI systems to safely test hypotheses before taking physical action. However, “world models” is an umbrella term that encompasses several distinct architectural approaches. 

That has produced three distinct architectural approaches, each with different tradeoffs.

JEPA: built for real-time

The first main approach focuses on learning latent representations instead of trying to predict the dynamics of the world at the pixel level. Endorsed by AMI Labs, this method is heavily based on the Joint Embedding Predictive Architecture (JEPA). 

JEPA models try to mimic how humans understand the world. When we observe the world, we do not memorize every single pixel or irrelevant detail in a scene. For example, if you watch a car driving down a street, you track its trajectory and speed; you do not calculate the exact reflection of light on every single leaf of the trees in the background. 

V-JEPA

V-JEPA architecture (source: Meta FAIR)

JEPA models reproduce this human cognitive shortcut. Instead of forcing the neural network to predict exactly what the next frame of a video will look like, the model learns a smaller set of abstract, or “latent,” features. It discards the irrelevant details and focuses entirely on the core rules of how elements in the scene interact. This makes the model robust against background noise and small changes that break other models.

This architecture is highly compute and memory efficient. By ignoring irrelevant details, it requires much fewer training examples and runs with significantly lower latency. These characteristics make it suitable for applications where efficiency and real-time inference are non-negotiable, such as robotics, self-driving cars, and high-stakes enterprise workflows. 

For example, AMI is partnering with healthcare company Nabla to use this architecture to simulate operational complexity and reduce cognitive load in fast-paced healthcare settings. 

Yann LeCun, a pioneer of the JEPA architecture and co-founder of AMI, explained that world models based on JEPA are designed to be “controllable in the sense that you can give them goals, and by construction, the only thing they can do is accomplish those goals” in an interview with Newsweek.

Gaussian splats: built for space

A second approach leans on generative models to build complete spatial environments from scratch. Adopted by companies like World Labs, this method takes an initial prompt (it could be an image or a textual description) and uses a generative model to create a 3D Gaussian splat. A Gaussian splat is a technique for representing 3D scenes using millions of tiny, mathematical particles that define geometry and lighting. Unlike flat video generation, these 3D representations can be imported directly into standard physics and 3D engines, such as Unreal Engine, where users and other AI agents can freely navigate and interact with them from any angle.

The primary benefit here is a drastic reduction in the time and one-time generation cost required to create complex interactive 3D environments. It addresses the exact problem outlined by World Labs founder Fei-Fei Li, who noted that LLMs are ultimately like “wordsmiths in the dark,” possessing flowery language but lacking spatial intelligence and physical experience. World Labs’ Marble model gives AI that missing spatial awareness. 

While this approach is not designed for split-second, real-time execution, it has massive potential for spatial computing, interactive entertainment, industrial design, and building static training environments for robotics. The enterprise value is evident in Autodesk’s heavy backing of World Labs to integrate these models into their industrial design applications.

End-to-end generation: built for scale

The third approach uses an end-to-end generative model to process prompts and user actions, continuously generating the scene, physical dynamics, and reactions on the fly. Rather than exporting a static 3D file to an external physics engine, the model itself acts as the engine. It ingests an initial prompt alongside a continuous stream of user actions, and it generates the subsequent frames of the environment in real-time, calculating physics, lighting, and object reactions natively. 

DeepMind’s Genie 3 and Nvidia’s Cosmos fall into this category. These models provide a highly simple interface for generating infinite interactive experiences and massive volumes of synthetic data. DeepMind demonstrated this natively with Genie 3, showcasing how the model maintains strict object permanence and consistent physics at 24 frames per second without relying on a separate memory module.

This approach translates directly into heavy-duty synthetic data factories. Nvidia Cosmos uses this architecture to scale synthetic data and physical AI reasoning, allowing autonomous vehicle and robotics developers to synthesize rare, dangerous edge-case conditions without the cost or risk of physical testing. Waymo (a fellow Alphabet subsidiary) built its world model on top of Genie 3, adapting it for training its self-driving cars.

The downside to this end-to-end generative method is the great compute cost required to continuously render physics and pixels simultaneously. Still, the investment is necessary to achieve the vision laid out by Hassabis, who argues that a deep, internal understanding of physical causality is required because current AI is missing critical capabilities to operate safely in the real world.

What comes next: hybrid architectures

LLMs will continue to serve as the reasoning and communication interface, but world models are positioning themselves as foundational infrastructure for physical and spatial data pipelines. As the underlying models mature, we are seeing the emergence of hybrid architectures that draw on the strengths of each approach. 

For example, cybersecurity startup DeepTempo recently developed LogLM, a model that integrates elements from LLMs and JEPA to detect anomalies and cyber threats from security and network logs. 

Learning physical understand Ways world
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
Previous ArticleGoogle Just Made Four Big Upgrades to Android Gaming on Your Windows PC
Next Article Low-Key Debut for Entra ID Backup and Recovery
team_scrolltonic
  • Website

Related Posts

When product managers ship code: AI just broke the software org chart

March 30, 2026

Amazon Spring Sale: Save $500 on this Samsung 40-inch G7 gaming monitor

March 29, 2026

What Is the Best Garmin Watch Right Now? (2026)

March 28, 2026
Add A Comment
Leave A Reply Cancel Reply

Top Posts

Must-Have AI Tools for Work and Personal Productivity

February 9, 2026735 Views

Best AI Daily Tools for Notes and Task Planning

January 25, 2026730 Views

Punkt Has a New Smartphone for People Who Hate Smartphones

January 5, 2026727 Views
Stay In Touch
  • Facebook
  • Pinterest

Subscribe to Updates

Stay updated with Smart Gadgets, AI tools, productivity apps, digital well-being tips, and smart home office ideas.

Keep Scrolling. Stay Refreshed. Live Smart.
A modern digital lifestyle blog simplifying tech for everyday productivity and well-being.

Categories
  • AI & Daily Tools
  • Digital Well-Being
  • Home Office Setup
  • Productivity Apps
  • Smart Gadgets
  • Uncategorized
QUick Links
  • About Us
  • Contact Us
  • Disclaimer
  • Privacy Policy
  • Terms & Conditions

© 2026 Scroll Tonic | Keep Scrolling. Stay Refreshed. Live Smart.

Type above and press Enter to search. Press Esc to cancel.