The Shift to Physical AI — and What It Will Be Built On

We are moving from language AI to world AI — and two skills are becoming foundational.

If you zoom out and look at where AI is heading, one thing becomes obvious:

We are moving from language AI → world AI.

Not just models that describe the world, but systems that understand, simulate, and act inside it.

That shift has a name: Physical AI.

And if you want to be relevant in that world, two skills stand out above the rest:

  1. 3D Gaussian Splatting (3DGS)
  2. Diffusion models — especially Rectified Flow

The big shift: from tokens to worlds

Look at what’s happening across the ecosystem:

World Labs (founded by Fei-Fei Li) is building spatial intelligence — models that understand 3D worlds, not just text.

Yann LeCun launched Advanced Machine Intelligence (AMI) Labs focused on world models, not LLMs.

These are not random bets.

They’re signals.

The next generation of AI systems will need to:

That requires two core capabilities:

👉 A way to represent the world (3D)
👉 A way to generate and complete the world (generative models)

1. 3D Gaussian Splatting: the new default for 3D

3D Gaussian Splatting (3DGS) is quietly becoming the most practical 3D representation we’ve seen in years.

Why?

Because it hits a rare combination:

Instead of meshes or implicit neural fields, 3DGS represents a scene as millions of small 3D Gaussians — optimized directly from images.

And crucially:

It’s fast enough to actually use in real systems.

Why this matters for Physical AI

If you're building anything that touches the real world:

You need a way to capture reality quickly and render it interactively.

That’s exactly where 3DGS shines.

SAM3D

SAM3D is a perfect example of where things are going.

It can take an image and output:

That second option is an important one.

3DGS is no longer experimental — it’s becoming a standard output format.

This is the key shift:

👉 Mesh is no longer the only way to represent 3D
👉 3DGS is becoming almost as important as mesh

If tools natively export 3DGS, then:

Knowing how to work with 3DGS becomes a core skill.

2. Diffusion models → Rectified Flow: the generative engine

Diffusion models changed everything in generative AI.

But they also came with a problem:

Too many steps. Too slow.

That’s where Rectified Flow comes in.

The intuition behind Rectified Flow

Instead of:

Rectified Flow:

Why is that powerful?

Because:

In some cases, you can get:
👉 near one-step generation

Why this matters for Physical AI

In the physical world, you need:

Diffusion (especially modern variants like Rectified Flow) gives you:

It becomes the imagination layer on top of your 3D world.

The real story: it’s not 3DGS vs diffusion

It’s:

3DGS + diffusion = Physical AI stack

Think of it like this:

Layer Role
3DGS Capture & represent reality
Diffusion / Rectified Flow Generate, augment, simulate

Examples:

This is exactly the kind of pipeline people are building right now.

Why these two skills matter now

We are at a transition point similar to early deep learning.

Back then:

Today:

Because:

What to do about it (seriously)

If you’re an engineer, researcher, or builder:

Don’t just read about this stuff.

Learn it properly.

1. Learn 3D Gaussian Splatting

If you want to go from:

“I ran a repo” → “I actually understand and can modify this”

👉 https://3dgaussiansplattingcourse.com/

2. Learn Rectified Flow & modern diffusion (SD 3.5)

If you want to understand:

👉 https://rectifiedflowcourse.com/

3. Stay ahead (3DGS updates)

This space is moving insanely fast.

If you want curated updates:

👉 https://3dgs.teachable.com/p/newsletter-3d-gaussian-splatting

Final thought

The industry is converging on something very clear:

And when it does:

The people who understand 3D representations + generative models will be the ones building the future.

Everything else will feel like prompt engineering.