
That’s the question sparked by two events making waves across the AI community. First, Meta’s Chief AI Scientist Yann LeCun, one of the founding fathers of modern AI, announced his departure from the company to start a new lab focused on world models. Second, Stanford’s Dr. Fei-Fei Li published an essay declaring that spatial intelligence—AI that perceives and interacts with the real world—is the next major frontier.
Together, their ideas signal a shift from language-driven systems toward AI that can sense, reason, and act in space. This evolution could redefine robotics, scientific research, and human-AI collaboration for decades to come. Taking a tech certification will help you understand the intricacies of AI.
How AI and Industry Are Learning to Work Together
Before diving into the research side of AI’s evolution, consider the cultural shift happening in how industries collaborate with it.
Voice tech company 11 Labs just launched the Iconic Voices Marketplace, a platform that licenses celebrity voices for content and advertising. Users can now legally use synthetic versions of voices like Michael Caine, Judy Garland, John Wayne, and Maya Angelou. The company calls it a “consent-based, performer-first approach”—a response to earlier backlash over unauthorized voice cloning.
Michael Caine himself endorsed the idea, saying, “We can preserve and share voices, not just mine, but anyone’s. It’s not about replacing voices, it’s about amplifying them.” Actor Matthew McConaughey has also joined, allowing his voice to be used in a Spanish-language adaptation of his newsletter Lyrics of Living.
This marks the beginning of a larger trend: industries that once viewed AI as a threat are now learning to collaborate with it. Artists and creators are licensing their likenesses. Companies are monetizing AI tools responsibly. The commercialization of creative IP for AI has begun.
SoftBank’s Massive AI Investment and the Nvidia Exit
While the cultural side of AI matures, the financial side is in overdrive. SoftBank recently revealed it sold all 32.1 million of its Nvidia shares—worth around $5.8 billion—to help fund its $30 billion investment in OpenAI. The last $22.5 billion is due in December following OpenAI’s conversion to a for-profit entity.
The sale shocked markets, with Nvidia stock slipping 3% and SoftBank shares dropping 10%. Still, this wasn’t a retreat—it was a reallocation. CEO Masayoshi Son is doubling down on AI infrastructure, even borrowing $5 billion against his stake in Arm Holdings to fund the deal.
Analyst Gavin Baker called it typical Son behavior. The billionaire famously sold his entire Nvidia stake in 2019, missing out on more than $100 billion in gains. Now he’s back in, betting that OpenAI’s future outweighs short-term market volatility.
At the same time, Blue Owl Capital announced a $3 billion equity contribution for OpenAI’s Stargate data center in New Mexico, alongside $18 billion in bank debt. It’s one of the largest private capital allocations in the history of cloud computing. Together, these moves highlight the enormous scale of AI’s physical backbone—the data centers that power the models of tomorrow.
AMD’s Ambitious Challenge to Nvidia
Not every player in the AI chip race is standing still. At a recent event, AMD CEO Lisa Su laid out a bold vision to carve out a double-digit share of the data center GPU market within five years. She projected 35% annual revenue growth, with the data center segment growing 60%, fueled by what she called “insatiable demand for AI chips.”
AMD’s new MI400X servers, containing 72 GPUs per rack, are designed for large-scale AI workloads. OpenAI has already committed to deploying a gigawatt of AMD hardware, while Meta and Oracle have signed long-term partnerships.
Next year will be the true test. If AMD can compete with Nvidia on performance and efficiency, it could reshape the global AI hardware market.
Meta’s Surprising Momentum
Amid all this, Meta’s AI division is having an unexpected resurgence. The Meta AI web app saw 105% traffic growth between September and October, making it the fastest-growing AI web platform that month.
Over the past year, Meta’s AI traffic has grown 149%, compared to 305% for Gemini and 68% for ChatGPT. Many analysts credit the surge to Meta’s new Sora competitor, Vibes, which quietly gained traction after its September release.
While skeptics question the sustainability of this growth, others argue it reflects how mainstream users interact differently than the hyper-online AI community. As one analyst put it, “The hardcore AI world might be in a bubble. Most people just want AI that feels natural—and Meta seems to be delivering that.”
The End of an Era: Yann LeCun Leaves Meta
Then came the news that rocked AI research. Yann LeCun, Meta’s long-time Chief AI Scientist and a 2018 Turing Award winner, is leaving to start his own company focused on next-generation AI systems.
LeCun joined Meta in 2013, well before OpenAI existed, and built the FAIR lab—the birthplace of Meta’s LLaMA models. Under his leadership, Meta helped define the modern deep learning landscape. But the writing was on the wall once Mark Zuckerberg formed the Superintelligence Division led by Alexander Wang and Nat Friedman, signaling a pivot from research toward productization.
LeCun’s FAIR team was eventually folded into Wang’s new organization, and his role became increasingly academic. Industry watchers say the split was inevitable. Pedro Domingos noted that the announcement alone wiped $30 billion off Meta’s market cap.
Reactions were divided. Some called it a routine leadership change, others saw it as an intellectual clash. John Hernandez argued that LeCun’s theoretical mindset clashed with Meta’s “wartime” execution culture. Jeffrey Emanuel added that LeCun belongs in “a Bell Labs or Xerox PARC setting,” where discovery, not deadlines, drives progress.
Still, many believe LeCun’s next venture will attract billions in funding. If his new company succeeds, it could easily become a prime acquisition target for DeepMind or Google.
Inside LeCun’s Vision: The Rise of World Models
At the heart of LeCun’s work is a radical new direction—world models. Unlike language models that learn from text, world models learn from video, spatial data, and physical interaction.
They aim to build AI that can understand the real world the way humans do—by perceiving motion, physics, and causality. These models wouldn’t just predict the next word; they’d predict the next state of the environment.
Sources close to the project say LeCun’s new lab will focus entirely on this idea, potentially taking a decade or more to develop fully. He believes world models could power truly autonomous systems that reason about the physical world instead of just describing it.
Fei-Fei Li’s Call for Spatial Intelligence
At the same time, Dr. Fei-Fei Li, co-director of Stanford’s Human-Centered AI Institute, published an essay titled “From Words to World: Spatial Intelligence Is AI’s Next Frontier.” Her thesis echoes LeCun’s but expands on its philosophical foundation.
Li argues that spatial intelligence—the ability to perceive, imagine, and reason about the physical world—is the next major leap. She describes today’s large language models as “wordsmiths in the dark: eloquent but inexperienced, knowledgeable but ungrounded.”
According to Li, LLMs can generate fluent text, but they can’t navigate space, estimate distance, or reason about physical interactions. They’re “blind storytellers,” impressive but detached from reality.
She defines spatial intelligence as the bridge between perception and action, between seeing and doing. It’s what allows humans to manipulate objects, imagine structures, and plan movements—all crucial for general intelligence.
What World Models Can Actually Do
Li outlines three defining traits of world models:
- Generative: They can create realistic, physics-consistent worlds that follow natural laws like gravity and light dynamics.
- Multimodal: They process diverse inputs—images, text, gestures, depth maps—to simulate complex environments.
- Interactive: They predict what happens next when an agent takes action.
In practice, this means AI that can not only describe a room but move through it, pick up objects, or imagine new layouts.
Fei-Fei Li says such models could revolutionize robotics, drug discovery, materials science, and education. Imagine AI that understands how molecules interact in 3D space, or a robot that learns physical coordination through simulation rather than trial and error.
Why Language Isn’t Enough
Li points out that even today’s most advanced multimodal models struggle with spatial reasoning. They can describe a photo but can’t estimate depth, orientation, or distance accurately.
They fail at basic physical tasks—like rotating objects or predicting trajectories—and AI-generated videos lose coherence after a few seconds.
Language is a one-dimensional signal; the world is multi-dimensional and governed by physics. To build true intelligence, AI must grasp those physical rules, not just the words we use to describe them.
Research Challenges and Breakthroughs Ahead
World model research introduces new challenges. Unlike language data, which is easily tokenized, spatial data involves geometry, time, and causality. Representing that in machine learning requires entirely new architectures.
Fei-Fei Li’s lab is working on:
- New task functions for training spatial reasoning.
- Extracting 3D understanding from 2D videos and images.
- Creating model architectures that blend symbolic reasoning with continuous perception.
She calls this the most complex challenge AI has ever faced. But the payoff could redefine what machines are capable of—giving AI a genuine sense of space, context, and interaction.
Beyond Language: The Future of Intelligent Systems
Both LeCun and Li agree that the next stage of AI won’t replace language models but expand beyond them. They see a dual-track future:
- Language-centered AI, driven by OpenAI, Anthropic, and Google, mastering text and reasoning.
- World-centered AI, led by LeCun and Li, focusing on embodied intelligence—AI that learns by doing.
For professionals building skills to navigate this dual paradigm, programs like Deep Tech Certification provide critical knowledge about spatial data, multimodal modeling, and physical reasoning. For leaders and strategists, the Marketing and Business Certification helps connect these technical advances with enterprise strategy and innovation.
Why This Matters
The language model revolution gave machines the ability to communicate. The next revolution will give them the ability to understand the world they talk about.
If LeCun’s world models and Li’s spatial intelligence succeed, AI won’t just generate text or images—it will learn to experience, reason, and act.
It’s a monumental shift that could reshape industries from robotics to healthcare. In Li’s words, “Spatial intelligence will transform how we create and interact with real and virtual worlds.”
The AI era of words is maturing. The era of worlds is about to begin.