Latent Space: The Hidden Infrastructure of Intelligence

The First Time You Realize AI Sees the World Differently

Latent space is one of those concepts that feels deceptively simple but quickly becomes mind-bending the deeper you go. In my MIT coursework, the moment it truly clicked wasn’t when someone showed a diagram or equation—it was when I watched two very different inputs land right next to each other in a high‑dimensional embedding space. Suddenly, you realize: AI doesn’t see categories the way we do. It sees geometry. And that geometry is the beating heart of modern AI.

This post is my attempt to make latent space both intuitive and technically sound—a tour of the hidden mathematical world that lets AI models generalize, reason, and occasionally surprise the hell out of us. If you’ve read my earlier posts like Efficiency Reckoning or AI Is Eating Software That Is Eating the World, you’ll recognize a recurring theme: exponential capability often hides in plain sight until you learn to see the structure underneath.

What Is Latent Space?

Latent space is the compressed mathematical world where AI stores meaning. Instead of memorizing data, models learn dense vector representations that capture the essence of concepts—objects, actions, styles, emotions, operational patterns. Similar ideas cluster together; different ideas drift apart. Geometry becomes understanding.

Key ideas:

  • Embeddings: Numerical vectors that represent the meaning of inputs (words, images, tokens). Their position and direction encode semantic relationships.
  • Distance: A mathematical measure (often cosine similarity or Euclidean distance) that indicates how similar two embeddings are. Closer = more related.
  • Manifolds: Lower‑dimensional, structured surfaces within the high‑dimensional latent space where meaningful data naturally clusters. Models “discover” these during training.

Everything the model “knows” lives somewhere in this hidden space.

Why Latent Space Is AI’s Superpower

Latent spaces give models the ability to:

  • Generalize beyond what they’ve seen.
  • Recognize analogies and patterns.
  • Perform zero-shot reasoning (answer questions they were never explicitly trained on).
  • Compress knowledge into a shape that can be navigated, manipulated, and queried.

This geometry is the fuel behind why large models feel so shockingly capable. It’s the same idea I explored in The Law of Accelerating Returns, systems don’t merely improve—they reshape the surface beneath our feet. Latent space is the mathematical expression of that reshaping. We’re no longer programming rules; we’re shaping the very spaces where meaning lives.

How Latent Spaces Are Built

Latent space emerges naturally during training, driven by the model’s need to predict missing information.

The Compression Process

Self-supervision forces the model to strip away noise and preserve structure. This compression yields abstract, high-dimensional patterns that capture relationships rather than raw inputs.

Transformation Through Layers

Embeddings pass through dozens or hundreds of transformer layers. Each layer rotates, stretches, and refines meaning until stable semantic structures emerge.

The Result: A Structured World

By late training, the model has carved out clear neighborhoods for concepts—objects clustering near the actions they relate to, pricing signals gravitating toward contextual cues, operational patterns forming their own orbits.

How Latent Space Behaves

Despite being abstract and high dimensional, latent spaces exhibit surprisingly intuitive properties. Humans naturally build mental maps to navigate ambiguity, and AI does something similar—just at a scale and dimensionality far beyond our own., latent spaces exhibit surprisingly intuitive properties.

Smoothness

Small moves yield gradual changes in meaning, enabling interpolation, transformation, and reinterpretation.

Relational Structure

Directional changes encode relationships—analogy, comparison, and categorization become geometric operations.

Compositionality

Concepts can combine fluidly: an object + context + constraint forms a new point in space that the model can reason about without explicit rules.

Natural Clustering

Clusters form organically, often better than human taxonomies—but also reflecting limitations or hidden biases.

Where Latent Space Breaks Down

As magical as it feels, latent space isn’t perfect:

  • Latent collapse: Everything clusters too tightly.
  • Overfitting: Geometry becomes brittle.
  • Bias: Prejudices become encoded as spatial structure.
  • Out-of-distribution drift: The model hallucinating outside the manifold.

These limitations matter when deploying AI into real operational environments, where edge cases are everywhere.

The Future: Latent Space as the New Programming Model

We are entering an era where latent space isn’t just a byproduct of AI models—it becomes the substrate of software itself. The shift is profound: instead of writing rules, we increasingly shape geometry, influence structure, and design the conditions under which models discover meaning.

Several forces are driving this transformation:

  • Geometry replaces logic. Traditional programming encodes explicit steps; latent‑space systems embed intent, relationships, and constraints into the shape of the space itself. We’re not prescribing behavior—we’re defining the terrain.
  • Agents operate like navigators, not executors. Agents don’t follow deterministic paths. They explore, sample, and move through conceptual regions, selecting actions by proximity, similarity, and predicted outcomes. This is closer to robotics in a physical world than software in a deterministic one.
  • World models introduce simulation as a first‑class primitive. When a model can simulate consequences inside its latent space, it stops behaving like a tool and starts behaving like a planner. Software becomes anticipatory rather than reactive.
  • Developers shift from coding workflows to curating spaces. The primary task becomes shaping embeddings, conditioning behavior, tuning representations, and steering emergent structure. Infrastructure teams will manage vector spaces the way they once managed databases.

This isn’t incremental. It is a foundational rewrite of how software is conceived and built—arguably the most important transition since cloud computing abstracted away hardware. Latent space abstracts away rules themselves.

Personal Reflection: Why This Matters to Me

Returning to technical study has been a joy—and a humbling reminder that AI is a field where intuition and mathematics collide. Latent space, more than any other concept, embodies that collision in computational form.

Understanding it doesn’t just make you a better builder of AI systems.
It makes you a better interpreter of AI behavior.

And maybe a bit more forgiving when a model wanders too far off the manifold.

The World Models Within Us

From Asimov to AGI: The Rise of Predictive Minds

“A robot may not injure a human being, or, through inaction, allow a human being to come to harm.” — Isaac Asimov

Isaac Asimov, one of my favorite authors, is best known for his Robot series — books that deeply influenced how I think about technology and ethics. The series wasn’t just about robots; it was about the emergence of intelligence — about what happens when a machine begins not just to obey, but to understand. His characters like Dr. Susan Calvin grappled with robots who broke the Three Laws not out of rebellion, but out of deeper reasoning — they’d built internal models of the world complex enough to predict human consequences.

That, in essence, is the story of world models in modern AI. Machines are learning not just to process data, but to imagine futures.

The Foundation of World Models

While most consumer-facing AI tools we see in the news or use at home or work — like ChatGPT, image generators, or voice assistants — excel at pattern recognition, true world models go a step further. They don’t just respond to input; they build internal representations of how the world works, allowing them to reason about context, dynamics, and cause-effect relationships. This makes them capable of simulating and anticipating outcomes rather than simply reproducing patterns from past data. A world model is a latent representation of reality that allows an intelligent system to simulate, plan, and predict. In technical terms, it’s what allows agents like MuZero or GPT-based systems to operate beyond mere pattern recognition. Instead of memorizing, they model — forming compressed internal maps of how the world behaves.

If a neural network is like a camera capturing pixels, a world model is like a mapmaker building an atlas. It abstracts, generalizes, and anticipates — and crucially, it allows for planning. A good world model can take incomplete information and still simulate likely futures. It’s the engine behind self-driving cars navigating uncertainty, generative agents predicting user intent, and language models inferring context.

This idea originated in reinforcement learning research (Ha & Schmidhuber’s World Models, DeepMind’s MuZero, OpenAI’s Sora). But the same principle now underpins multimodal systems like GeminiGato, and GPT-5, which combine perception, reasoning, and action under one unified architecture. These systems don’t just respond — they simulate, project, and plan.

From Logic to Learning: How We Got Here

The lineage of AI traces a fascinating arc. It began with symbolic AI, when we tried to encode knowledge into explicit rules — much like Asimov’s Three Laws. Those systems could reason, but only within the narrow confines of their logic. Then came the revolution of statistical learning, where we replaced rules with probabilities and let data teach the machine. That gave us vision systems, speech recognition, and language models — but not true understanding.

Now we’re entering the age of world modeling — a synthesis between symbolic precision and statistical intuition. These new systems learn not only to recognize patterns, but to simulate how the world changes over time. It’s a step closer to how humans think. When you reach for a coffee mug, you don’t calculate every possible trajectory — your brain’s world model has already predicted the motion.

World models are the missing bridge between perception and agency. They turn observation into imagination and allow intelligent agents to act not just reactively, but intentionally.

Why World Models Matter

Every intelligent system — human or machine — must answer the same question: What will happen if I act? That is the essence of intelligence — the ability to simulate the future and choose accordingly.

In AI, this predictive capacity is fundamental:

  • In AGI research, world models enable self-supervised learning, planning, and reasoning across tasks.
  • In enterprise applications, they power agents that can simulate business outcomes before executing decisions.
  • In human cognition, predictive coding lets our brains anticipate the world milliseconds before our senses confirm it.

World models, then, are not just a technical concept. They’re a philosophy of intelligence — a recognition that foresight, not memory, is the truest marker of understanding. They allow systems to bridge the gap between data and decision, between reaction and anticipation.

The Business Parallel

In business and technology alike, we build world models to make sense of complexity. Whether we call them strategy frameworks, simulation tools, or digital twins, these systems help us forecast outcomes, reduce uncertainty, and align execution. A well-constructed model of operations or markets acts as a simplified reflection of reality, letting leaders test scenarios before taking action.

Organizations that refine their world models — capturing customer behavior, market dynamics, and operational constraints — become more adaptive. Pricing, scheduling, and resource allocation evolve from guesswork into predictive, data-informed decisions. The fidelity of a company’s model of its world determines how effectively it can act within it.

Every system — human, corporate, or artificial — competes on the fidelity of its world model. The sharper the map, the smarter the motion.

The Rise of Embodied Intelligence

Recent advancements in robotics have brought world models from simulation to the physical world. Robots are no longer confined to factory floors or research labs—they’re navigating complex, unstructured environments using predictive world modeling to anticipate, adapt, and learn.

Take Tesla’s OptimusBoston Dynamics’ Atlas, or Figure AI’s humanoid robots—each of these systems relies on internal models to understand and predict physical interactions. They don’t just follow pre-programmed paths; they build internal maps of how objects, gravity, and force behave. When a robot walks across uneven terrain or picks up a delicate object, it’s using its world model to simulate potential outcomes before acting.

Even in household robotics, we’re seeing the impact. Modern robot vacuums and warehouse automation bots use spatial and behavioral world models to improve navigation and decision-making over time.

This convergence of embodied AI and world modeling is what gives machines agency in the real world. The same predictive reasoning that allows a language model to plan a paragraph now lets a robot plan a movement. As the fidelity of these physical world models increases, robots will move from reactive tools to proactive collaborators—learning continuously through experience.

The Limits and the Next Frontier

All models are simplifications, and even the most advanced world model is still a shadow of reality. The world shifts faster than any simulation can adapt. World models hallucinate, overfit, and misinterpret context — much like humans with bias and intuition. But even imperfect models can be powerful tools for alignment: between perception and truth, intent and impact.

Researchers like LeCun, Schmidhuber, and the teams at DeepMind are exploring architectures (JEPA, Dreamer, Genie) that enable machines to learn continuously from their environment, refining their world models the way children do — through play, prediction, and correction. It’s a vision of AI that doesn’t just compute but grows.

The next frontier is not just bigger models, but better models of the world — ones that learn cause and effect, adapt to change, and stay grounded in reality.

Asimov’s Mirror

Asimov’s fiction gave us robots wrestling with morality; our reality gives us agents wrestling with understanding. His positronic brains modeled the world to serve human ends. Ours must now do the same — but with humility.

Asimov’s Robot stories ultimately asked whether understanding the world made robots more human — or humans more predictable. That same question echoes today. The more our machines model us, the more we must model ourselves.

“His robots learned to imagine the consequences of their actions. We’re finally teaching ours to do the same.”

Every great civilization — and every great AI — begins with a world model.

Reflections on AI: Context and Memory – The Gateway to AGI

Introduction: Why AGI is Different from Narrow AI

Today’s frontier models are wonders of engineering. They can write code, draft legal arguments, and create poetry on command. But for all their power, they are fundamentally transient. Once a session ends, the model resets. The insights, the rapport, the shared understanding—it all vanishes. It’s like having a brilliant conversation with someone who develops amnesia the moment you walk away.

This is the core limitation of Narrow AI. Artificial General Intelligence (AGI), the long-sought goal of creating a truly autonomous and adaptive intelligence, requires something more: persistence. AGI must have the ability to remember, adapt, and apply knowledge not just within a single conversation, but over time. True intelligence emerges when raw predictive power is paired with persistent context and memory.

A Brief History: AI Without Memory

The quest for AI has been a story of brilliant but forgetful machines. Each era pushed the boundaries of computation but ultimately fell short of creating lasting intelligence.

  • Expert Systems (1980s): These were the first commercial AIs, functioning like digital encyclopedias. They operated on vast, hard-coded rule-based systems. While effective for specific tasks like medical diagnosis, they had no memory of past interactions and couldn’t learn from experience.
  • Deep Blue (1997): IBM’s chess-playing supercomputer famously defeated world champion Garry Kasparov. It could analyze hundreds of millions of positions per second, a monumental feat of brute-force computation. Yet, each game was a clean slate. Deep Blue had no memory of Kasparov’s style from previous matches; it was a tactical genius with zero long-term continuity.
  • Early Machine Learning (2000s): The rise of statistical models brought pattern recognition to the forefront. These systems could classify images or predict market trends but were narrow and forgetful. A model trained to identify cats couldn’t learn to identify dogs without being completely retrained, often forgetting its original skill in a process known as “catastrophic forgetting.”
  • Modern LLMs: Today’s large language models possess massive context windows and demonstrate emergent reasoning abilities that feel like a step-change. Yet, they remain fundamentally stateless. Their “memory” is confined to the length of the current conversation. Close the tab, and the world resets.

The takeaway is clear: across decades of innovation, AI has lacked true continuity. Context and memory are the missing ingredients.

Context as the Fuel of Intelligence

If intelligence is an engine, context is its high-octane fuel. We can define context as an AI’s active working state—everything that is “in mind” right now. It’s the collection of recent inputs, instructions, and generated outputs that the model uses to inform its next step.

In recent years, context windows have exploded, growing from a few thousand tokens to over a million. Models can now process entire codebases or novels in a single prompt. They are also becoming multimodal, ingesting text, images, and audio to build a richer, more immediate understanding of the world.

A useful analogy is to think of context as RAM. It’s temporary, volatile, and absolutely vital for processing the task at hand. But just like RAM, its contents expire. Without a mechanism to save that working state, intelligence resets the moment the power is cut.

Memory as the Backbone of Learning

This is where memory comes in. Memory is the mechanism that transforms fleeting context into lasting knowledge. It’s the backbone of learning, allowing an intelligence to build a persistent model of the world and its place in it.

We can draw parallels between human and AI memory systems:

  • Short-Term / Working Memory: This is analogous to an AI’s context window—the information currently being processed.
  • Episodic Memory: This involves recalling specific experiences or past events. In AI, this is mirrored by storing conversation histories or specific interaction logs in vector databases, allowing a model to retrieve relevant “memories” based on semantic similarity.
  • Semantic Memory: This is generalized knowledge about the world—facts, concepts, and skills. This is what LLMs are pre-trained on, but the goal of continual learning is to allow models to update this semantic memory over time without starting from scratch.

Memory is what allows an AI to move beyond one-off transactions. It’s the bridge that connects past experiences to present decisions, enabling true learning and adaptation.

Why Context + Memory Together Are Transformational

Separately, context and memory are powerful but incomplete. It’s their synthesis that unlocks the potential for higher-order intelligence.

  • Context without memory is a clever amnesiac. It can solve complex problems within a given session but can’t build on past successes or learn from failures.
  • Memory without context is a passive archive. A database can store infinite information, but it can’t reason about it, apply it to a new problem, or act on it in real time.

When fused, they create a virtuous cycle of adaptive, continuous reasoning. The system can hold a real-time state (context) while simultaneously retrieving and updating a persistent knowledge base (memory). A better analogy combines the previous ones: context is the CPU + RAM, processing the present moment, while memory is the hard disk, providing the long-term storage that gives the system continuity and depth.

Case Study: From Jarvis to Real-World Architectures

Perhaps the best fictional illustration of this concept is Tony Stark’s AI assistant, Jarvis. While still science fiction, the principles that make Jarvis feel like a true AGI are actively being engineered into real-world systems today.

  • Context as Real-Time Awareness: Jarvis’s ability to multitask—monitoring the Iron Man suit, Stark Industries, and geopolitical threats simultaneously—is a conceptual parallel to the massive context windows of modern models. For example, Google’s Gemini 1.5 Pro demonstrated a context window of 1 million tokens, capable of processing hours of video or entire codebases at once. This mirrors Jarvis’s immense capacity for real-time situational awareness.
  • Memory as Persistent Knowledge: Jarvis’s deep memory of Stark’s habits, history, and humor is now being approximated by Retrieval-Augmented Generation (RAG) architectures. As detailed in research from Meta AI and others, RAG systems connect LLMs to external knowledge bases (like vector databases). When a query comes in, the system first retrieves relevant documents or past interactions—its “memories”—and feeds them into the model’s context window. This allows the AI to provide responses grounded in specific, persistent information, much like how Jarvis recalls past battles to inform present strategy.

The takeaway is that the magic of Jarvis is being deconstructed into an engineering roadmap. The fusion of enormous context windows (the “present”) with deep, retrievable knowledge bases (the “past”) is the critical step toward creating an AI with a genuine sense of continuity.

Architectures Emerging Today

The good news is that we are moving from science fiction to engineering reality. The architecture for persistent AI is being built today.

  • Extended Context Windows: Models from companies like Anthropic and Google are pushing context windows to a million tokens and beyond, allowing for much longer and more complex “sessions.”
  • Memory-Augmented Agents: Frameworks like LangChain and LlamaIndex are creating systems that allow LLMs to connect to external vector databases, giving them a persistent long-term memory they can query.
  • Hybrid Neuro-Symbolic Models: Researchers are exploring models that blend the pattern-recognition strengths of neural networks with the structured, logical reasoning of symbolic AI, creating a more robust framework for knowledge representation.
  • Continual Learning: The holy grail is developing agents that can continuously update their own parameters in real time based on new information, truly learning as they go without needing to be retrained.

How Close Are We? An Opinion

While the architectural components for a persistent AI are falling into place, it’s crucial to distinguish between having the blueprints and having a finished skyscraper. We are in the early stages of the construction phase—the foundation is poured and the first few floors are framed, but the penthouse is still a long way off.

  • The Good News: Concepts like Retrieval-Augmented Generation (RAG) and massive context windows have moved from research papers to practical frameworks in just a few years. We now have the basic tools to give models a semblance of long-term memory. This is a monumental step forward. This rapid acceleration from theory to practice is a clear example of the Law of Accelerating Returns, a concept I explored in a previous post.
  • The Hard Reality: The primary challenge is no longer about possibility but about integration and autonomy. Current RAG systems are often brittle and slow. Determining what information is truly “relevant” for retrieval is a complex challenge in itself. More importantly, we haven’t solved continual learning. Today’s agents “read” from their memory; they don’t truly “learn” from it in a way that fundamentally reshapes their internal understanding of the world. They are more like interns with access to a perfect library than seasoned experts who have internalized that library’s knowledge.

We are likely years, not months, away from systems that can learn and adapt autonomously over long periods in a way that truly resembles human-like persistence. The scaffolding is visible, but the hard work of seamless integration, optimization, and achieving genuine learning has only just begun.

The AGI Threshold

When these pieces come together, we will begin to approach the AGI threshold. The key ingredients of general intelligence can be framed as follows:

  1. Context: The ability to reason effectively in the present moment.
  2. Memory: The ability to persist knowledge and learn across time.
  3. Agency: The ability to act on that reasoning and learning to achieve goals and improve oneself.

Crossing the threshold from Narrow AI to AGI won’t be about a single breakthrough. It will be an evolution toward systems that can “live” across days, months, or even years, learning continuously from their interactions with the world and with us.

Risks & Ethical Dimensions

Of course, creating AI with perfect, persistent memory introduces profound ethical challenges.

  • Privacy: What should an AI be allowed to remember about its users? A system that never forgets could become the ultimate surveillance tool.
  • Bias and Malice: False or malicious memories, whether introduced accidentally or deliberately, could permanently shape an AI’s behavior in harmful ways.
  • The Importance of Forgetting: Human memory decays, and this is often a feature, not a bug. Forgetting allows for forgiveness, healing, and moving past trauma. A perfectly eidetic AI may lack this crucial aspect of wisdom.
  • Governance: This new reality will demand robust governance frameworks, including clear audit trails, explicit user consent for memory storage, and a “right to be forgotten” that allows users to wipe an AI’s memory of them.

Conclusion: Context + Memory as the True Gateway

For years, the race toward AGI has been framed as a race for scale—bigger models, more data, more compute. While these are important, they are not the whole story. The true gateway to AGI will not be opened by raw computational power alone, but by the development of persistent, contextual intelligence.

The Jarvis analogy, once pure fantasy, is now a design specification. It shows us what’s possible when an AI can remember everything yet act on that knowledge with immediate, contextual awareness. The great AI race of the next decade will not be about building the biggest brain, but about building the one with the best memory.

Reflections on AI: AI is Eating Software that is Eating the World

In the summer of 2011, Marc Andreessen published a seminal essay in the Wall Street Journal that defined the next decade of technology and business: “Why Software Is Eating the World.” His argument was as elegant as it was prophetic. He posited that we were in the middle of a fundamental economic shift, where software companies were poised to invade and overturn established industry structures. This wasn’t a cyclical tech bubble, he argued, but a tectonic change in how businesses are built and operated. Nearly every industry was becoming a software industry, and those that failed to adapt would be “eaten.”

He was right. Software did eat the world. We watched as Netflix, a software company, devoured Blockbuster. We saw Amazon, a software company with warehouses attached, consume traditional retail. The arc was clear: build a software-centric model and disrupt the incumbents.

That essay landed with particular force for me. My second daughter Brooklyn had just been born, and inspired by the dawn of the mobile era, I had quit my job to launch an augmented reality startup. It was a time of immense learning and, as my wife Sarah loves to remind me, questionable timing. We were building on the new wave, combining sensors on the new iPhones with marketing and gaming. While the startup ultimately didn’t go the distance, the experience was invaluable. It taught me about the immense weight of the word “disruption” and the grit required to survive it—whether you’re the one disrupting or the one being disrupted, both are incredibly difficult.

For over a decade, Andreessen’s thesis was the undisputed law of the digital jungle. But a new, apex predator has emerged. The cycle of disruption has accelerated to a dizzying pace, and in a deeply meta twist, the disruptors from the past two decades are now the ones being disrupted.

AI is now eating the software that is eating the world.

An abstract depiction of the Earth being engulfed by a colorful, swirling cosmic force, symbolizing disruption and transformation.

What Disruption Really Means

Andreessen’s essay heralded a wave of software-driven change that felt unstoppable. But what does it actually feel like to be on the receiving end of that disruption? It’s not just about a new competitor; it’s about the ground shifting beneath your feet.

  1. Loss of Control Over the Value Chain: Disruptors rewire how value is delivered—removing steps, middlemen, or entire business models before you even notice.
  2. Customer Expectations Shift Overnight: When a new player offers instant, personalized, cheaper, or more delightful experiences, your “good enough” becomes “not even close.”
  3. Margin Compression Becomes Existential: Disruptive technologies often enable radically lower cost structures. Software doesn’t sleep, unionize, or take vacations. Your 20% margin looks quaint next to their 80%.
  4. Your Competitive Moat Turns Into a Puddle: Scale, legacy systems, and brand used to be strengths. But disruption turns those into anchors, slowing adaptation while nimble upstarts sprint past.
  5. Innovation Moves Outside the Building: Disruption often comes from adjacent industries or unexpected entrants. Amazon didn’t ask bookstores for permission; OpenAI didn’t wait for Google to modernize.
  6. Talent Starts Leaving for the Cool Kids: The best engineers, designers, and product thinkers want to build the future, not maintain the past. When you’re being disrupted, your best people become a leading indicator of decline.
  7. It Feels Like a Tech Problem, But It’s Actually a Culture Problem: Many incumbents respond by buying new software or hiring consultants. But the real challenge is rewiring how they think, decide, and act.
  8. You’re Not Competing With Companies—You’re Competing With Capabilities: AI, APIs, open-source, no-code… disruptive tools are making individuals and small teams exponentially more powerful.

The Disruptors Disrupted: Modern Examples

Andreessen gave us the classic examples: Blockbuster falling to Netflix, traditional retail to Amazon, Kodak to digital photos. But the most fascinating part of this new wave is seeing the disruptors of that era facing their own existential threats.

Google vs. ChatGPT: The Search for Answers

Google built an empire on software that indexed the world’s information and presented it as ten blue links. SEO became the science of ranking on that list. But AI is eating that model. While Google still dominates the raw volume of search, a significant behavioral shift is happening faster than anyone predicted.

According to a recent Wall Street Journal article, AI-powered search is growing more quickly than expected, with traffic to leading AI chatbots like ChatGPT and Perplexity AI surging. One analytics firm, Similarweb, noted that combined traffic to the top 10 AI chatbots grew 34% in the first part of this year alone. This isn’t just a niche trend; it’s a mainstream migration for certain types of queries. Users are flocking to conversational AI for complex, informational tasks—research, brainstorming, coding help, and problem-solving. We see real-world examples of this constantly. A user on Quora recounted struggling to find a half-remembered book using Google; ChatGPT found it instantly from a vague, partially incorrect description. This is a fundamentally new type of search—one based on context and conversation, not just keywords. The game is shifting from Search Engine Optimization (SEO) to Answer Engine Optimization (AEO) and Generative Engine Optimization (GEO). Users no longer just want a list of links to search through; they want the answer.

Uber vs. Waymo: The End of the Driver

Uber used software to disrupt the taxi industry by creating a massive, efficient marketplace for drivers. Their former CEO pushed hard into autonomous driving, recognizing the existential threat. But in a classic innovator’s dilemma, the new leadership divested from that costly, long-term bet to focus on near-term profitability. Now, companies like Waymo and Tesla are rolling out robotaxi services that threaten to eat Uber’s core business model by removing the driver—and their associated costs—entirely.

The IDE vs. AI: The Changing Nature of Code

The very process of building software is being consumed. For decades, developers have relied on Integrated Development Environments (IDEs) like Microsoft’s Visual Studio or JetBrains’ IntelliJ IDEA. These were the definitive software-building tools. Now, AI-native environments like Cursor and Replit are upending that. They don’t just help you write code; they write it with you and for you.

This has profound implications. What happens when the cost to build software approaches zero?

  • Explosion in Software Supply: Software is no longer a scarce, expensive resource—it becomes ubiquitous infrastructure.
  • Margins Collapse for Custom Development: Dev agencies, especially those competing on cost, face commoditization unless they move up the value chain to strategy and architecture.
  • Shift from “Build” to “Compose”: Software creation becomes more about orchestration and configuration than hard engineering.
  • Rise of Citizen Developers: Domain expertise becomes more valuable than knowledge of syntax.
  • Incumbent Software Vendors Get Eaten: Legacy vendors must reinvent themselves or be disrupted out of existence.
  • Regulation Struggles to Keep Up: Governance models must evolve—fast.
  • Software Becomes Embedded Everywhere: The world becomes hyper-personalized and hyper-automated.
  • Engineering Roles Evolve: The “10x engineer” becomes the “10x AI collaborator.”
  • Economic Leverage Shifts: Distribution, branding, and user insight become more valuable than the underlying code.
  • Everything Speeds Up: Strategic agility becomes the only true competitive advantage.

The Crumbling Moats of Enterprise Software

Every traditional enterprise software vendor is seeing their moats dry up. For years, the high cost of replacement was a powerful defense. But that changes as monolithic platforms give way to a diverse ecosystem of best-of-breed SaaS players. Data is becoming more accessible through APIs, and workflows are easier to replace. Additionally, companies are getting wiser to the enterprise sales games. Just because a vendor bought a company doesn’t mean its technology is well-integrated into the platform. We will see the emergence of AI-native enterprise platforms that are built from the ground up to automate, predict, and advise—making their predecessors look like relics.

The Existential Question for Every Company

In 2011, Andreessen argued that every company needed to become a software company to survive. In 2025, the stakes are even higher. What happens to companies—even the software-savvy ones—that don’t evolve into AI-native organizations?

The bottom line is they risk becoming irrelevant, uncompetitive, or extinct. That isn’t a threat; it’s the emerging reality.

  • They get outpaced by faster, cheaper, smarter rivals.
  • Innovation freezes while bureaucracy expands.
  • Knowledge work gets bottlenecked in human siloes.
  • Margins shrink as defensibility moats evaporate.
  • Top talent leaves for companies where AI is an amplifier, not a threat.
  • Customers expect magic, but they deliver forms and call centers.
  • Legacy infrastructure becomes an existential debt.
  • Strategy becomes guesswork without the real-time data fabric to train and validate AI.

The imperative has evolved. In 2011, the call was to become a software company. Today, every company must become an AI company. This isn’t about buying a few AI tools or launching a chatbot. It’s about fundamentally re-architecting the business around data, intelligence, and automation. It means fostering a culture that thinks in terms of models, probabilities, and feedback loops, and embedding intelligent capabilities into the core of every product, service, and process.

Why Now? The Perfect Storm for Disruption

This isn’t happening in a vacuum. A confluence of factors has created a perfect storm for this AI-driven disruption. As I explored in my previous posts on Accelerating Returns and the Stochastic Era, we’ve hit a critical inflection point.

  1. Foundation Models Changed the Game: General-purpose models like GPT can now write, debug, and refactor software, crossing a critical capability threshold.
  2. OpenAI (and others) Made It Accessible: The interface to intelligence is now an API call, not a research lab.
  3. Software Was Ripe for Disruption: Ironically, much of the software world had become bloated, slow, and ripe for a leaner, smarter alternative.
  4. Cheap Cloud + Ubiquitous GPUs = Acceleration: The hardware finally caught up with the ambition.
  5. We Finally Have Enough Training Data: The internet created the massive corpus of code, text, and images needed to train these models.
  6. Human-Machine Collaboration Just Got Real: The technology is not just smart—it’s usable, amplifying human potential across every role.
  7. Software Economics Just Collapsed: When AI can write the code, the cost to create software plummets, and the speed to ship skyrockets.

The Great Leapfrog Moment

One of the wildest things about this era? It’s a leapfrog moment. You don’t need to be the biggest, richest, or most established player anymore—you just need to be the fastest learner.

A scrappy team with a bold vision can outmaneuver giants. The stack is flatter, the tools are open, and the pace of change is brutal. Where you started matters less than how fast you move. This isn’t just for startups. Older companies can leapfrog, too. In fact, they might be in the best position—if they’re willing to change. They have the customers, the data, the brand, and the operational knowledge. What they often lack is urgency and imagination.

The age of the “5-year digital roadmap” is over. The game now is a chaotic, high-stakes parkour race.

Conclusion

In his 2011 essay, Marc Andreessen famously wrote that he was optimistic about the future growth of the economy, predicting it would be driven by these new software-based disruptors. He encouraged every company to embrace this change, to become a software company.

Today, I am also incredibly optimistic, but for a different reason. We are witnessing a second, more profound wave of disruption that is unlocking human potential on an unprecedented scale. The ability to create, to solve problems, and to build is being democratized by AI. Companies that embrace this new reality—that become AI-native at their core—will not only survive but will define the next era of innovation and value creation.

More and more major businesses and industries are being run on artificial intelligence and delivered as intelligent, automated services. The smart ones will be AI-first. The rest will be dinner.

Reflections on AI: The Stochastic Era

I’ve always loved jazz and improvisational music. My wife, Sarah, appreciates the perfect, tight structure of a three-minute song, and I get it. There’s a real beauty in that precision. But for me, the magic happens in the exploratory freedom of a 10, 15, or even 25-minute musical journey. It’s about letting go of a rigid plan to discover something new and amazing in the moment.

I was thinking about this recently, remembering a weekend back in August of 1996. I was standing on a decommissioned Air Force base in Plattsburgh, New York, with three good friends and a huge smile on my face. We were at The Clifford Ball, Phish’s first festival, and the band was on fire. During the second set of the second night, they launched into “Run Like An Antelope.” The jam that followed was pure improvisational genius—a high-energy, tight-but-loose exploration that broke free from the song’s structure to create something utterly unique and unrepeatable. The entire festival was like that, a masterclass in creative freedom.

I’m a firm believer in what Steve Jobs called standing at the “crossroads of technology and the liberal arts.” That Phish jam is a perfect example of the artistic side: letting go of a rigid structure can lead to something far more profound. It feels counterintuitive, but for my entire career in technology, I’ve seen the other side—a world built on perfect, deterministic machines. Now, we’re standing at a new crossroads, and the same principle of letting go is about to change everything.

Steve Jobs famously said, ” — it’s technology married with liberal arts, married with the humanities, that yields us the results that make our hearts sing … ”

A Jarring Shift in Thinking

For as long as I’ve been a software engineer and a technology leader, computers have been defined by their deterministic nature. They are perfect, logical calculators. Input A always produces Output B. 2 + 2 will always equal 4. But we are now entering a new era: the Stochastic Era.

The most powerful large language models today, the ones that can generate art, write poetry, and are changing our world, are fundamentally not deterministic. At their core, they are probabilistic engines making sophisticated guesses. Letting go of rigid structure has allowed for the room for what feels like creativity. This is a massively jarring shift in thinking. How can this randomness—this seeming imperfection—be the essential ingredient for building true, human-like intelligence?

From Certainty to Probability: What is Stochastic Thinking?

To understand this shift, we need to contrast two mindsets.

  • Deterministic Thinking: This is like following a precise recipe to bake a cake. You use the exact same ingredients and instructions every time, and you get the exact same cake. It’s predictable and reliable.
  • Stochastic Thinking: This is like a skilled chef improvising a meal. They have a deep understanding of ingredients and techniques, but they create a dish based on what’s fresh and available. The meal is different every time, but it’s creative, adapted, and often brilliant.

It’s crucial to understand that this isn’t just chaos or random noise. It’s principled randomness. A stochastic system uses probability distributions to make the best possible guess based on the vast amount of data it has learned from.

The Engine of Modern AI: How LLMs Actually Work

The generative AI revolution we are living through was ignited by a single research paper. In 2017, researchers at Google published a paper titled “Attention Is All You Need.” It introduced a new architecture called the Transformer, which is the blueprint for every modern Large Language Model (LLM), from ChatGPT to Gemini.

Before the Transformer, AI models processed language sequentially, one word at a time, often forgetting the context of earlier words. The Transformer’s breakthrough was a mechanism called self-attention, which allows the model to look at all the words in a sentence at once and weigh their relevance to each other. This enabled a far deeper understanding of context and, crucially, allowed for massive parallelization in training.

Stochastic thinking is not just an add-on to this architecture; it is its fundamental operating principle.

  1. The Core Engine: A Probabilistic Word Predictor. At its heart, an LLM is predicting the most probable next word in a sequence. Its creativity comes from the fact that it doesn’t always pick the #1 most likely word. Instead, it samples from a distribution of likely candidates, allowing for variety and novelty.
  2. Controllable Randomness: Temperature and Top-P Sampling. We can control this randomness with parameters. Temperature acts as a creativity dial—low temperature makes the AI more factual and predictable, while high temperature makes it more creative and surprising. Top-P sampling provides another lever, telling the model to only consider a set of the most likely words.
  3. The Learning Process: Stochastic Gradient Descent. Even the training process is stochastic. It would be impossible to learn from the entire internet at once. Instead, models learn using Stochastic Gradient Descent (SGD), where they take a small, random batch of data, learn from it, and adjust. This random sampling makes learning efficient and helps the model generalize its knowledge.

The Wall of Determinism: Why Old AI Hit a Limit

An old vintage pickup truck parked on a dirt road in a scenic landscape with grasslands and rolling hills under a colorful twilight sky.

For decades, AI research focused on rule-based “expert systems.” This deterministic approach could never lead to AGI for a few key reasons:

  • The Real World is Messy: The world isn’t a clean set of IF-THEN statements. It’s ambiguous, nuanced, and unpredictable.
  • Brittleness: Rule-based systems are brittle. They fail the moment they encounter a situation not explicitly covered by their hand-crafted rules.
  • The Creativity Problem: A deterministic system can only follow its programming. It can never create something truly novel or surprising.

The Bitter Lesson

In 2019, AI pioneer Rich Sutton wrote a now-famous essay called “The Bitter Lesson.” His central point was that, in the long run, general-purpose methods that leverage massive computation (like learning and search) will always outperform systems where humans try to hand-craft their knowledge.

This is the ultimate validation of the stochastic approach. Instead of trying to teach an AI all the grammatical rules of English, we let a general learning algorithm discover the patterns for itself from trillions of words. This is exactly how LLMs work, and it’s a lesson that connects directly to the ideas in my previous post on the Law of Accelerating Returns. When you combine The Bitter Lesson (let computation do the work) with the Stochastic Engine of LLMs and place it on the exponential curve of Accelerating Returns, you get the explosive, transformative moment in AI that we are witnessing right now.

How Stochasticity Unlocks Intelligence

This new approach is the bridge to AGI because it enables capabilities that were impossible before:

  1. Creativity and Exploration: Randomness allows an AI to explore novel combinations of ideas and generate content that has never existed before.
  2. Robustness and Adaptability: A probabilistic model can handle the uncertainty of the real world, making informed guesses instead of breaking down.
  3. Efficient Learning: It is the only way to effectively learn from the planet-scale datasets required to achieve general intelligence.

I saw early glimpses of this in my career. I had the incredible opportunity to be mentored by Steve Kirsch, the founder of Infoseek and a true tech pioneer. We worked together on algorithms for blocking spam for major clients like Yahoo Mail. The techniques we used were essentially early stochastic models, employing Bayesian probability to “guess” if an email was spam based on patterns, rather than relying on rigid rules. That company was later sold to Proofpoint, but the core lesson about the power of probabilistic systems stayed with me.

Even today, my role as CTO for O2E Brands is a stochastic exercise. I’m constantly weighing probabilities—the likelihood of a project’s success, market adoption, potential risks—to make the best strategic bets with the available data. It’s never about one certain answer.

The Art of the Guess

An abstract digital artwork featuring swirling purple and golden lines set against a dark blue background, reminiscent of intricate neural connections.

Looking ahead, these non-deterministic, stochastic models will power the next wave of systems on the path to AGI, from autonomous agents that can navigate unpredictable environments to scientific AIs that can form novel hypotheses.

The journey to AGI isn’t about building a faster, more powerful calculator. It’s about building a more sophisticated and intuitive “guesser.” We’ve spent a century trying to make machines perfectly logical. It turns out, to make them truly intelligent, we first have to teach them the art of probability. The messy, jarring concept of randomness is not a bug—it’s the feature that will finally get us to AGI.

Thank you for reading. Leave a comment if you have thoughts or comments.

Reflections on AI: The Law of Accelerating Returns

Looking back on my 25 years in technology, I can’t help but feel an immense sense of gratitude. It has been an amazing ride, and I feel incredibly lucky to have had a front-row seat—and often, a place on the stage—for some of the most profound technological shifts in human history.

My career has spanned the dot-com boom, the rise of enterprise software, the mobile revolution, the shift to the cloud, and now, the dawn of the AI era. Each wave was built on the last, creating a foundation for the next leap forward. The amount of change we’ve packed into the last quarter-century is staggering. It makes you wonder: if this is what we saw in the last 25 years, where could we possibly be in the next 25?

It feels impossible to predict, but some people make it their life’s work. One of the most compelling thinkers on this topic is the inventor and futurist, Ray Kurzweil.

The Prophet of Exponential Growth

Ray Kurzweil is a towering figure in computer science. He’s an author, inventor, and one of the most prominent futurists of our time. He’s probably one of the oldest living computer scientists and even worked with the legendary Marvin Minsky at MIT—an institution I’ve grown particularly fond of since my daughter started attending.

Ray Kurzweil speaking at a technology conference, smiling and wearing a dark blazer over a plaid shirt.

Kurzweil is best known for his mind-bending books like The Singularity Is Near and his brand new follow-up, The Singularity Is Nearer. In them, he argues that humanity is approaching a “Singularity”—a point in the near future where technological growth becomes uncontrollable and irreversible, resulting in unforeseeable changes to human civilization. He predicts we will achieve Artificial General Intelligence (AGI), an AI that can understand or learn any intellectual task that a human being can, by 2029, and that the Singularity itself will occur around 2045.

He frames this journey through his “Six Epochs of Evolution”:

  1. Physics and Chemistry: Information in atomic structures.
  2. Biology: Information in DNA.
  3. Brains: Information in neural patterns.
  4. Technology: Information in hardware and software.
  5. The Merger: The fusion of technology and human intelligence.
  6. The Universe Wakes Up: The point where intelligence saturates the cosmos.

According to Ray, we are living in the 5th Epoch right now.

Why We Fail to See the Future

Kurzweil’s predictions can feel like science fiction because our brains are wired to think linearly. We struggle to grasp the power of exponential growth. Think about it: if you take 30 linear steps, you end up 30 meters away. If you take 30 exponential steps (doubling each time), you travel over a billion meters—enough to circle the Earth 26 times!

A Futurist, in my mind, is someone who can intuitively grasp exponential growth. They don’t just see the next step; they see the curve. This understanding is the key to Kurzweil’s central thesis: The Law of Accelerating Returns.

The Law of Accelerating Returns

This law is the engine driving us toward the Singularity. It states that the rate of technological advancement—and evolution in general—is not linear, but exponential. This happens because of powerful feedback loops of innovation. Each new generation of technology provides better tools to create the next generation, which is then created faster and more efficiently.

Think of it as a form of societal reinforcement learning. We create a tool, learn from it, and use that knowledge to build a better tool, accelerating the cycle. Moore’s Law, which famously predicted the doubling of transistors on a chip every two years, is just one famous example of this law in action. But Kurzweil argues it applies to all information-based technologies. The law of accelerating returns is happening now, and it has been for a long time. The evidence is all around us.

Mary Meeker’s “Trends, Artificial Intelligence” Report

For decades, anyone in tech has eagerly awaited Mary Meeker’s annual “Internet Trends Report.” She is a bit of a celebrity in our circles, first publishing her report in 1995 and updating it yearly until 2019. Just a few weeks back, she and her team at BOND Capital dropped a new gem: “Trends, Artificial Intelligence”.

Reading through the 340-slide deck, I couldn’t help but see it as a stunning validation of Kurzweil’s Law of Accelerating Returns. The AI pace of change is off the charts. The compounding effect of AI technology, its ecosystem, and user adoption is completely unheralded.

The arc of Meeker’s deck proves that this AI wave is built upon all the technology that came before it: computing power, vast datasets, advanced algorithms, and global communications networks. It’s the ultimate feedback loop.

Here are a few slides that stood out to me:

  • Slide 20: Google Disruption: The pace at which new AI-native search products are challenging Google is breathtaking. This isn’t a slow, decade-long battle; it’s happening in months.
A graph comparing annual searches for ChatGPT and Google from 1998 to 2025, highlighting that ChatGPT reached 365 billion searches in just 2 years, while Google took 11 years.
  • Slide 26: Wisdom, Not Just Knowledge: “Wisdom” is why products like ChatGPT and Gemini Search will win. Traditional search gives you knowledge (a list of links). AI-powered search provides wisdom—synthesized, contextual answers. It’s a fundamental upgrade.
Slide featuring a quote by Martin H. Fischer: 'Knowledge is a process of piling up facts; wisdom lies in their simplification.' The slide is attributed to BOND and highlights the theme of knowledge distribution over six centuries.
  • Slide 43: Passing the Turing Test: We are already starting to pass the Turing Test in various modalities. AI is becoming indistinguishable from human-created content, a milestone Kurzweil predicted for 2029. We’re right on schedule, if not ahead.
An image illustrating a Turing Test conversation between two witnesses, A and B, showcasing the realism of AI-generated responses compared to human dialogue.
  • Slide 302: Waymo vs. Lyft: In San Francisco, Waymo’s autonomous vehicles have surpassed Lyft in market share. Think about that. A technology that was science fiction a decade ago is now out-competing an established, human-powered incumbent in a major city. The disruption is real and it is happening now.
Graph showing the market share of Waymo's fully-autonomous vehicles compared to Uber and Lyft in San Francisco over a period from August 2023 to April 2025.

Buckle Up

As someone who has spent a career building things, I find it impossible to look at Kurzweil’s theories and Meeker’s data with anything but immense optimism. This isn’t a moment for fear; it’s a moment for builders. The scale of this transformation is unlike anything we’ve ever seen, presenting an unprecedented opportunity to redefine what’s possible and build a better future.

There will be fear and resistance; there always is. This is not a new phenomenon. When the printing press emerged in the 15th century, the scribal class and religious authorities feared a loss of control, calling it a technology that would spread dangerous ideas to the masses. In the 19th century, the Luddites famously smashed the automated looms that threatened their craft and livelihoods. And in our own lifetimes, people protested the introduction of calculators in schools, fearing students would forget basic math. The AI revolution is, of course, something much bigger, but the pattern of anxiety and opposition is the same. We cannot turn back the clock. The feedback loop of innovation is spinning faster than ever.

So much change is ahead of us. Buckle up.

Artificial Intelligence & Me

In the vast realm of science fiction, few names shine as brightly as Isaac Asimov. As a young reader, I found myself captivated by the intricate universes he wove in novels like Foundation and I, Robot. These stories, where artificial intelligence and robotics played pivotal roles, ignited a spark in me that would later pave the path of my own career. Today, as I delve into the intricate tapestry of my journey with artificial intelligence, it’s impossible not to look back at those early days of wonder, where Asimov’s words provided not just entertainment, but inspiration for a future I had yet to envision.

It’s official, I have secured Renato.mascardo.ai.  This has moved up the value of my blog by 10x.  Kidding. Kinda. Artificial Intelligence has created both rational and irrational reactions.  One would argue AI has propped up the stock market when everything AI related has killed while everything else looks “eh”.  Seven of the top 10 largest companies in the United States by market cap have major AI value propositions.

My career has seen its waves of innovation.  I am so grateful for having been part of such a fruitful period for computer science.  It has been such a fun ride and it doesn’t seem to be letting up.

As one would expect, each wave builds upon the next and is a combination of previous achievements and learnings.  AI is a combination of data, copious amount of distributed compute, innovation in gpu chips and foundational data / data science / machine learning / deep learning practices.

So, here are a few common questions I’ve better getting recently about AI —

What is you perspective on Artificial Intelligence? 

Artificial Intelligence represents both an opportunity and existential risk for companies.  Similar to other technology waves, something very fundamental is shifting under existing companies.  This shift can create new markets and provide opportunities to wedge into existing markets.  We should see new interaction models and all new levels of efficiency with AI.  

This is where it’s critical to be in continuous first principle thinking.  Assumptions should be challenged. Experimentation should be the norm.  Action should be taken. 

What experience do you have with machine learning / deep learning / Artificial Intelligence? 

  • Mercury Interactive / HP Software (2004) — processed the millions / billions of production alerts to provide an elevated alerting or automate action. 
  • Abaca Technology (2007) — created a totally new AI approach for handling email SPAM.
  • DigitalGlobe (2010) — built one of the first and largest at the time GPU based supercomputers that was used to run advanced algorithms for higher level processing on digital satellite imagery.  These algorithms did things like cloud cover detection, change detection and image feature recognition.
  • Atari (2012) — built AI based real-time adjustment of gameplay based on individual progress through the game. 
  • Rosetta Stone (2014) — built AI based content creation platform.   
  • Recurly (2018) — built AI based Dunning management solution. 

What will the next 5-10 years of AI innovation bring us? 

Buckle up.  The pace of AI improvements is fierce and everyone wants to get there first.  Watch out for the hyperbole and emotions through the technology adoption life cycle but rest assured we’ll see the shift on the other side.  This feels like it’s going to go faster than we think.

Thank you for reading.  Please drop me a note if you have a comment.  I appreciate you. 

-rjm