The Holiday Break That Broke the Software Development Lifecycle

The 2025 holiday break was supposed to be slow. It has been a long year of hard work and I needed to take a break.

My parents are older now. Cruises are their speed. Elevators instead of staircases. Early dinners. My dad triple-checking the daily schedule like it’s a mission briefing. My mom insisting we all order dessert because “it’s free”.  Time feels different when you realize there are fewer of these trips ahead than behind.

I told myself I would unplug. I even announced it at dinner the first night. “No laptop this week.” My dad nodded approvingly. My mom smiled like she didn’t quite believe me.

I didn’t last very long.

On the second afternoon, I slipped away with my laptop and headphones and found a quiet spot at the back of the boat. Espresso in hand. Salt air hanging thick. The low mechanical hum of the engines vibrating through the deck. The wake trailing behind us like a white scar across a blue horizon.

I opened my terminal.

I started catching up.

Claude Code, Claude 3.5 Sonnet. GPT-4.1 and GPT-4o. Early Codex CLI previews. Terminal-native agents gaining traction. IDEs that no longer just suggested code but executed meaningful chunks of it. Parallel task execution in the cloud. Multi-step tool use that didn’t collapse after the second function call.

At first, it felt like noise. 2025 had trained us to expect noise. But then I started testing.

I delegated something non-trivial. It finished. Cleanly.

I handed off something messier. It reasoned through it. Returned with structure.

I felt that strange combination of excitement and unease — the same feeling I had when I was working on code complete and refactoring in Borland JBuilder. Felt like magic.

This wasn’t incremental improvement.

Something had crossed a line.

The Delegation Threshold

There is a moment in any technological system when progress stops being additive and starts being structural.

The iPhone wasn’t merely a better phone — it reorganized behavior. AWS wasn’t cheaper hosting — it changed who could start a company. ChatGPT didn’t just answer questions — it altered how people thought about knowledge itself.

Over the final months of 2025, engineering crossed a similar boundary.

I call it the Delegation Threshold.  The point at which models become reliable enough to safely delegate meaningful work.  Not perfect. Not autonomous. Not magic.  But dependable enough that you stop hovering over every output.  Before the threshold, AI assisted me.  After the threshold, I found myself assigning work.  That subtle shift — from helper to delegate — reorganizes the system.  It reorganized me.

When Delegation Becomes Real

Individually, the releases were impressive.  Long-context windows that could actually be trusted. Tool use that didn’t unravel after the second function call. Parallel execution that made decomposition practical instead of aspirational. IDE-native agents that moved from autocomplete to autonomy.  Collectively, they crossed a reliability boundary.  When you can hand a model a non-trivial task and reasonably expect it to finish, you stop thinking in files.  You start thinking in systems.  Tasks decompose differently. Parallelism becomes natural. Orchestration becomes the new bottleneck.

The wildly stupid term “vibe coding” was making way for “agentic engineering”.

The scarcity shifts.

Code is no longer the constraint. Judgment is.

That realization didn’t feel triumphant.

It felt destabilizing.

Grassroots Discovery

What fascinated me most wasn’t the press releases.  It was what developers started doing once delegation became viable.  Recursive loops appeared — bash scripts feeding output back into models until the system converged. What looked chaotic was actually a crude form of evaluation harnessing.  

How our brains work was surfacing in the methods being discovered.

Agent swarms emerged — one human intent spawning multiple concurrent threads across files, features, and research tasks.  I tried it myself.  It was messy. Threads collided. Context drifted. One agent confidently refactored something another agent had just rewritten.

And yet — it worked often enough that you couldn’t ignore it.

When delegation becomes safe, orchestration becomes the craft.

The community was rediscovering harness engineering before it had a name.

A Collective Synchronization

Why did it feel sudden?

Because thousands of engineers had something rare during that break: uninterrupted time.

No sprint planning. No performance reviews. No backlog theater.

Just space.

Space to test. To push. To hand something real to a model and see if it came back intact.

When everyone returned, the tone had changed.

Group chats weren’t debating whether agents were useful. They were debating how to structure them.

The tools hadn’t just improved. The collective mental model had updated.

The Structural Repricing

When early 2026 reports began quantifying adoption curves and productivity deltas, the data validated the intuition.  Agent usage wasn’t experimental anymore. It was habitual.  I felt it in my own workflow. Tasks I would have blocked off half a day for were now something I delegated while I thought about architecture. And then came formalization.  OpenAI described building a product with near-zero human-written code. Engineers we ren’t typing implementation — they were designing constraints, evaluation loops, and execution environments. Harnesses.

That word stuck with me.

Harness.

Not autopilot. Not replacement.

Harness.

This is what happens after a threshold crossing.

When execution becomes cheap, environment design becomes expensive.

Feature velocity stops being durable advantage. System durability does.

The SDLC Inverted

We like to say we sped up the Software Development Lifecycle.  But sitting on that cruise ship, watching agents complete tasks while my parents took a nap, I realized that wasn’t quite right.

We didn’t speed it up.

We inverted it.

The old world moved in phases: requirements, design, code, test, deploy.

The new world begins with intent and harness design. Agents execute. Evaluation runs continuously. Humans orchestrate and judge.

Phase boundaries blur.

Code stops being the bottleneck. Evaluation and architecture take its place.

That isn’t speed.

That’s reorganization.

The Identity Question

This is the part I keep coming back to.

If writing code is no longer the center of gravity, what is?

I learned to love engineering by writing it line by line. By debugging at 2 a.m. By feeling the satisfaction of something compiling cleanly after hours of friction.

What becomes craftsmanship in a world of delegation? How do junior engineers learn the texture of a system if they rarely touch its internals? Where does pride migrate?

Delegation changes identity, not just workflow.

Every structural shift does.

What Comes Next

Crossing a threshold does not eliminate constraints. It relocates them.

Orchestration complexity can explode. Evaluation can become the new choke point. Reliability can plateau.

But thresholds, once crossed, rarely uncross.

The genie isn’t the story.

The threshold is.

And once crossed, systems reorganize themselves.

On that cruise ship, I closed my laptop and went back inside. My dad sent me a text and asked if I was done working.

“Yeah,” I said.

But I wasn’t done thinking.

We are still early on this journey.

Let’s go.

Move 37 and the Shape of What’s Next

There’s something about forced stillness that creates space for the unexpected.

Molly was a few weeks into recovery from her second ankle surgery—the one where she “won the ankle injury lottery in the worst way possible.” Her boot-clad ankle propped up next to me on the couch, she wasn’t going anywhere. Neither was I. So we did what any reasonable father-daughter duo does when escape isn’t an option: we binged two documentaries about artificial intelligence.

First, AlphaGo—the 2017 film about DeepMind’s AI beating the world champion at Go. Then The Thinking Game—the 2024 documentary that follows DeepMind’s broader quest toward artificial general intelligence, filmed over five years by the same team.

What I didn’t expect was that this double feature would turn into one of the best conversations we’ve ever had.

Molly is a Computational Biology major. I’m a lifelong computer science nerd. Our worlds were about to collide in the best possible way. (Fun aside: one of her friends from MIT—his dad appears in one of the documentaries. We got a good laugh out of that.)

Why Go Matters (And Why No One Thought This Would Happen)

If you’re not familiar with Go, here’s the short version: it’s a 2,500-year-old board game that makes chess look like tic-tac-toe.

Chess has roughly 10^47 possible game states. Go has 10^170. For perspective, there are approximately 10^80 atoms in the observable universe. Go has more possible positions than there are atoms—by a factor of 10^90. Let that sink in.

This isn’t just trivia. It means you can’t brute-force Go. You can’t calculate every possibility the way Deep Blue did against Kasparov in 1997, evaluating 200 million positions per second. That approach simply doesn’t work here. The game is too vast. Too deep.

For decades, the best Go programs were… embarrassing. They played at the level of a decent amateur, routinely getting crushed by club players. Experts confidently predicted that AI beating a professional Go player was 10-20 years away. Some said it might never happen.

In 2016, DeepMind did it anyway.

How AlphaGo Actually Works

Neural Networks: Teaching Intuition

AlphaGo wasn’t programmed with rules about how to play Go. Nobody sat down and wrote “if your opponent plays here, respond there.” That approach had been tried for decades. It didn’t work.

Instead, AlphaGo was trained.

First, it studied millions of games played by human masters. The neural network learned to “see” the board—not as a grid of black and white stones, but as patterns. Shapes. Flows. The kind of intuition that takes a human player decades to develop, encoded in the weights of a neural network.

This is the key insight that changed everything: intuition can be learned. It’s not magic. It’s not some mystical human quality that machines can never possess. It’s pattern recognition at scale.

Reinforcement Learning: Playing Itself

But learning from humans only gets you so far. Humans, after all, are limited.

So after learning from human games, AlphaGo started playing against itself. Millions of games. Billions of moves. Twenty-four hours a day, at speeds no human could match.

This is reinforcement learning—trial and error at superhuman velocity. And here’s the kicker: through self-play, AlphaGo discovered strategies that no human had ever seen. Not because they were wrong. Because we never thought to try them.

The machine had started to see things we couldn’t.

Monte Carlo Tree Search: Guided Exploration

AlphaGo doesn’t evaluate every possible move—that’s mathematically impossible, remember? Instead, it uses its neural network intuition to guide its search.

Think of it like this:

  • Policy Network: “What move looks promising?” (Intuition)
  • Value Network: “Who’s winning from this position?” (Evaluation)
  • Monte Carlo Tree Search: “Let me simulate a bunch of games from here to check.” (Verification)

It’s intuition combined with calculation. The machine equivalent of a grandmaster “feeling” that a move is right, then verifying it with deep analysis.

Human experts have both systems too—the gut and the grind. AlphaGo unified them into something more powerful than either alone.

Move 37: The Moment Everything Changed

Game 2 of the match against Lee Sedol. If you haven’t seen the documentary, go watch it. If you have, you know exactly what I’m about to describe.

Lee Sedol is one of the greatest Go players in history. Eighteen world championships. A player of profound intuition and legendary fighting spirit. He sat across from AlphaGo expecting a battle. He got something else entirely.

Move 37.

AlphaGo places a stone on the fifth line—a move that looks, to the trained human eye, wrong. Commentators are confused. Experts call it a mistake. Lee Sedol leaves the room, visibly shaken. The move violates centuries of accumulated Go wisdom.

And then it wins the game.

Move 37 wasn’t in any textbook. It wasn’t copied from any human game in AlphaGo’s training data. The machine had discovered something new about a game humans have played for 2,500 years.

What does it feel like to watch a machine be creative? I still don’t have a great answer. But I know it changes how you think about intelligence—artificial and otherwise.

From Go to Protein Folding: Where Our Worlds Cross

This is where The Thinking Game picks up the story.

Here’s the thing about DeepMind: they weren’t just trying to win at board games. Go was a proving ground. A demonstration. The real target was always bigger.

Enter AlphaFold. And enter my daughter’s world.

The Protein Folding Problem

Proteins are the workhorses of biology. They do almost everything—carry oxygen in your blood, fight infections, make your muscles contract, replicate your DNA. And every protein is built from a chain of amino acids that folds into a specific three-dimensional shape.

Here’s the critical insight: the shape is the function. A protein’s 3D structure determines what it does. Get the shape wrong, and the protein doesn’t work. Misfolded proteins cause diseases like Alzheimer’s, Parkinson’s, and cystic fibrosis.

The problem? We know the amino acid sequences for over 200 million proteins. But determining the 3D structure experimentally—using X-ray crystallography, cryo-electron microscopy, or nuclear magnetic resonance—is brutally slow and expensive. In 60 years of global scientific effort, we had solved about 170,000 structures.

Predicting how a protein folds from its sequence alone? That was the “50-year grand challenge” of biology. The Mount Everest of molecular science.

CASP: The Olympics of Protein Prediction

Every two years, computational biologists compete in CASP—the Critical Assessment of Structure Prediction. It’s basically the Olympics of protein folding. Teams submit predictions for protein structures that have been experimentally determined but not yet published. Then they get graded.

For years, scores hovered around 40 out of 100 for the hardest targets. Progress was incremental. Slow. Scientists would publish papers celebrating a 2-point improvement.

Then AlphaFold 2 showed up in 2020.

The “Holy Shit” Moment

There’s no other way to describe it.

AlphaFold 2 scored above 90 on two-thirds of the targets. Some predictions were so accurate they were essentially indistinguishable from experimental results. The competition wasn’t close. It wasn’t even a competition anymore.

The judges called it “astounding.” One researcher said it was “like landing on the moon.” Another said protein structure prediction had been “solved.”

I looked at Molly. This is her field. Transformed overnight.

How AlphaFold Works

Like AlphaGo, AlphaFold uses neural networks. But the architecture is different—it’s built on attention mechanisms, similar to the transformers that power GPT and other large language models.

The key insight is co-evolution.

Here’s the intuition: if two amino acids that are far apart in the sequence consistently mutate together across many different species, they’re probably close together in the 3D structure. Evolution leaves fingerprints. AlphaFold learned to read them.

The system analyzes millions of protein sequences, looking for these co-evolutionary patterns. Then it uses that information—combined with geometric reasoning and iterative refinement—to predict the spatial relationship between every pair of amino acids.

It’s pattern recognition. The same fundamental idea as AlphaGo—but applied to the language of life itself.

AlphaFold 3 and the Nobel Prize

In 2024, DeepMind released AlphaFold 3. It doesn’t just predict individual protein structures—it predicts how proteins interact with DNA, RNA, and small molecules. The implications for drug discovery, gene therapy, and understanding disease are enormous.

Oh, and Demis Hassabis and John Jumper won the Nobel Prize in Chemistry for their work on AlphaFold. No big deal. Just the highest honor in science for a system that started with a board game.

What It Meant to Watch This Together

Here’s what I didn’t tell Molly while we watched: I was so damn excited for her.

She’s walking into a world where the tools to understand life at the molecular level are suddenly, radically more powerful. The intersection of computer science and biology isn’t a niche curiosity anymore—it’s the frontier. And she’s not watching it from the sidelines. She’s studying it. She’s going to use it.

I’ve spent my career in tech, watching waves come and go. I’ve seen hype cycles inflate and collapse. But this one feels different. Her timing is impeccable.

We didn’t plan this documentary double feature as a “teaching moment.” It was just couch time—her ankle in a boot, me with the remote, nowhere to be. But somewhere between Move 37 in AlphaGo and the protein folding breakthrough in The Thinking Game, something clicked.

Her world and my world aren’t separate anymore. They’re the same world.

And she’s going to take it places I can’t even imagine.

The Thread

These two documentaries tell one continuous story—a thread that runs from a board game in Seoul to a protein database that covers all of life.

It’s not about computers being smarter than humans. It’s about building tools that let us see what we couldn’t see before.

AlphaGo showed us a move no human had imagined in 2,500 years of play. AlphaFold showed us the shapes of 200 million proteins that would have taken centuries to solve experimentally.

What’s next? I don’t know.

But I have a feeling Molly is going to help figure it out.

And I’ll be cheering from the couch—boot or no boot.

2025 Year-in-Review: Read. Write. Code.

2025 felt like one of those years where you don’t fully understand what happened until you look backward and realize: oh… that was a phase shift.

The year had a rhythm to it. The front half was dominated by reading — books, papers, blog posts, half-formed ideas scribbled into notes apps. The back half turned into writing, coding and building things that didn’t exist before.

AI was the through-line. Not as a trend. Not as a tool. But as a force.

It reminded me of the 1990s — long nights, endless curiosity, learning by building, and that quiet feeling that something foundational was changing. Web 1.0 energy, but faster. Much faster.

Work, especially at O2E Brands, felt like skiing the same black diamond run for the tenth time. Familiar terrain. Different snow. Same steep pitch. New ice patterns. And no matter how well you know the run, there are always shifts and turns you don’t expect — moments that force you to react, adjust, and stay fully present. You still need to stay on edge.

This was a year of reading, writing, and coding — in that order, and then all at once.

Read

I read a lot in 2025. Mostly in the first half. Books gave me the long arcs — the kind of perspective you don’t get from timelines or hot takes.

Books I Read in 2025

Favorite Books of the Year

Reading Themes

A few patterns emerged:

  • Acceleration is not optional. It’s structural.
  • Agency matters more than raw intelligence.
  • Understanding beats adoption.
  • Strategy still matters — maybe more now than ever.
  • AI is power infrastructure, not a feature.

Trying to keep track of every white paper, blog post, or research thread was impossible. The firehose won. That itself felt like a signal.

Write

If reading was about absorbing, writing was about sense-making.

Most of my writing in 2025 happened in the second half of the year, once the ideas had compressed enough to feel coherent.

Blog Posts in 2025

Earlier

Writing Themes

Across all of it, a few ideas kept resurfacing:

  • AI as physics, not hype
  • Focus as a competitive advantage
  • Context and memory as primitives
  • Enterprise transformation as inevitable
  • Human judgment still matters — a lot

Writing became less about explaining what and more about explaining why it feels different.

Code

Coding felt fresh again.

That alone is worth pausing on.

I found myself deep in computer science concepts — state machines, orchestration, evaluation, memory — like I was back in college. Only now, the feedback loops were instant.

Coding Projects & Experiments

  • Vibe coding with Cursor and Anthropic Claude
  • Vibe coding with GitHub Copilot
  • LangChain, LangGraph, and LangSmith prototypes
  • OpenAI API multi-modal applications
  • Custom LLM hosting and development

“English” has quietly become the most important programming language — but calling this “vibe coding” misses the point. It’s a dumb term because it implies looseness where there is actually rigor. You still need to understand what’s happening under the hood. The real leverage comes from knowing where to apply the tool and why — not from blindly trusting it.

At the same time, multimodal finally feels real. Text, images, audio, video — not as demos or promises, but as things you actually experience while building. That shift alone changes how software feels to write.

As a result, classic enterprise software patterns suddenly look unfamiliar. Some are breaking. Some are reforming. Many are being replaced by agent-driven workflows. And in the middle of all of that, I’m seeing pockets of productivity that are honestly mind-bending.

Closing: Read. Write. Code.

Everything feels different now — not in a loud, hype-driven way, but in a quieter, more structural sense. The kind of different you feel in your hands while you’re building, or halfway through a problem when you realize the old mental shortcuts no longer apply.

There will be people who get it, and people who don’t. And the frustrating part is that those who get it often struggle to explain it to those who don’t. Not because it’s secret, but because it’s experiential. You have to do the work to feel the shift.

The productivity is real. The leverage is real. But it only shows up when reading turns into understanding, writing turns into clarity, and coding turns into creation.

That loop — read, write, code — is what kept me grounded this year. It’s how I made sense of the acceleration without getting lost in it. And it’s why, despite all the change, the craft itself still feels familiar.

Read. Write. Code. Repeat.

Onward.

Latent Space: The Hidden Infrastructure of Intelligence

The First Time You Realize AI Sees the World Differently

Latent space is one of those concepts that feels deceptively simple but quickly becomes mind-bending the deeper you go. In my MIT coursework, the moment it truly clicked wasn’t when someone showed a diagram or equation—it was when I watched two very different inputs land right next to each other in a high‑dimensional embedding space. Suddenly, you realize: AI doesn’t see categories the way we do. It sees geometry. And that geometry is the beating heart of modern AI.

This post is my attempt to make latent space both intuitive and technically sound—a tour of the hidden mathematical world that lets AI models generalize, reason, and occasionally surprise the hell out of us. If you’ve read my earlier posts like Efficiency Reckoning or AI Is Eating Software That Is Eating the World, you’ll recognize a recurring theme: exponential capability often hides in plain sight until you learn to see the structure underneath.

What Is Latent Space?

Latent space is the compressed mathematical world where AI stores meaning. Instead of memorizing data, models learn dense vector representations that capture the essence of concepts—objects, actions, styles, emotions, operational patterns. Similar ideas cluster together; different ideas drift apart. Geometry becomes understanding.

Key ideas:

  • Embeddings: Numerical vectors that represent the meaning of inputs (words, images, tokens). Their position and direction encode semantic relationships.
  • Distance: A mathematical measure (often cosine similarity or Euclidean distance) that indicates how similar two embeddings are. Closer = more related.
  • Manifolds: Lower‑dimensional, structured surfaces within the high‑dimensional latent space where meaningful data naturally clusters. Models “discover” these during training.

Everything the model “knows” lives somewhere in this hidden space.

Why Latent Space Is AI’s Superpower

Latent spaces give models the ability to:

  • Generalize beyond what they’ve seen.
  • Recognize analogies and patterns.
  • Perform zero-shot reasoning (answer questions they were never explicitly trained on).
  • Compress knowledge into a shape that can be navigated, manipulated, and queried.

This geometry is the fuel behind why large models feel so shockingly capable. It’s the same idea I explored in The Law of Accelerating Returns, systems don’t merely improve—they reshape the surface beneath our feet. Latent space is the mathematical expression of that reshaping. We’re no longer programming rules; we’re shaping the very spaces where meaning lives.

How Latent Spaces Are Built

Latent space emerges naturally during training, driven by the model’s need to predict missing information.

The Compression Process

Self-supervision forces the model to strip away noise and preserve structure. This compression yields abstract, high-dimensional patterns that capture relationships rather than raw inputs.

Transformation Through Layers

Embeddings pass through dozens or hundreds of transformer layers. Each layer rotates, stretches, and refines meaning until stable semantic structures emerge.

The Result: A Structured World

By late training, the model has carved out clear neighborhoods for concepts—objects clustering near the actions they relate to, pricing signals gravitating toward contextual cues, operational patterns forming their own orbits.

How Latent Space Behaves

Despite being abstract and high dimensional, latent spaces exhibit surprisingly intuitive properties. Humans naturally build mental maps to navigate ambiguity, and AI does something similar—just at a scale and dimensionality far beyond our own., latent spaces exhibit surprisingly intuitive properties.

Smoothness

Small moves yield gradual changes in meaning, enabling interpolation, transformation, and reinterpretation.

Relational Structure

Directional changes encode relationships—analogy, comparison, and categorization become geometric operations.

Compositionality

Concepts can combine fluidly: an object + context + constraint forms a new point in space that the model can reason about without explicit rules.

Natural Clustering

Clusters form organically, often better than human taxonomies—but also reflecting limitations or hidden biases.

Where Latent Space Breaks Down

As magical as it feels, latent space isn’t perfect:

  • Latent collapse: Everything clusters too tightly.
  • Overfitting: Geometry becomes brittle.
  • Bias: Prejudices become encoded as spatial structure.
  • Out-of-distribution drift: The model hallucinating outside the manifold.

These limitations matter when deploying AI into real operational environments, where edge cases are everywhere.

The Future: Latent Space as the New Programming Model

We are entering an era where latent space isn’t just a byproduct of AI models—it becomes the substrate of software itself. The shift is profound: instead of writing rules, we increasingly shape geometry, influence structure, and design the conditions under which models discover meaning.

Several forces are driving this transformation:

  • Geometry replaces logic. Traditional programming encodes explicit steps; latent‑space systems embed intent, relationships, and constraints into the shape of the space itself. We’re not prescribing behavior—we’re defining the terrain.
  • Agents operate like navigators, not executors. Agents don’t follow deterministic paths. They explore, sample, and move through conceptual regions, selecting actions by proximity, similarity, and predicted outcomes. This is closer to robotics in a physical world than software in a deterministic one.
  • World models introduce simulation as a first‑class primitive. When a model can simulate consequences inside its latent space, it stops behaving like a tool and starts behaving like a planner. Software becomes anticipatory rather than reactive.
  • Developers shift from coding workflows to curating spaces. The primary task becomes shaping embeddings, conditioning behavior, tuning representations, and steering emergent structure. Infrastructure teams will manage vector spaces the way they once managed databases.

This isn’t incremental. It is a foundational rewrite of how software is conceived and built—arguably the most important transition since cloud computing abstracted away hardware. Latent space abstracts away rules themselves.

Personal Reflection: Why This Matters to Me

Returning to technical study has been a joy—and a humbling reminder that AI is a field where intuition and mathematics collide. Latent space, more than any other concept, embodies that collision in computational form.

Understanding it doesn’t just make you a better builder of AI systems.
It makes you a better interpreter of AI behavior.

And maybe a bit more forgiving when a model wanders too far off the manifold.

The Physics of Focus

I believe that focus is the single most important variable in the equation of execution. Yet, time and time again, I see executive teams unable to summon it. My theory is that true focus requires a level of strength and accountability resilience that most executives simply don’t have. It demands an “all or nothing” mentality—a trait most often found in founders and entrepreneurs, but rarely in established corporate structures.

It is the difference between movement and velocity. In physics, movement is scalar—it’s just speed. Velocity is a vector—it requires speed plus direction.

In my career leading product and engineering organizations, I have seen brilliant teams work incredibly hard, burning massive amounts of energy, only to achieve zero net displacement. Why? Because their vectors were misaligned. They were moving fast, but they weren’t moving forward.

The most idyllic example of this is the second coming of Steve Jobs at Apple in 1997.

The CEOs who preceded him—Michael Spindler and Gil Amelio—fell into the classic trap of trying to manage their way out of a crisis by adding more. They believed that capturing every niche meant building a specific product for every niche. They lacked the strength to say “no,” so they said “yes” to everything. The result was a confused portfolio of Performa, Quadra, and PowerBook models that baffled customers. They were chasing revenue, but by fracturing their focus, they were destroying value.

When Jobs returned, he saw that Apple wasn’t suffering from a lack of talent. It was suffering from a surplus of “good” ideas. The product roadmap was a mess of printers, servers, and slightly varied Macintoshes. Jobs didn’t ask the teams to double their output. He walked to a whiteboard, drew a simple 2×2 grid (Consumer/Pro, Desktop/Portable), and wiped the rest of the table clean.

He deleted 70% of the roadmap. He understood that to make the iMac great, he had to let the Newton die.

This is the hardest lesson for leaders to learn: Focus isn’t about concentration; it’s about subtraction.

I like to think of this in terms of physics.

The Equation: Pressure = Force / Area

In engineering, pressure is determined by dividing force by the surface area (P = F/A).

  • If you apply your force over a large area, you create very little pressure. You are essentially leaning against the wall.
  • If you apply that same force to the tip of a needle, you can punch through steel.

Most of us—and most companies—operate like a flat palm pushing against a wall. We spread our limited bandwidth across a dozen priorities, creating zero penetration. We mistake “busy” for “impact.”

The Entropy of Success: Why Focus is So Difficult

If the physics are so clear—reduce the surface area to increase pressure—why is it so rare to find a truly focused organization?

In my experience, five forces conspire against focus:

  1. Too Much Opportunity: Paradoxically, opportunity is the biggest threat to execution. When you have resources and options, the temptation to chase every rabbit is overwhelming.
  2. Lack of Constraints: Innovation thrives on constraints. When there are no hard limits—whether imposed by the market or self-imposed by leadership—strategies bloat.
  3. Lack of Discipline: As I mentioned, focus is an “all or nothing” game. It requires a level of behavioral discipline that is uncomfortable for most.
  4. The “General Manager” Problem: You cannot make surgical cuts if you don’t understand the anatomy. We have too many general managers who lack the detailed, functional understanding of their business. They can’t decide what to cut because they don’t truly understand how the engine works so they hedge their bets.
  5. Old School Executive Thinking: There is a pervasive belief that an executive’s job is to ask for more—more features, more markets, more growth. I fundamentally disagree. I believe an executive’s primary job is to ask for less. Less confusion. Less drag. Less surface area.

Here is a framework for how to apply this “reduction of surface area” from the boardroom down to your calendar.

1. Macro Focus: The Corporate Strategy

The Principle: Strategic Abandonment

In a company, the enemy of focus isn’t the bad idea. Bad ideas are typically self-correcting because the data screams “stop.”

The real enemy is the “good” idea. The project that is profitable enough. The feature that customers like somewhat. These are the “Zombie Projects.” They are dangerous because they consume engineering cycles, marketing budget, and leadership headspace, yet they prevent you from putting sufficient pressure on the one or two initiatives that actually drive the business.

This is Strategic Drag. Every “good” project you keep alive slows down the velocity of your “great” projects.

The Framework: Adopt a “Hell Yeah or No” policy. If a new initiative is just “reasonable,” kill it. “Reasonable” is the enemy of “Exceptional.” If you have ten priorities, you actually have zero. You need the courage to starve the profitable “good” to feed the potential “great.”

2. Operational Focus: The Digital Rebirth

The Principle: Subtraction Before Digitization

This physics equation applies just as ruthlessly to how a company operates. Most legacy corporations are clogged with “process sediment”—layers of analog habits, manual workarounds, and spreadsheet safety nets that have accumulated over decades.

When these companies try to modernize, they often fail because they try to add digital workflows on top of analog mindsets. They implement the new ERP system but keep the manual approval chain “just in case.” They build the data lake but keep the Excel trackers. This splits the organization’s focus between the past and the future.

The Framework: You cannot iterate your way from a horse to a car; you have to abandon the horse. For a process to be truly digitized, the analog version must be stopped, not just deprecated.

  • Burn the Boats: If you are digitizing an operation, you must remove the ability to revert to the manual way.
  • The Rebirth: Treat operations like products. Don’t just pave the cow path. Wipe the table clean (just like Jobs did) and ask, “If we started this company today with zero history, how would this workflow exist?” Build that, and delete the rest.

3. Meso Focus: Career & Ambition

The Principle: The Competence Trap

As we grow in our careers, we accumulate skills. We become capable of doing many things well. This leads to the Competence Trap: Just because you can do something doesn’t mean you should.

There is a mental model often attributed to Warren Buffett regarding this. He purportedly asked his pilot to list his top 25 goals, then circle the top 5. The pilot assumed he should work on the top 5 and do the other 20 in his spare time.

Buffett corrected him: The bottom 20 are the Avoid-At-All-Costs list. They are the things you care about just enough to distract you from the few things that matter.

The Framework: Audit your commitments. Identify the things you are doing simply because you are capable of doing them, not because they drive your long-term value. If it’s not in your top 5, it is a distraction. Cut it loose.

4. Micro Focus: The Daily Task

The Principle: The Context Switching Penalty

This is where the battle is won or lost on a Tuesday morning.

We live in an economy designed to fracture our attention. We deceive ourselves into thinking we can multitask, but the science says otherwise. Researchers call it “Attention Residue.” When you switch from a strategic document to check an email, a significant portion of your CPU (your brain) remains allocated to the email even after you switch back.

This is latency. If you switch contexts every 15 minutes, you are effectively operating with a fraction of your IQ. You are introducing lag into your own processing power.

The Framework: You cannot rely on willpower; you must rely on systems.

  • Defensive Design: Remove the option to be distracted. Physical separation from the phone. Turn off notifications.
  • The Deep Work Block: Schedule a 4-hour block for your hardest task. Treat it like a flight—you are buckled in, the door is closed, and there is no WiFi. This is how you generate density.

Final Thought

Focus is painful.

It hurts to kill a product line that people worked on. It hurts to force a team to abandon a familiar manual process.

But that pain is the feeling of the surface area shrinking. It is the feeling of you becoming the needle rather than the palm.

If you want to execute, stop adding. Start subtracting.

The Great Corporate Efficiency Reckoning

Introduction: Welcome to the Reckoning

We’ve reached a moment where corporate transformation is no longer optional—it’s being forced. AI isn’t just another technology wave; it’s an existential efficiency reckoning. Companies that aren’t ruthlessly exposing and eliminating inefficiencies are already behind.

There’s no hiding from this one.

The Myth of Smooth Transformation

Executives often talk about digital transformation like it’s a one-time project. It’s not.

Transformation is constant. The companies who survive are the ones who operate under continuous reinvention. But here’s the painful truth: most companies don’t change until they’re forced to. The veneer of stability has covered up shockingly deep inefficiencies.

Most workflows inside organizations aren’t even documented. Ask an ops leader to produce a process map and watch them sweat. As Peter Drucker said: “You can’t manage what you can’t measure.” But what if you can’t even find the work?

The AI Catalyst — and the Hard Truth

AI hasn’t created inefficiency. It’s simply exposed it.

We’re seeing companies push out CTOs because they “aren’t doing enough with AI.” But the problem isn’t just technical leadership—it’s organizational culture. AI isn’t an engineering project. It’s an “all hands on deck” transformation.

“You can’t outsource transformation—it’s cultural, not contractual.”

AI has thrown every department—from finance to field ops—into the blender. And now executives are realizing the painful truth: they haven’t built for this era.

Efficiency Beyond AI

Ironically, the quest to implement AI has exposed the many areas where companies are simply… bloated.

  • Work expands to fill the time and headcount you give it.
  • Revenue teams buried in manual tasks.
  • Ops teams drowning in redundancy.
  • Leaders who’d rather tweak PowerPoints than face structural inefficiency.

Sometime you have to cut until it bleeds. Then you fix the wound.

The companies that resist necessary cuts won’t survive.

Cut to Transform — or Get Cut

Look no further than Twitter’s transformation into X under Elon Musk. In 2022, Musk cut nearly 50% of the workforce in just weeks—an unprecedented move that shocked the tech world. Critics called it reckless, but it forced radical simplification: collapsing dozens of teams, canceling bloated contracts, and removing layers of slow decision-making. Whether you agree with the method or not, the rapid downsizing catalyzed operational reinvention and aggressive experimentation—at a pace impossible under the old structure.

In your own organization, the same holds true. It’s not about being heartless. It’s about being real. The days of “everyone has a job forever” are gone. And here’s the hidden risk: cultures that protect themselves from this kind of forced reinvention—clinging to the idea of being one big “family”—are often worse off. A family tolerates inefficiency and preserves comfort. A world-class team embraces accountability and expects constant growth. One survives disruption, the other gets disrupted. Meanwhile, cultures that cling to the comfort of the past—celebrating legacy wins instead of inventing their future—won’t just fall behind; they’ll be blindsided. Progress demands letting go of nostalgia and facing the work with fresh eyes.

The Agentification Era

We’re entering a phase where systems manage systems and humans orchestrate outcomes rather than manually triggering workflows. This is the era of agentification—a shift where autonomous or semi-autonomous agents are entrusted with achieving outcomes, not just performing tasks.

Frameworks like LangChainGemini Enterprise, and the emerging OpenAI Agent environments are infrastructure-level systems. They allow developers to chain models, tools, and APIs into cohesive, self-directed workflows capable of context-driven execution. In practical terms, this means a pricing agent could access CRM data, query inventory, consult policy libraries, infer customer sentiment from voice transcripts, and autonomously adjust pricing—all while handling exceptions, logging actions, and escalating when needed.

Agent-based systems are not merely “automated scripts”. They are dynamic reasoning engines managing both state and decision trees over time. But these promises come with significant complexity: memory management, context windows, hallucination control, permissions when tools perform actions (especially destructive ones), and transparent auditability for governance.

“The next great company won’t just automate workflows—it will orchestrate layers of intelligence across every function.”

A crucial challenge remains: from pilot to production. While demos often succeed in controlled environments, the real world introduces unpredictable data, unique edge cases, and higher stakes. Companies often stop at “AI-powered mock-ups” because operationalizing agent workflows requires not only technical sophistication but organizational readiness—clear workflows, defined permission tiers, and a system for rapid correction.

Most companies will never scale beyond prototypes—not because of model limitations, but because they lack the operational discipline to productionize intelligence safely and repeatably.

Where Efficiency Is Already Winning

  • Engineering productivity: Teams are achieving 2–5x development velocity using AI copilots, large-context-aware IDE extensions, and context caching. One of my own technical leads recently noted: “My problem isn’t writing code anymore—it’s code reviewing all the code that’s being generated.” Toolchains like GitHub Copilot, TabNine, and custom agents integrated into developer workflows are producing high volumes of synthetic code, shifting the bottleneck to validation, testing, and deployment pipelines. This is no longer a human-speed system; it’s systems managing systems, with automated PR reviews, AI-driven linting, and even self-healing deployment scripts.
  • Voice & call center automation: Modern real-time voice agents now run on hybrid ASR/NLP stacks powered by models like Whisper, Deepgram, or AssemblyAI. These systems detect intent, sentiment, and compliance in milliseconds, and many now integrate directly with ticketing and ERP systems. This allows for fully automated routing, booking, and billing resolution, driving 30–60% reductions in cost and a massive boost in consistency.
  • Finance automation: AI-enabled reconciliation engines and anomaly detection models are turning manual month-end close processes into continuous close operations. Systems ingest GL data in real time, align it against prior period models, and auto-escalate exceptions. What once required multi-week manual preparation is now accomplished in hours, with precision.

These examples underscore a simple truth: efficiency doesn’t always require full platform rewrites—it often involves augmenting existing systems with intelligence layers. Companies achieving the biggest gains are systematically inserting AI into workflow steps where context is well-structured and feedback cycles are measurable, allowing rapid iteration and ROI tracking.

Are Jobs Actually Going Away?

Some jobs are disappearing—but most aren’t. According to recent estimates from the Wharton School, only about 1% of jobs could be fully automated today—but more than 25% of U.S. roles include tasks where 90–99% of the work could be automated by AI. That’s not just a shift—it’s a restructuring of human work.

McKinsey estimates that up to 30% of current work hours could be automated by 2030 due to generative AI, with productivity boosts concentrated in knowledge work, customer service, finance, and coding-intensive roles, yet only 5% of occupations are expected to be fully replaced.

A recent BCG report suggests AI can support or automate as much as 80% of the tasks within corporate functions like operations and finance, with teams reclaiming between 26–36% of their time when used effectively.

Humans won’t disappear—but they will need to adapt. The Gallup Workplace Index notes that usage of AI tools at work has nearly doubled in the last two years, from 21% to 40%, signaling broad adoption—not displacement. The real shift is in how work is structured and what skills are needed.

Bubble or Not?

Are we in a hype cycle? Maybe.

Yes, It’s a Bubble: The exuberance has led to inflated expectations and a flood of pilots that never make it to production. As one analyst put it, “AI has become the new currency of hype—and hype has a cost.” For every success story, there are failed proofs of concept, governance blind spots, and companies rushing in without readiness.

No, This Is Real: If this is a bubble, it’s a bubble filled with fundamentals—real productivity gains, measurable cost reductions, and transformative potential across sectors. One CTO remarked, “The hype isn’t the point—the results are.” The companies winning right now aren’t just riding hype; they’re building lasting operational muscle.

But here’s what’s real: this efficiency wave isn’t optional. And this feels fundamentally different from the Dot-Com 1.0 boom I lived through as an engineer at Scient—a company that once embodied the innovation narrative and now exists only as a cautionary tale. Back then, we were building for a promise, often without the means to fulfill it. Today, companies are already seeing results before they’ve even finished the slide decks. This isn’t speculative. It’s happening in real time—and the companies who fail to adapt won’t just fade quietly—they’ll be overtaken.

“Once you’ve watched an AI model build a product roadmap in 5 minutes, you can’t go back to waiting 6 months for slide decks.”

Closing: What’s Next for Corporate Efficiency

Leaders must stop treating AI like an IT problem. It’s an operating system shift.

The future belongs to companies who act on inefficiency—not those who fear the discomfort of change. The winners will treat efficiency as a core value.

We’re not just implementing AI. We’re rewriting how companies work.

Welcome to the reckoning.

Recursion: We Have to Go Deeper

Opening Scene: Inception & the Dream Layers

You’re in a van. It’s falling. But somehow… you are not.

Gravity tilts sideways. Time stretches like taffy. And as the van free-falls, you suddenly find yourself in a hotel hallway — where everyone is also falling, but standing up straight, because physics took the night off.

Then — snap — you’re in a snowy fortress. Guns. Ski masks. Zero explanation.

You are multiple layers deep into the same mission. Each layer depends on the one above. Each layer runs slower than the one below.

This is recursion in IMAX — brought to you by director Christopher Nolan, the reigning champion of layered storytelling (and one of my all‑time favorite filmmakers).

A state inside a copy of that state. A system inside a smaller version of the same system. A repeated descent… until you find the exit condition.

Because without a way back — without a base case — you’re not a hacker or an architect or a programmer. You’re lost. In Limbo. Running a function that calls itself forever.

Movie poster for 'Inception', featuring a man in a suit standing in a flooded urban environment, with skyscrapers in the background. The title 'Inception' and credits are displayed prominently.

Dinner in Boston with My Daughter

Recently, I was in Boston having dinner with my daughter — a Computational Biology major (which means she studies things that evolve recursively in nature and then complains when recursion shows up in class).

She’s brilliant, curious, and probably smarter than me already. It’s fine. I’m fine. But… recursion? Recursion has been her intellectual villain — like a video game boss whose health respawns every time you think you’ve finally won.

I told her, “It’s just like Inception — a dream within a dream.” She laughed, rolled her eyes, and informed me that my analogy was both helpful and deeply unhelpful at the same time. But the truth is: recursion requires a different mental model — one where you trust the call stack and hope you remembered your base case.

Parenting is similar.

What Recursion Actually Is

At its heart, recursion is when a function solves a problem by solving a smaller version of itself. Like zooming into a photo… then zooming again… and again — each layer revealing the same picture but slightly smaller, until eventually you hit pixels and either succeed… or question all your life choices.

Every recursive solution needs two things:

  • Recursive case — keep going deeper
  • Base case — stop before everything breaks
function solve(problem):
    if problem is tiny enough:
        return solution  // base case
    else:
        return solve(smaller problem)

Elegant. Terrifying. Very Nolan.

Why Recursion Feels Hard (and How Iteration Tries to Save Us)

Humans love straight lines. Step 1 → Step 2 → Step 3 → victory royale.

Recursion hands you a spinning top and says: “Just solve the problem… by solving the problem… inside the problem.”

Your brain becomes a call stack:

  • “Which version of me is running?”
  • “Is this the dream or the dream inside the dream?”
  • “Wait… did I already solve this?”

One missing base case and — boom: Stack overflow. Your program panics harder than Ariadne when the Paris street folds in half.

This is where iteration strolls in like the chill coworker:

  • Recursion: Elegant. Reads like poetry. The Leonardo DiCaprio of code.
  • Iteration: Practical. Gets things done. Drives a minivan.

In fact, recursion and iteration can solve most of the same problems. But recursion does it with flair — and sometimes significantly worse performance if you’re not careful about stack depth and repeated work.

Example: Fibonacci.

# Recursive Fibonacci
function fib(n):
    if n <= 1:
        return n
    return fib(n-1) + fib(n-2)

Run with n = 40 and you may have time to:

  • Make coffee
  • Drink that coffee
  • Earn a PhD in Computational Biology
  • Explain recursion again over dinner

Why? Because it repeats the same work exponentially.

Now iteration:

# Iterative Fibonacci
function fib(n):
    a = 0
    b = 1
    for i in range(2, n+1):
        temp = a + b
        a = b
        b = temp
    return b

Same output. Runs faster than Dom Cobb when the van is five seconds from splashdown. No duplicate work. No stack explosions.

Performance efficiency matters — and recursion does not always care about your CPU budget.

Still… for problems that nest, branch, or fracture into self‑similar layers, recursion is the genius iteration calls when it’s out of ideas.

Where Recursion Secretly Rules Software

If you’ve made it this far, congratulations — you’ve followed the call stack down a few layers without losing track of who you are. My daughter would be proud.

And yes… we have to go a bit deeper.

Even if production engineers avoid it like unbounded memory leaks, recursion is the hidden backbone of:

  • File systems (folders inside folders inside folders)
  • UI frameworks like React (components inside components)
  • Sorting algorithms (QuickSort’s entire personality is “let’s descend”)
  • Tree and graph search (the internet is basically recursion with ads)

Recursion is the invisible stage crew of computing — unseen but essential.

Bonus nerd moment: Computational biology uses recursion constantly — phylogenetic trees, protein folding… nature recurses.

Recursion in Modern AI (Yes, Even This Post)

Recursion isn’t just a CS 101 rite of passage — it’s powering many of the smartest systems we build today:

  • Neural networks backpropagate through layers recursively.
  • Game‑playing AIs explore huge game trees with recursive search.
  • Recursive neural networks understand language hierarchies.
  • Graph algorithms break complex problems into smaller ones… repeatedly.

Even large language models — hello 👋 — reason over nested patterns that behave recursively, even if there’s no literal call().

Basically, recursion graduated, joined a research lab, and now publishes papers. Iteration still fixes production bugs.

Real Talk: Why We Don’t Use It Everywhere

Production systems hate drama. Recursion can bring:

  • Memory overhead (every function call needs space)
  • Unpredictable depth (sometimes life decides to go four dreams down)
  • Debugging sessions that feel like you’re chasing your own tail
  • That “oh no” moment when your stack detonates in prod

Bring a totem: unit tests, depth limits, and logs — so you always know which layer you’re in. Loops are the steady minivan — recursion is the flashy sports car you only drive when the weather is perfect.

Recursion in Pop Culture & Reality

If recursion feels abstract or like something only math people obsess over… it already found you.

Recursion is everywhere:

  • Mirrors facing mirrors 🎭
  • A photo of someone holding a photo of themselves 📸
  • Russian nesting dolls 🪆
  • Your company org chart (boss → boss → board → shareholders → chaos)

Parenting might actually be recursion: Teach kid → kid grows up → teaches their kid → reboot universe.

The Recursive Ending

Back in Boston, my daughter asked: “So recursion is just… doing the same thing, but smaller, until you can stop?”

I nodded. She spun an imaginary top and said, “Wake me when we hit the base case.”

That’s recursion: Define the base case. Trust the descent. Don’t lose your totem.

If this keeps going, we’re in Limbo.

return "We have to go deeper."

The World Models Within Us

From Asimov to AGI: The Rise of Predictive Minds

“A robot may not injure a human being, or, through inaction, allow a human being to come to harm.” — Isaac Asimov

Isaac Asimov, one of my favorite authors, is best known for his Robot series — books that deeply influenced how I think about technology and ethics. The series wasn’t just about robots; it was about the emergence of intelligence — about what happens when a machine begins not just to obey, but to understand. His characters like Dr. Susan Calvin grappled with robots who broke the Three Laws not out of rebellion, but out of deeper reasoning — they’d built internal models of the world complex enough to predict human consequences.

That, in essence, is the story of world models in modern AI. Machines are learning not just to process data, but to imagine futures.

The Foundation of World Models

While most consumer-facing AI tools we see in the news or use at home or work — like ChatGPT, image generators, or voice assistants — excel at pattern recognition, true world models go a step further. They don’t just respond to input; they build internal representations of how the world works, allowing them to reason about context, dynamics, and cause-effect relationships. This makes them capable of simulating and anticipating outcomes rather than simply reproducing patterns from past data. A world model is a latent representation of reality that allows an intelligent system to simulate, plan, and predict. In technical terms, it’s what allows agents like MuZero or GPT-based systems to operate beyond mere pattern recognition. Instead of memorizing, they model — forming compressed internal maps of how the world behaves.

If a neural network is like a camera capturing pixels, a world model is like a mapmaker building an atlas. It abstracts, generalizes, and anticipates — and crucially, it allows for planning. A good world model can take incomplete information and still simulate likely futures. It’s the engine behind self-driving cars navigating uncertainty, generative agents predicting user intent, and language models inferring context.

This idea originated in reinforcement learning research (Ha & Schmidhuber’s World Models, DeepMind’s MuZero, OpenAI’s Sora). But the same principle now underpins multimodal systems like GeminiGato, and GPT-5, which combine perception, reasoning, and action under one unified architecture. These systems don’t just respond — they simulate, project, and plan.

From Logic to Learning: How We Got Here

The lineage of AI traces a fascinating arc. It began with symbolic AI, when we tried to encode knowledge into explicit rules — much like Asimov’s Three Laws. Those systems could reason, but only within the narrow confines of their logic. Then came the revolution of statistical learning, where we replaced rules with probabilities and let data teach the machine. That gave us vision systems, speech recognition, and language models — but not true understanding.

Now we’re entering the age of world modeling — a synthesis between symbolic precision and statistical intuition. These new systems learn not only to recognize patterns, but to simulate how the world changes over time. It’s a step closer to how humans think. When you reach for a coffee mug, you don’t calculate every possible trajectory — your brain’s world model has already predicted the motion.

World models are the missing bridge between perception and agency. They turn observation into imagination and allow intelligent agents to act not just reactively, but intentionally.

Why World Models Matter

Every intelligent system — human or machine — must answer the same question: What will happen if I act? That is the essence of intelligence — the ability to simulate the future and choose accordingly.

In AI, this predictive capacity is fundamental:

  • In AGI research, world models enable self-supervised learning, planning, and reasoning across tasks.
  • In enterprise applications, they power agents that can simulate business outcomes before executing decisions.
  • In human cognition, predictive coding lets our brains anticipate the world milliseconds before our senses confirm it.

World models, then, are not just a technical concept. They’re a philosophy of intelligence — a recognition that foresight, not memory, is the truest marker of understanding. They allow systems to bridge the gap between data and decision, between reaction and anticipation.

The Business Parallel

In business and technology alike, we build world models to make sense of complexity. Whether we call them strategy frameworks, simulation tools, or digital twins, these systems help us forecast outcomes, reduce uncertainty, and align execution. A well-constructed model of operations or markets acts as a simplified reflection of reality, letting leaders test scenarios before taking action.

Organizations that refine their world models — capturing customer behavior, market dynamics, and operational constraints — become more adaptive. Pricing, scheduling, and resource allocation evolve from guesswork into predictive, data-informed decisions. The fidelity of a company’s model of its world determines how effectively it can act within it.

Every system — human, corporate, or artificial — competes on the fidelity of its world model. The sharper the map, the smarter the motion.

The Rise of Embodied Intelligence

Recent advancements in robotics have brought world models from simulation to the physical world. Robots are no longer confined to factory floors or research labs—they’re navigating complex, unstructured environments using predictive world modeling to anticipate, adapt, and learn.

Take Tesla’s OptimusBoston Dynamics’ Atlas, or Figure AI’s humanoid robots—each of these systems relies on internal models to understand and predict physical interactions. They don’t just follow pre-programmed paths; they build internal maps of how objects, gravity, and force behave. When a robot walks across uneven terrain or picks up a delicate object, it’s using its world model to simulate potential outcomes before acting.

Even in household robotics, we’re seeing the impact. Modern robot vacuums and warehouse automation bots use spatial and behavioral world models to improve navigation and decision-making over time.

This convergence of embodied AI and world modeling is what gives machines agency in the real world. The same predictive reasoning that allows a language model to plan a paragraph now lets a robot plan a movement. As the fidelity of these physical world models increases, robots will move from reactive tools to proactive collaborators—learning continuously through experience.

The Limits and the Next Frontier

All models are simplifications, and even the most advanced world model is still a shadow of reality. The world shifts faster than any simulation can adapt. World models hallucinate, overfit, and misinterpret context — much like humans with bias and intuition. But even imperfect models can be powerful tools for alignment: between perception and truth, intent and impact.

Researchers like LeCun, Schmidhuber, and the teams at DeepMind are exploring architectures (JEPA, Dreamer, Genie) that enable machines to learn continuously from their environment, refining their world models the way children do — through play, prediction, and correction. It’s a vision of AI that doesn’t just compute but grows.

The next frontier is not just bigger models, but better models of the world — ones that learn cause and effect, adapt to change, and stay grounded in reality.

Asimov’s Mirror

Asimov’s fiction gave us robots wrestling with morality; our reality gives us agents wrestling with understanding. His positronic brains modeled the world to serve human ends. Ours must now do the same — but with humility.

Asimov’s Robot stories ultimately asked whether understanding the world made robots more human — or humans more predictable. That same question echoes today. The more our machines model us, the more we must model ourselves.

“His robots learned to imagine the consequences of their actions. We’re finally teaching ours to do the same.”

Every great civilization — and every great AI — begins with a world model.

Bold Moves: The Antidote to the Status Quo

I still remember the feeling—the mix of excitement and terror—as I packed the last box into a U-Haul after college. Two buddies of mine drove with me across the United States driving a U-Haul and my 1999 Toyota 4-Runner (that I still own and drive today). The destination was San Francisco, California, the goal was the Silicon Valley dream. Along the way, we attempted to visit as many Major League Baseball ballparks we could get to. Our favorite was Wrigley Park in Chicago. As a Computer Science major, I was drawn to the epicenter of the digital gold rush. It was 1999, and the headlines were intoxicating: companies like GeoCities and theGlobe.com were having record-breaking IPOs despite having no profits. The air was thick with stories of 20-somethings becoming overnight millionaires, and the promise of a ‘new economy’ fueled by giants like Yahoo! and countless other dot-coms felt limitless. It truly felt like the center of the universe, the only place to be. But in reality, it was less a carefully planned career step and more a blind leap of faith. I didn’t need to do it. I had job offers in Washington DC and New York City for lucrative programming jobs. Wow, my life would have been different if I didn’t do this. Looking back, I can see how that single, impulsive move set the tone for my entire life. It was my first real lesson in a principle I now live by: the life you get is a reflection of the bold decisions you’re willing to make.

The Seduction of the Status Quo (and the Gravity of Safety)

Life has a funny way of pulling us toward the center. Both personally and professionally, there’s a natural gravity toward safety, predictability, and the well-trodden path. It’s the comfort of the known, the security of the status quo.

The problem is, safety is an illusion. The real risk isn’t in taking a leap; it’s in standing still. The cost of avoiding bold moves is stagnation. It’s a slow fade into irrelevance as the world moves on without you. The comfortable path inevitably leads to a place of regret, wondering “what if?”

My Bold Moves: Personal Stories That Shaped Me

That U-Haul to San Francisco was just the first of many bets I’ve made on myself. Each one felt like defying gravity at the time.

  • Starting a company right after our second child was born: On paper, it was the worst possible time. The responsible move would have been to find a stable job with a predictable paycheck. But the pull of building something from the ground up was stronger than the fear of instability. That company became one of the most formative experiences of my life.
  • Leaving stable corporate jobs for startups: I’ve done this a few times in my life. It meant leaving the safety of a clear career path for the chaotic, high-stakes world of a startup. Each time, it was a bet on impact and accelerated learning over the comfort of certainty, and each time it paid off.
  • Moving my family to Park City during COVID: The world was shutting down, and we decided to uproot everything. We left the familiar for the mountains, seeking a different quality of life. It was a bet on a lifestyle, and it paid off in ways we couldn’t have imagined.

Each of these moves was a conscious push against the gravity of safety. And each one returned dividends in growth, learning, and fulfillment that far outweighed the perceived risks.

What Bold Moves Create (in Business & Life)

This principle isn’t just personal; it’s the engine of progress in business. History is littered with examples of companies that won or lost based on their appetite for boldness.

Think of Apple launching the iPhone, a bet-the-company move that cannibalized their successful iPod business. Or Netflix going all-in on streaming when their DVD-by-mail service was at its peak. Or Amazon Web Services, a wild idea that had nothing to do with e-commerce but now powers a significant portion of the internet.

Conversely, think of the corporate graveyards filled with companies that clung to the status quo. Kodak invented the digital camera but buried it to protect its film business. Blockbuster laughed Netflix out of the room. BlackBerry was convinced its physical keyboard was invincible. They all played it safe, and they all lost. Boldness is what scales outcomes, both for individuals and for empires.

Bold Moves Don’t Always Mean Giant Leaps

But boldness doesn’t have to be a U-Haul across the country or a nine-figure business bet. Sometimes, the boldest moves are the small ones that accumulate over time.

It’s speaking up in a meeting when everyone else is silent. It’s making the cold call you’ve been dreading. It’s making the difficult decision to part ways with a team member who was perfect for the company’s past but has outgrown its future. It’s saying “no” to a good opportunity to protect your time for a great one. These small acts of courage build the muscle for bigger leaps. They create a compounding effect, where each small, bold move creates the foundation for the next.

The Fear Factor: Why Boldness Feels So Hard

Let’s be honest: bold moves are terrifying. The fear is real. It’s the fear of failure, the fear of judgment, the fear of leaving the stability we’ve worked so hard to build. With every big decision I made, fear was a constant companion. When starting a company with a young family, the fear of not being able to provide was immense.

But I learned that fear isn’t a stop sign. It’s a compass. It points you toward the areas where you have the most to grow. Leaning into that fear, acknowledging it, and moving forward anyway is what unlocks progress.

The Payoff: Why Boldness Wins

The beautiful thing about bold moves is that they create momentum, even when they appear to “fail.” We spend too much time measuring success in dollars or fearing what others might think when they see something fall short. But a failed startup teaches you more than a decade in a safe corporate job. A move that doesn’t work out still expands your perspective and builds resilience. There is no such thing as a failed bold move, only learning opportunities that propel you forward. Each step, successful or not, compounds over time, building a life and career defined by growth, not stagnation.

Bold Moves Are Required in Startups and Transformations

This mindset is non-negotiable in the worlds I operate in. In a startup, playing it safe is a death sentence. The only way to break through the noise and overcome the inertia of established players is to make bold bets.

The same is true for corporate transformations. Companies don’t pivot from legacy models to future-proof businesses by making incremental tweaks. It requires fundamental, bold shifts. History is a testament to this: Sears clung to its catalog model while Amazon built the future of retail, Nokia dismissed the iPhone to protect its existing phone business, and Yahoo had the chance to buy Google but played it safe. In my work, I’ve seen what happens when companies embrace this, but the truth is, I don’t see it enough. The winning companies are the ones making bold moves in their product strategy, aggressively adopting AI to reinvent “non-tech” industries, and challenging every assumption about how their business should run. Without this commitment to boldness, any company is destined for the corporate graveyard alongside Kodak and Blockbuster.

Making Bold Moves a Habit

Boldness isn’t a personality trait; it’s a practice.

  • Mindset: Start reframing risk not as a threat, but as an investment in your future growth.
  • Strategy: Use a barbell approach. Protect your core (pay the bills, maintain key relationships), but make bold, asymmetric bets on the edges.
  • Practice: Constantly ask yourself, “Am I living and working in a way that is bold enough to generate bold outcomes?”

Your Challenge

That U-Haul journey to a California wasn’t just a trip; it was a decision to choose the unknown over the known. Life’s gravity will always pull you toward safety, and the only way to break free is through conscious, bold moves—big and small.

So, what’s your U-Haul moment? What bold move are you avoiding right now?

Reflections on AI: Context and Memory – The Gateway to AGI

Introduction: Why AGI is Different from Narrow AI

Today’s frontier models are wonders of engineering. They can write code, draft legal arguments, and create poetry on command. But for all their power, they are fundamentally transient. Once a session ends, the model resets. The insights, the rapport, the shared understanding—it all vanishes. It’s like having a brilliant conversation with someone who develops amnesia the moment you walk away.

This is the core limitation of Narrow AI. Artificial General Intelligence (AGI), the long-sought goal of creating a truly autonomous and adaptive intelligence, requires something more: persistence. AGI must have the ability to remember, adapt, and apply knowledge not just within a single conversation, but over time. True intelligence emerges when raw predictive power is paired with persistent context and memory.

A Brief History: AI Without Memory

The quest for AI has been a story of brilliant but forgetful machines. Each era pushed the boundaries of computation but ultimately fell short of creating lasting intelligence.

  • Expert Systems (1980s): These were the first commercial AIs, functioning like digital encyclopedias. They operated on vast, hard-coded rule-based systems. While effective for specific tasks like medical diagnosis, they had no memory of past interactions and couldn’t learn from experience.
  • Deep Blue (1997): IBM’s chess-playing supercomputer famously defeated world champion Garry Kasparov. It could analyze hundreds of millions of positions per second, a monumental feat of brute-force computation. Yet, each game was a clean slate. Deep Blue had no memory of Kasparov’s style from previous matches; it was a tactical genius with zero long-term continuity.
  • Early Machine Learning (2000s): The rise of statistical models brought pattern recognition to the forefront. These systems could classify images or predict market trends but were narrow and forgetful. A model trained to identify cats couldn’t learn to identify dogs without being completely retrained, often forgetting its original skill in a process known as “catastrophic forgetting.”
  • Modern LLMs: Today’s large language models possess massive context windows and demonstrate emergent reasoning abilities that feel like a step-change. Yet, they remain fundamentally stateless. Their “memory” is confined to the length of the current conversation. Close the tab, and the world resets.

The takeaway is clear: across decades of innovation, AI has lacked true continuity. Context and memory are the missing ingredients.

Context as the Fuel of Intelligence

If intelligence is an engine, context is its high-octane fuel. We can define context as an AI’s active working state—everything that is “in mind” right now. It’s the collection of recent inputs, instructions, and generated outputs that the model uses to inform its next step.

In recent years, context windows have exploded, growing from a few thousand tokens to over a million. Models can now process entire codebases or novels in a single prompt. They are also becoming multimodal, ingesting text, images, and audio to build a richer, more immediate understanding of the world.

A useful analogy is to think of context as RAM. It’s temporary, volatile, and absolutely vital for processing the task at hand. But just like RAM, its contents expire. Without a mechanism to save that working state, intelligence resets the moment the power is cut.

Memory as the Backbone of Learning

This is where memory comes in. Memory is the mechanism that transforms fleeting context into lasting knowledge. It’s the backbone of learning, allowing an intelligence to build a persistent model of the world and its place in it.

We can draw parallels between human and AI memory systems:

  • Short-Term / Working Memory: This is analogous to an AI’s context window—the information currently being processed.
  • Episodic Memory: This involves recalling specific experiences or past events. In AI, this is mirrored by storing conversation histories or specific interaction logs in vector databases, allowing a model to retrieve relevant “memories” based on semantic similarity.
  • Semantic Memory: This is generalized knowledge about the world—facts, concepts, and skills. This is what LLMs are pre-trained on, but the goal of continual learning is to allow models to update this semantic memory over time without starting from scratch.

Memory is what allows an AI to move beyond one-off transactions. It’s the bridge that connects past experiences to present decisions, enabling true learning and adaptation.

Why Context + Memory Together Are Transformational

Separately, context and memory are powerful but incomplete. It’s their synthesis that unlocks the potential for higher-order intelligence.

  • Context without memory is a clever amnesiac. It can solve complex problems within a given session but can’t build on past successes or learn from failures.
  • Memory without context is a passive archive. A database can store infinite information, but it can’t reason about it, apply it to a new problem, or act on it in real time.

When fused, they create a virtuous cycle of adaptive, continuous reasoning. The system can hold a real-time state (context) while simultaneously retrieving and updating a persistent knowledge base (memory). A better analogy combines the previous ones: context is the CPU + RAM, processing the present moment, while memory is the hard disk, providing the long-term storage that gives the system continuity and depth.

Case Study: From Jarvis to Real-World Architectures

Perhaps the best fictional illustration of this concept is Tony Stark’s AI assistant, Jarvis. While still science fiction, the principles that make Jarvis feel like a true AGI are actively being engineered into real-world systems today.

  • Context as Real-Time Awareness: Jarvis’s ability to multitask—monitoring the Iron Man suit, Stark Industries, and geopolitical threats simultaneously—is a conceptual parallel to the massive context windows of modern models. For example, Google’s Gemini 1.5 Pro demonstrated a context window of 1 million tokens, capable of processing hours of video or entire codebases at once. This mirrors Jarvis’s immense capacity for real-time situational awareness.
  • Memory as Persistent Knowledge: Jarvis’s deep memory of Stark’s habits, history, and humor is now being approximated by Retrieval-Augmented Generation (RAG) architectures. As detailed in research from Meta AI and others, RAG systems connect LLMs to external knowledge bases (like vector databases). When a query comes in, the system first retrieves relevant documents or past interactions—its “memories”—and feeds them into the model’s context window. This allows the AI to provide responses grounded in specific, persistent information, much like how Jarvis recalls past battles to inform present strategy.

The takeaway is that the magic of Jarvis is being deconstructed into an engineering roadmap. The fusion of enormous context windows (the “present”) with deep, retrievable knowledge bases (the “past”) is the critical step toward creating an AI with a genuine sense of continuity.

Architectures Emerging Today

The good news is that we are moving from science fiction to engineering reality. The architecture for persistent AI is being built today.

  • Extended Context Windows: Models from companies like Anthropic and Google are pushing context windows to a million tokens and beyond, allowing for much longer and more complex “sessions.”
  • Memory-Augmented Agents: Frameworks like LangChain and LlamaIndex are creating systems that allow LLMs to connect to external vector databases, giving them a persistent long-term memory they can query.
  • Hybrid Neuro-Symbolic Models: Researchers are exploring models that blend the pattern-recognition strengths of neural networks with the structured, logical reasoning of symbolic AI, creating a more robust framework for knowledge representation.
  • Continual Learning: The holy grail is developing agents that can continuously update their own parameters in real time based on new information, truly learning as they go without needing to be retrained.

How Close Are We? An Opinion

While the architectural components for a persistent AI are falling into place, it’s crucial to distinguish between having the blueprints and having a finished skyscraper. We are in the early stages of the construction phase—the foundation is poured and the first few floors are framed, but the penthouse is still a long way off.

  • The Good News: Concepts like Retrieval-Augmented Generation (RAG) and massive context windows have moved from research papers to practical frameworks in just a few years. We now have the basic tools to give models a semblance of long-term memory. This is a monumental step forward. This rapid acceleration from theory to practice is a clear example of the Law of Accelerating Returns, a concept I explored in a previous post.
  • The Hard Reality: The primary challenge is no longer about possibility but about integration and autonomy. Current RAG systems are often brittle and slow. Determining what information is truly “relevant” for retrieval is a complex challenge in itself. More importantly, we haven’t solved continual learning. Today’s agents “read” from their memory; they don’t truly “learn” from it in a way that fundamentally reshapes their internal understanding of the world. They are more like interns with access to a perfect library than seasoned experts who have internalized that library’s knowledge.

We are likely years, not months, away from systems that can learn and adapt autonomously over long periods in a way that truly resembles human-like persistence. The scaffolding is visible, but the hard work of seamless integration, optimization, and achieving genuine learning has only just begun.

The AGI Threshold

When these pieces come together, we will begin to approach the AGI threshold. The key ingredients of general intelligence can be framed as follows:

  1. Context: The ability to reason effectively in the present moment.
  2. Memory: The ability to persist knowledge and learn across time.
  3. Agency: The ability to act on that reasoning and learning to achieve goals and improve oneself.

Crossing the threshold from Narrow AI to AGI won’t be about a single breakthrough. It will be an evolution toward systems that can “live” across days, months, or even years, learning continuously from their interactions with the world and with us.

Risks & Ethical Dimensions

Of course, creating AI with perfect, persistent memory introduces profound ethical challenges.

  • Privacy: What should an AI be allowed to remember about its users? A system that never forgets could become the ultimate surveillance tool.
  • Bias and Malice: False or malicious memories, whether introduced accidentally or deliberately, could permanently shape an AI’s behavior in harmful ways.
  • The Importance of Forgetting: Human memory decays, and this is often a feature, not a bug. Forgetting allows for forgiveness, healing, and moving past trauma. A perfectly eidetic AI may lack this crucial aspect of wisdom.
  • Governance: This new reality will demand robust governance frameworks, including clear audit trails, explicit user consent for memory storage, and a “right to be forgotten” that allows users to wipe an AI’s memory of them.

Conclusion: Context + Memory as the True Gateway

For years, the race toward AGI has been framed as a race for scale—bigger models, more data, more compute. While these are important, they are not the whole story. The true gateway to AGI will not be opened by raw computational power alone, but by the development of persistent, contextual intelligence.

The Jarvis analogy, once pure fantasy, is now a design specification. It shows us what’s possible when an AI can remember everything yet act on that knowledge with immediate, contextual awareness. The great AI race of the next decade will not be about building the biggest brain, but about building the one with the best memory.