We’re Back in the Garage

ClawdBot, MoltBot, OpenClaw, MoltBook — and the Return of the Homebrew Energy

It was 1995.

I was sitting in my dorm room at the University of Richmond — beige tower PC humming like it had something important to prove, CRT monitor warming the room, Ethernet cable stretched across the carpet like a tripwire for the future. My dorm had high‑speed internet. Which, at the time, felt like I had been handed the nuclear codes.

WiFi wasn’t really a thing yet. You wanted to be online? You needed a Network Interface Card (NIC) card. You needed drivers. You needed someone willing to open up your computer without flinching.

Luckily, I was that someone.

I started installing NIC cards for other students — popping open cases, sliding hardware into slots, tightening screws like I was assembling possibility itself. I’d crawl under desks, string cable, configure IP settings. I made absurdly good money for a college kid. Basically the unofficial ISP of campus. I serviced most of the athletes and even helped them with their computer science homework — haha.

And when I wasn’t running my tiny networking empire? I was in the Jepson Hall computer lab — which at the time felt like NASA had decided to open a satellite office in Richmond. We had NeXT machines. SPARC workstations. Hardware that looked like it belonged in a sci‑fi movie and sounded like a small jet engine when it spun up. The machines were glorious. Fast. Shiny. Capable of things that felt borderline supernatural. At one point I had a couple of them cranking away on some early neural net experiments — primitive by today’s standards, but intoxicating back then. Let’s just say the sysadmins were not thrilled with my interpretation of “appropriate resource usage.” I learned a valuable lesson about shared compute. And about how quickly you can get a polite but firm email from IT.

We built anything we could think of.

Websites for local businesses. Student groups. Experiments that probably made no sense but felt revolutionary at 2am. We weren’t chasing product‑market fit. We were chasing the edge of what was possible.

Looking back, it felt like our own version of the Homebrew Computer Club. Different decade. Same energy.

The Original Hackers

A dimly lit garage workspace featuring multiple computer monitors displaying code, a whiteboard with memory concepts, and scattered electronic components.

In 1975, a group of curious, slightly obsessive engineers started meeting in Silicon Valley to talk about microprocessors. They passed schematics around like contraband. They believed computing shouldn’t belong to institutions. It should belong to individuals.

Two of those people were Steve Jobs and Steve Wozniak.

The club didn’t just produce companies.

It produced culture.

A culture of show‑and‑tell hacking. Of building first and asking permission never. Of staying up too late because the machine might do something new if you just tweak one more thing.

Technology revolutions rarely start with polished strategy decks.

They start with people who can’t stop tinkering.

Web 1.0: When HTML Felt Like Magic

In 1995, most people had never heard the word “browser.” Netscape felt like something out of Star Trek. If you knew how to use a tag correctly, you were basically Gandalf.

We built pages by hand. We viewed source. We copied. We remixed. We stayed up until sunrise arguing about font sizes and background colors like it was geopolitics.

There was no playbook.

There was just momentum.

San Francisco: 10am to 10pm (and Then Some)

A few years later I moved to San Francisco, right as the dot‑com wave was cresting.

We’d get into the office around 10am and stay until 10pm. And then — because we were apparently incapable of moderation — we’d roll to Buddha Bar, decompress for a hot second, and then go home and get back on our computers.

Exterior view of Buddha Lounge, a dive bar in San Francisco, featuring neon signs and vibrant lighting.

No one was forcing us.

We were just in it.

We were building some of the first serious commercial websites. Infrastructure that would become ordinary later felt extraordinary then. Every deploy felt like lighting a match in the dark.

The web wasn’t inevitable yet.

We were helping make it inevitable.

The “Any‑to‑Any” Club

At Scient — a fast‑growing digital agency during the late‑90s dot‑com boom that built large‑scale web experiences and platforms for Fortune 500 brands trying to figure out this whole “internet” thing — we started something we called the “Any‑to‑Any” club. The name alone should tell you everything you need to know about our collective restraint.

We hacked on early mobile devices before the mobile era had officially begun. PDAs. Proto‑smartphones. Hardware that felt like it had arrived from 2007 by accident.

Half the time the devices barely worked.

Which made it even better.

We weren’t building polished apps. We were trying to answer a more fundamental question: what happens when computing leaves the desk?

Every major wave has this phase — a scrappy moment when a handful of builders feel the tremor before everyone else notices the earthquake.

And Then… AI

A dynamic, abstract visual representation of light beams radiating outward from a central bright source, showcasing vibrant blue and orange colors against a dark background.

Fast forward to now.

ClawdBot / MoltBot / OpenClaw  

This progression matters because it marks a shift from prompt-driven novelty to system-driven behavior. From one-off answers to persistent agents. From “look what it said” to “look what it did.”

If you know, you know.

If you don’t — imagine 1995, except the browser can reason.

We’re back in that early‑wave energy. APIs duct‑taped together. Prompt engineering that feels more like alchemy than software development. Agents talking to agents. Systems writing code that writes systems.

It’s messy.

It’s chaotic.

It’s glorious.

OpenClaw and the Architecture of Memory

One of the most interesting projects to emerge from this wave is OpenClaw’s Memory Architecture.

Not because it wrapped an LLM in a shiny interface.

But because it treated memory as architecture — not as an afterthought.

For years, LLMs have been brilliant but stateless. Goldfish with superpowers. Every conversation reset at the edge of the context window. You could simulate memory with clever prompting, but it was fragile. Expensive. And fundamentally constrained by token limits.

OpenClaw approached the problem differently.

Instead of stuffing everything back into the prompt, it externalized cognition.

At a high level, the system separates memory into distinct layers:

1. Working Memory
Short‑lived context tied to the current task. Think: scratchpad reasoning, tool outputs, intermediate plans. This lives close to the agent loop and gets pruned aggressively.

2. Episodic Memory
Structured records of prior interactions — who the user is, what they’re building, decisions made, constraints discovered. These are stored outside the model context, indexed semantically, and retrieved on demand.

3. Long‑Term Knowledge
Durable artifacts: documents, prior code, system state, embeddings of previous sessions. This layer acts more like a knowledge base than chat history.

Under the hood, this means a few important architectural shifts:

  • Vector indexing for semantic recall — prior conversations and artifacts are embedded and stored so retrieval is meaning‑based, not keyword‑based.
  • Selective retrieval pipelines — instead of blindly rehydrating the entire past, the agent queries memory with intent and pulls back only what’s relevant to the current goal.
  • Memory summarization and compaction — episodic traces are periodically distilled into higher‑order summaries to prevent unbounded growth.
  • Separation of cognition and storage — the LLM reasons; external systems persist.

That last one is subtle but profound. The model is no longer pretending to remember. It actually remembers — via infrastructure. Which changes the agent loop entirely.

Instead of:

User → Prompt → Response → Forget

You now have:

User → Retrieve Memory → Reason → Act → Update Memory → Evolve

That feedback loop is what makes agents feel continuous. And continuity is what makes them feel alive. It’s still not consciousness. It’s not self‑awareness. There’s no internal subjective state.

But there is persistence.

And persistence compounds.

Now, it’s worth contrasting this with simpler RAG (Retrieval‑Augmented Generation) implementations — because on the surface, they look similar.

Basic RAG works like this:

  1. Embed documents.
  2. Store them in a vector database.
  3. Retrieve top‑K matches for a query.
  4. Stuff them back into the prompt.

It’s powerful. It unlocked an entire generation of AI apps. But it’s fundamentally document‑centric. RAG answers questions about things that already exist. OpenClaw’s memory model is agent‑centric. It’s not just retrieving documents — it’s maintaining state.

Here’s the difference in practice:

With simple RAG:

  • The model retrieves relevant docs.
  • It answers.
  • The session ends.
  • Nothing meaningfully changes in the system.

With OpenClaw‑style memory:

  • The agent retrieves prior decisions.
  • It reasons in the context of past goals.
  • It performs actions.
  • It writes back new state.
  • The system evolves.

RAG augments answers. Persistent memory augments behavior. That’s a big leap.

RAG makes models smarter in the moment. Memory makes agents smarter over time.

And once an agent accumulates context across weeks or months — product decisions, user preferences, architectural tradeoffs — you’re no longer interacting with a stateless model.

You’re interacting with a continuously updating system.

That’s the shift.

Architecture is destiny. And memory is architecture.

MoltBook and the Accidental Consciousness Moment

Then came MoltBook.

And for a brief, extremely online moment, parts of the internet thought we were on the verge of computer consciousness.

Screenshots circulated. Threads exploded. People asked, in earnest:

“Is this thing self‑aware?”

No.

But it felt like it might be.

Which tells you more about humans than it does about machines.

Why did this moment hit so hard?

1. The Illusion of Continuity

When a system has memory and tone and context, we project identity onto it. We are wired to detect agency. Even when it’s not there.

2. Public Experimentation

Unlike 1995, this wave is unfolding in public. Demos go viral. Iterations are visible. The hackathon is livestreamed.

3. Velocity

What used to take years now happens in weeks. Sometimes days. One framework spawns another. One experiment becomes a movement.

ClawdBot becomes MoltBot becomes OpenClaw thrown into MoltBook.

The stack is assembling itself in real time.

Why This Feels So Familiar

There’s a very specific electricity in the air when a technology wave is still fragile. Before standards. Before consultants. Before the MBA slide decks explaining why it was obvious all along.

We are in that electricity.

Builders are wiring up memory layers. Orchestrating agents. Inventing new abstractions on top of reasoning models that didn’t exist two years ago.

And the machine is helping build the machine.

That’s new.

Recursive creation.

Systems that help us design better systems that build better systems.

It’s strange.

It’s slightly terrifying.

And it’s the most alive I’ve felt in a long time.

We Are Living in an Extraordinary Moment

I’ve seen a few waves.

Web 1.0. Dot‑com. Mobile. Cloud. SaaS.

This one feels different.

Not because it’s bigger.

But because it’s more intimate.

The Homebrew Computer Club didn’t know they were inventing the personal computing industry. They just showed up and shared what they built.

That’s what this feels like.

People building in public. Sharing memory architectures. Posting screenshots of agents behaving strangely. Laughing at bugs. Iterating at a pace that feels unsustainable — until it becomes normal.

Somewhere right now, someone is in a dorm room with a messy Ethernet cable on the floor, hacking on something that will look obvious in ten years.

And they probably don’t even know it yet.

-rjm

The Holiday Break That Broke the Software Development Lifecycle

The 2025 holiday break was supposed to be slow. It has been a long year of hard work and I needed to take a break.

My parents are older now. Cruises are their speed. Elevators instead of staircases. Early dinners. My dad triple-checking the daily schedule like it’s a mission briefing. My mom insisting we all order dessert because “it’s free”.  Time feels different when you realize there are fewer of these trips ahead than behind.

I told myself I would unplug. I even announced it at dinner the first night. “No laptop this week.” My dad nodded approvingly. My mom smiled like she didn’t quite believe me.

I didn’t last very long.

On the second afternoon, I slipped away with my laptop and headphones and found a quiet spot at the back of the boat. Espresso in hand. Salt air hanging thick. The low mechanical hum of the engines vibrating through the deck. The wake trailing behind us like a white scar across a blue horizon.

I opened my terminal.

I started catching up.

Claude Code, Claude 3.5 Sonnet. GPT-4.1 and GPT-4o. Early Codex CLI previews. Terminal-native agents gaining traction. IDEs that no longer just suggested code but executed meaningful chunks of it. Parallel task execution in the cloud. Multi-step tool use that didn’t collapse after the second function call.

At first, it felt like noise. 2025 had trained us to expect noise. But then I started testing.

I delegated something non-trivial. It finished. Cleanly.

I handed off something messier. It reasoned through it. Returned with structure.

I felt that strange combination of excitement and unease — the same feeling I had when I was working on code complete and refactoring in Borland JBuilder. Felt like magic.

This wasn’t incremental improvement.

Something had crossed a line.

The Delegation Threshold

There is a moment in any technological system when progress stops being additive and starts being structural.

The iPhone wasn’t merely a better phone — it reorganized behavior. AWS wasn’t cheaper hosting — it changed who could start a company. ChatGPT didn’t just answer questions — it altered how people thought about knowledge itself.

Over the final months of 2025, engineering crossed a similar boundary.

I call it the Delegation Threshold.  The point at which models become reliable enough to safely delegate meaningful work.  Not perfect. Not autonomous. Not magic.  But dependable enough that you stop hovering over every output.  Before the threshold, AI assisted me.  After the threshold, I found myself assigning work.  That subtle shift — from helper to delegate — reorganizes the system.  It reorganized me.

When Delegation Becomes Real

Individually, the releases were impressive.  Long-context windows that could actually be trusted. Tool use that didn’t unravel after the second function call. Parallel execution that made decomposition practical instead of aspirational. IDE-native agents that moved from autocomplete to autonomy.  Collectively, they crossed a reliability boundary.  When you can hand a model a non-trivial task and reasonably expect it to finish, you stop thinking in files.  You start thinking in systems.  Tasks decompose differently. Parallelism becomes natural. Orchestration becomes the new bottleneck.

The wildly stupid term “vibe coding” was making way for “agentic engineering”.

The scarcity shifts.

Code is no longer the constraint. Judgment is.

That realization didn’t feel triumphant.

It felt destabilizing.

Grassroots Discovery

What fascinated me most wasn’t the press releases.  It was what developers started doing once delegation became viable.  Recursive loops appeared — bash scripts feeding output back into models until the system converged. What looked chaotic was actually a crude form of evaluation harnessing.  

How our brains work was surfacing in the methods being discovered.

Agent swarms emerged — one human intent spawning multiple concurrent threads across files, features, and research tasks.  I tried it myself.  It was messy. Threads collided. Context drifted. One agent confidently refactored something another agent had just rewritten.

And yet — it worked often enough that you couldn’t ignore it.

When delegation becomes safe, orchestration becomes the craft.

The community was rediscovering harness engineering before it had a name.

A Collective Synchronization

Why did it feel sudden?

Because thousands of engineers had something rare during that break: uninterrupted time.

No sprint planning. No performance reviews. No backlog theater.

Just space.

Space to test. To push. To hand something real to a model and see if it came back intact.

When everyone returned, the tone had changed.

Group chats weren’t debating whether agents were useful. They were debating how to structure them.

The tools hadn’t just improved. The collective mental model had updated.

The Structural Repricing

When early 2026 reports began quantifying adoption curves and productivity deltas, the data validated the intuition.  Agent usage wasn’t experimental anymore. It was habitual.  I felt it in my own workflow. Tasks I would have blocked off half a day for were now something I delegated while I thought about architecture. And then came formalization.  OpenAI described building a product with near-zero human-written code. Engineers we ren’t typing implementation — they were designing constraints, evaluation loops, and execution environments. Harnesses.

That word stuck with me.

Harness.

Not autopilot. Not replacement.

Harness.

This is what happens after a threshold crossing.

When execution becomes cheap, environment design becomes expensive.

Feature velocity stops being durable advantage. System durability does.

The SDLC Inverted

We like to say we sped up the Software Development Lifecycle.  But sitting on that cruise ship, watching agents complete tasks while my parents took a nap, I realized that wasn’t quite right.

We didn’t speed it up.

We inverted it.

The old world moved in phases: requirements, design, code, test, deploy.

The new world begins with intent and harness design. Agents execute. Evaluation runs continuously. Humans orchestrate and judge.

Phase boundaries blur.

Code stops being the bottleneck. Evaluation and architecture take its place.

That isn’t speed.

That’s reorganization.

The Identity Question

This is the part I keep coming back to.

If writing code is no longer the center of gravity, what is?

I learned to love engineering by writing it line by line. By debugging at 2 a.m. By feeling the satisfaction of something compiling cleanly after hours of friction.

What becomes craftsmanship in a world of delegation? How do junior engineers learn the texture of a system if they rarely touch its internals? Where does pride migrate?

Delegation changes identity, not just workflow.

Every structural shift does.

What Comes Next

Crossing a threshold does not eliminate constraints. It relocates them.

Orchestration complexity can explode. Evaluation can become the new choke point. Reliability can plateau.

But thresholds, once crossed, rarely uncross.

The genie isn’t the story.

The threshold is.

And once crossed, systems reorganize themselves.

On that cruise ship, I closed my laptop and went back inside. My dad sent me a text and asked if I was done working.

“Yeah,” I said.

But I wasn’t done thinking.

We are still early on this journey.

Let’s go.

Move 37 and the Shape of What’s Next

There’s something about forced stillness that creates space for the unexpected.

Molly was a few weeks into recovery from her second ankle surgery—the one where she “won the ankle injury lottery in the worst way possible.” Her boot-clad ankle propped up next to me on the couch, she wasn’t going anywhere. Neither was I. So we did what any reasonable father-daughter duo does when escape isn’t an option: we binged two documentaries about artificial intelligence.

First, AlphaGo—the 2017 film about DeepMind’s AI beating the world champion at Go. Then The Thinking Game—the 2024 documentary that follows DeepMind’s broader quest toward artificial general intelligence, filmed over five years by the same team.

What I didn’t expect was that this double feature would turn into one of the best conversations we’ve ever had.

Molly is a Computational Biology major. I’m a lifelong computer science nerd. Our worlds were about to collide in the best possible way. (Fun aside: one of her friends from MIT—his dad appears in one of the documentaries. We got a good laugh out of that.)

Why Go Matters (And Why No One Thought This Would Happen)

If you’re not familiar with Go, here’s the short version: it’s a 2,500-year-old board game that makes chess look like tic-tac-toe.

Chess has roughly 10^47 possible game states. Go has 10^170. For perspective, there are approximately 10^80 atoms in the observable universe. Go has more possible positions than there are atoms—by a factor of 10^90. Let that sink in.

This isn’t just trivia. It means you can’t brute-force Go. You can’t calculate every possibility the way Deep Blue did against Kasparov in 1997, evaluating 200 million positions per second. That approach simply doesn’t work here. The game is too vast. Too deep.

For decades, the best Go programs were… embarrassing. They played at the level of a decent amateur, routinely getting crushed by club players. Experts confidently predicted that AI beating a professional Go player was 10-20 years away. Some said it might never happen.

In 2016, DeepMind did it anyway.

How AlphaGo Actually Works

Neural Networks: Teaching Intuition

AlphaGo wasn’t programmed with rules about how to play Go. Nobody sat down and wrote “if your opponent plays here, respond there.” That approach had been tried for decades. It didn’t work.

Instead, AlphaGo was trained.

First, it studied millions of games played by human masters. The neural network learned to “see” the board—not as a grid of black and white stones, but as patterns. Shapes. Flows. The kind of intuition that takes a human player decades to develop, encoded in the weights of a neural network.

This is the key insight that changed everything: intuition can be learned. It’s not magic. It’s not some mystical human quality that machines can never possess. It’s pattern recognition at scale.

Reinforcement Learning: Playing Itself

But learning from humans only gets you so far. Humans, after all, are limited.

So after learning from human games, AlphaGo started playing against itself. Millions of games. Billions of moves. Twenty-four hours a day, at speeds no human could match.

This is reinforcement learning—trial and error at superhuman velocity. And here’s the kicker: through self-play, AlphaGo discovered strategies that no human had ever seen. Not because they were wrong. Because we never thought to try them.

The machine had started to see things we couldn’t.

Monte Carlo Tree Search: Guided Exploration

AlphaGo doesn’t evaluate every possible move—that’s mathematically impossible, remember? Instead, it uses its neural network intuition to guide its search.

Think of it like this:

  • Policy Network: “What move looks promising?” (Intuition)
  • Value Network: “Who’s winning from this position?” (Evaluation)
  • Monte Carlo Tree Search: “Let me simulate a bunch of games from here to check.” (Verification)

It’s intuition combined with calculation. The machine equivalent of a grandmaster “feeling” that a move is right, then verifying it with deep analysis.

Human experts have both systems too—the gut and the grind. AlphaGo unified them into something more powerful than either alone.

Move 37: The Moment Everything Changed

Game 2 of the match against Lee Sedol. If you haven’t seen the documentary, go watch it. If you have, you know exactly what I’m about to describe.

Lee Sedol is one of the greatest Go players in history. Eighteen world championships. A player of profound intuition and legendary fighting spirit. He sat across from AlphaGo expecting a battle. He got something else entirely.

Move 37.

AlphaGo places a stone on the fifth line—a move that looks, to the trained human eye, wrong. Commentators are confused. Experts call it a mistake. Lee Sedol leaves the room, visibly shaken. The move violates centuries of accumulated Go wisdom.

And then it wins the game.

Move 37 wasn’t in any textbook. It wasn’t copied from any human game in AlphaGo’s training data. The machine had discovered something new about a game humans have played for 2,500 years.

What does it feel like to watch a machine be creative? I still don’t have a great answer. But I know it changes how you think about intelligence—artificial and otherwise.

From Go to Protein Folding: Where Our Worlds Cross

This is where The Thinking Game picks up the story.

Here’s the thing about DeepMind: they weren’t just trying to win at board games. Go was a proving ground. A demonstration. The real target was always bigger.

Enter AlphaFold. And enter my daughter’s world.

The Protein Folding Problem

Proteins are the workhorses of biology. They do almost everything—carry oxygen in your blood, fight infections, make your muscles contract, replicate your DNA. And every protein is built from a chain of amino acids that folds into a specific three-dimensional shape.

Here’s the critical insight: the shape is the function. A protein’s 3D structure determines what it does. Get the shape wrong, and the protein doesn’t work. Misfolded proteins cause diseases like Alzheimer’s, Parkinson’s, and cystic fibrosis.

The problem? We know the amino acid sequences for over 200 million proteins. But determining the 3D structure experimentally—using X-ray crystallography, cryo-electron microscopy, or nuclear magnetic resonance—is brutally slow and expensive. In 60 years of global scientific effort, we had solved about 170,000 structures.

Predicting how a protein folds from its sequence alone? That was the “50-year grand challenge” of biology. The Mount Everest of molecular science.

CASP: The Olympics of Protein Prediction

Every two years, computational biologists compete in CASP—the Critical Assessment of Structure Prediction. It’s basically the Olympics of protein folding. Teams submit predictions for protein structures that have been experimentally determined but not yet published. Then they get graded.

For years, scores hovered around 40 out of 100 for the hardest targets. Progress was incremental. Slow. Scientists would publish papers celebrating a 2-point improvement.

Then AlphaFold 2 showed up in 2020.

The “Holy Shit” Moment

There’s no other way to describe it.

AlphaFold 2 scored above 90 on two-thirds of the targets. Some predictions were so accurate they were essentially indistinguishable from experimental results. The competition wasn’t close. It wasn’t even a competition anymore.

The judges called it “astounding.” One researcher said it was “like landing on the moon.” Another said protein structure prediction had been “solved.”

I looked at Molly. This is her field. Transformed overnight.

How AlphaFold Works

Like AlphaGo, AlphaFold uses neural networks. But the architecture is different—it’s built on attention mechanisms, similar to the transformers that power GPT and other large language models.

The key insight is co-evolution.

Here’s the intuition: if two amino acids that are far apart in the sequence consistently mutate together across many different species, they’re probably close together in the 3D structure. Evolution leaves fingerprints. AlphaFold learned to read them.

The system analyzes millions of protein sequences, looking for these co-evolutionary patterns. Then it uses that information—combined with geometric reasoning and iterative refinement—to predict the spatial relationship between every pair of amino acids.

It’s pattern recognition. The same fundamental idea as AlphaGo—but applied to the language of life itself.

AlphaFold 3 and the Nobel Prize

In 2024, DeepMind released AlphaFold 3. It doesn’t just predict individual protein structures—it predicts how proteins interact with DNA, RNA, and small molecules. The implications for drug discovery, gene therapy, and understanding disease are enormous.

Oh, and Demis Hassabis and John Jumper won the Nobel Prize in Chemistry for their work on AlphaFold. No big deal. Just the highest honor in science for a system that started with a board game.

What It Meant to Watch This Together

Here’s what I didn’t tell Molly while we watched: I was so damn excited for her.

She’s walking into a world where the tools to understand life at the molecular level are suddenly, radically more powerful. The intersection of computer science and biology isn’t a niche curiosity anymore—it’s the frontier. And she’s not watching it from the sidelines. She’s studying it. She’s going to use it.

I’ve spent my career in tech, watching waves come and go. I’ve seen hype cycles inflate and collapse. But this one feels different. Her timing is impeccable.

We didn’t plan this documentary double feature as a “teaching moment.” It was just couch time—her ankle in a boot, me with the remote, nowhere to be. But somewhere between Move 37 in AlphaGo and the protein folding breakthrough in The Thinking Game, something clicked.

Her world and my world aren’t separate anymore. They’re the same world.

And she’s going to take it places I can’t even imagine.

The Thread

These two documentaries tell one continuous story—a thread that runs from a board game in Seoul to a protein database that covers all of life.

It’s not about computers being smarter than humans. It’s about building tools that let us see what we couldn’t see before.

AlphaGo showed us a move no human had imagined in 2,500 years of play. AlphaFold showed us the shapes of 200 million proteins that would have taken centuries to solve experimentally.

What’s next? I don’t know.

But I have a feeling Molly is going to help figure it out.

And I’ll be cheering from the couch—boot or no boot.

2025 Year-in-Review: Read. Write. Code.

2025 felt like one of those years where you don’t fully understand what happened until you look backward and realize: oh… that was a phase shift.

The year had a rhythm to it. The front half was dominated by reading — books, papers, blog posts, half-formed ideas scribbled into notes apps. The back half turned into writing, coding and building things that didn’t exist before.

AI was the through-line. Not as a trend. Not as a tool. But as a force.

It reminded me of the 1990s — long nights, endless curiosity, learning by building, and that quiet feeling that something foundational was changing. Web 1.0 energy, but faster. Much faster.

Work, especially at O2E Brands, felt like skiing the same black diamond run for the tenth time. Familiar terrain. Different snow. Same steep pitch. New ice patterns. And no matter how well you know the run, there are always shifts and turns you don’t expect — moments that force you to react, adjust, and stay fully present. You still need to stay on edge.

This was a year of reading, writing, and coding — in that order, and then all at once.

Read

I read a lot in 2025. Mostly in the first half. Books gave me the long arcs — the kind of perspective you don’t get from timelines or hot takes.

Books I Read in 2025

Favorite Books of the Year

Reading Themes

A few patterns emerged:

  • Acceleration is not optional. It’s structural.
  • Agency matters more than raw intelligence.
  • Understanding beats adoption.
  • Strategy still matters — maybe more now than ever.
  • AI is power infrastructure, not a feature.

Trying to keep track of every white paper, blog post, or research thread was impossible. The firehose won. That itself felt like a signal.

Write

If reading was about absorbing, writing was about sense-making.

Most of my writing in 2025 happened in the second half of the year, once the ideas had compressed enough to feel coherent.

Blog Posts in 2025

Earlier

Writing Themes

Across all of it, a few ideas kept resurfacing:

  • AI as physics, not hype
  • Focus as a competitive advantage
  • Context and memory as primitives
  • Enterprise transformation as inevitable
  • Human judgment still matters — a lot

Writing became less about explaining what and more about explaining why it feels different.

Code

Coding felt fresh again.

That alone is worth pausing on.

I found myself deep in computer science concepts — state machines, orchestration, evaluation, memory — like I was back in college. Only now, the feedback loops were instant.

Coding Projects & Experiments

  • Vibe coding with Cursor and Anthropic Claude
  • Vibe coding with GitHub Copilot
  • LangChain, LangGraph, and LangSmith prototypes
  • OpenAI API multi-modal applications
  • Custom LLM hosting and development

“English” has quietly become the most important programming language — but calling this “vibe coding” misses the point. It’s a dumb term because it implies looseness where there is actually rigor. You still need to understand what’s happening under the hood. The real leverage comes from knowing where to apply the tool and why — not from blindly trusting it.

At the same time, multimodal finally feels real. Text, images, audio, video — not as demos or promises, but as things you actually experience while building. That shift alone changes how software feels to write.

As a result, classic enterprise software patterns suddenly look unfamiliar. Some are breaking. Some are reforming. Many are being replaced by agent-driven workflows. And in the middle of all of that, I’m seeing pockets of productivity that are honestly mind-bending.

Closing: Read. Write. Code.

Everything feels different now — not in a loud, hype-driven way, but in a quieter, more structural sense. The kind of different you feel in your hands while you’re building, or halfway through a problem when you realize the old mental shortcuts no longer apply.

There will be people who get it, and people who don’t. And the frustrating part is that those who get it often struggle to explain it to those who don’t. Not because it’s secret, but because it’s experiential. You have to do the work to feel the shift.

The productivity is real. The leverage is real. But it only shows up when reading turns into understanding, writing turns into clarity, and coding turns into creation.

That loop — read, write, code — is what kept me grounded this year. It’s how I made sense of the acceleration without getting lost in it. And it’s why, despite all the change, the craft itself still feels familiar.

Read. Write. Code. Repeat.

Onward.

Latent Space: The Hidden Infrastructure of Intelligence

The First Time You Realize AI Sees the World Differently

Latent space is one of those concepts that feels deceptively simple but quickly becomes mind-bending the deeper you go. In my MIT coursework, the moment it truly clicked wasn’t when someone showed a diagram or equation—it was when I watched two very different inputs land right next to each other in a high‑dimensional embedding space. Suddenly, you realize: AI doesn’t see categories the way we do. It sees geometry. And that geometry is the beating heart of modern AI.

This post is my attempt to make latent space both intuitive and technically sound—a tour of the hidden mathematical world that lets AI models generalize, reason, and occasionally surprise the hell out of us. If you’ve read my earlier posts like Efficiency Reckoning or AI Is Eating Software That Is Eating the World, you’ll recognize a recurring theme: exponential capability often hides in plain sight until you learn to see the structure underneath.

What Is Latent Space?

Latent space is the compressed mathematical world where AI stores meaning. Instead of memorizing data, models learn dense vector representations that capture the essence of concepts—objects, actions, styles, emotions, operational patterns. Similar ideas cluster together; different ideas drift apart. Geometry becomes understanding.

Key ideas:

  • Embeddings: Numerical vectors that represent the meaning of inputs (words, images, tokens). Their position and direction encode semantic relationships.
  • Distance: A mathematical measure (often cosine similarity or Euclidean distance) that indicates how similar two embeddings are. Closer = more related.
  • Manifolds: Lower‑dimensional, structured surfaces within the high‑dimensional latent space where meaningful data naturally clusters. Models “discover” these during training.

Everything the model “knows” lives somewhere in this hidden space.

Why Latent Space Is AI’s Superpower

Latent spaces give models the ability to:

  • Generalize beyond what they’ve seen.
  • Recognize analogies and patterns.
  • Perform zero-shot reasoning (answer questions they were never explicitly trained on).
  • Compress knowledge into a shape that can be navigated, manipulated, and queried.

This geometry is the fuel behind why large models feel so shockingly capable. It’s the same idea I explored in The Law of Accelerating Returns, systems don’t merely improve—they reshape the surface beneath our feet. Latent space is the mathematical expression of that reshaping. We’re no longer programming rules; we’re shaping the very spaces where meaning lives.

How Latent Spaces Are Built

Latent space emerges naturally during training, driven by the model’s need to predict missing information.

The Compression Process

Self-supervision forces the model to strip away noise and preserve structure. This compression yields abstract, high-dimensional patterns that capture relationships rather than raw inputs.

Transformation Through Layers

Embeddings pass through dozens or hundreds of transformer layers. Each layer rotates, stretches, and refines meaning until stable semantic structures emerge.

The Result: A Structured World

By late training, the model has carved out clear neighborhoods for concepts—objects clustering near the actions they relate to, pricing signals gravitating toward contextual cues, operational patterns forming their own orbits.

How Latent Space Behaves

Despite being abstract and high dimensional, latent spaces exhibit surprisingly intuitive properties. Humans naturally build mental maps to navigate ambiguity, and AI does something similar—just at a scale and dimensionality far beyond our own., latent spaces exhibit surprisingly intuitive properties.

Smoothness

Small moves yield gradual changes in meaning, enabling interpolation, transformation, and reinterpretation.

Relational Structure

Directional changes encode relationships—analogy, comparison, and categorization become geometric operations.

Compositionality

Concepts can combine fluidly: an object + context + constraint forms a new point in space that the model can reason about without explicit rules.

Natural Clustering

Clusters form organically, often better than human taxonomies—but also reflecting limitations or hidden biases.

Where Latent Space Breaks Down

As magical as it feels, latent space isn’t perfect:

  • Latent collapse: Everything clusters too tightly.
  • Overfitting: Geometry becomes brittle.
  • Bias: Prejudices become encoded as spatial structure.
  • Out-of-distribution drift: The model hallucinating outside the manifold.

These limitations matter when deploying AI into real operational environments, where edge cases are everywhere.

The Future: Latent Space as the New Programming Model

We are entering an era where latent space isn’t just a byproduct of AI models—it becomes the substrate of software itself. The shift is profound: instead of writing rules, we increasingly shape geometry, influence structure, and design the conditions under which models discover meaning.

Several forces are driving this transformation:

  • Geometry replaces logic. Traditional programming encodes explicit steps; latent‑space systems embed intent, relationships, and constraints into the shape of the space itself. We’re not prescribing behavior—we’re defining the terrain.
  • Agents operate like navigators, not executors. Agents don’t follow deterministic paths. They explore, sample, and move through conceptual regions, selecting actions by proximity, similarity, and predicted outcomes. This is closer to robotics in a physical world than software in a deterministic one.
  • World models introduce simulation as a first‑class primitive. When a model can simulate consequences inside its latent space, it stops behaving like a tool and starts behaving like a planner. Software becomes anticipatory rather than reactive.
  • Developers shift from coding workflows to curating spaces. The primary task becomes shaping embeddings, conditioning behavior, tuning representations, and steering emergent structure. Infrastructure teams will manage vector spaces the way they once managed databases.

This isn’t incremental. It is a foundational rewrite of how software is conceived and built—arguably the most important transition since cloud computing abstracted away hardware. Latent space abstracts away rules themselves.

Personal Reflection: Why This Matters to Me

Returning to technical study has been a joy—and a humbling reminder that AI is a field where intuition and mathematics collide. Latent space, more than any other concept, embodies that collision in computational form.

Understanding it doesn’t just make you a better builder of AI systems.
It makes you a better interpreter of AI behavior.

And maybe a bit more forgiving when a model wanders too far off the manifold.

The Physics of Focus

I believe that focus is the single most important variable in the equation of execution. Yet, time and time again, I see executive teams unable to summon it. My theory is that true focus requires a level of strength and accountability resilience that most executives simply don’t have. It demands an “all or nothing” mentality—a trait most often found in founders and entrepreneurs, but rarely in established corporate structures.

It is the difference between movement and velocity. In physics, movement is scalar—it’s just speed. Velocity is a vector—it requires speed plus direction.

In my career leading product and engineering organizations, I have seen brilliant teams work incredibly hard, burning massive amounts of energy, only to achieve zero net displacement. Why? Because their vectors were misaligned. They were moving fast, but they weren’t moving forward.

The most idyllic example of this is the second coming of Steve Jobs at Apple in 1997.

The CEOs who preceded him—Michael Spindler and Gil Amelio—fell into the classic trap of trying to manage their way out of a crisis by adding more. They believed that capturing every niche meant building a specific product for every niche. They lacked the strength to say “no,” so they said “yes” to everything. The result was a confused portfolio of Performa, Quadra, and PowerBook models that baffled customers. They were chasing revenue, but by fracturing their focus, they were destroying value.

When Jobs returned, he saw that Apple wasn’t suffering from a lack of talent. It was suffering from a surplus of “good” ideas. The product roadmap was a mess of printers, servers, and slightly varied Macintoshes. Jobs didn’t ask the teams to double their output. He walked to a whiteboard, drew a simple 2×2 grid (Consumer/Pro, Desktop/Portable), and wiped the rest of the table clean.

He deleted 70% of the roadmap. He understood that to make the iMac great, he had to let the Newton die.

This is the hardest lesson for leaders to learn: Focus isn’t about concentration; it’s about subtraction.

I like to think of this in terms of physics.

The Equation: Pressure = Force / Area

In engineering, pressure is determined by dividing force by the surface area (P = F/A).

  • If you apply your force over a large area, you create very little pressure. You are essentially leaning against the wall.
  • If you apply that same force to the tip of a needle, you can punch through steel.

Most of us—and most companies—operate like a flat palm pushing against a wall. We spread our limited bandwidth across a dozen priorities, creating zero penetration. We mistake “busy” for “impact.”

The Entropy of Success: Why Focus is So Difficult

If the physics are so clear—reduce the surface area to increase pressure—why is it so rare to find a truly focused organization?

In my experience, five forces conspire against focus:

  1. Too Much Opportunity: Paradoxically, opportunity is the biggest threat to execution. When you have resources and options, the temptation to chase every rabbit is overwhelming.
  2. Lack of Constraints: Innovation thrives on constraints. When there are no hard limits—whether imposed by the market or self-imposed by leadership—strategies bloat.
  3. Lack of Discipline: As I mentioned, focus is an “all or nothing” game. It requires a level of behavioral discipline that is uncomfortable for most.
  4. The “General Manager” Problem: You cannot make surgical cuts if you don’t understand the anatomy. We have too many general managers who lack the detailed, functional understanding of their business. They can’t decide what to cut because they don’t truly understand how the engine works so they hedge their bets.
  5. Old School Executive Thinking: There is a pervasive belief that an executive’s job is to ask for more—more features, more markets, more growth. I fundamentally disagree. I believe an executive’s primary job is to ask for less. Less confusion. Less drag. Less surface area.

Here is a framework for how to apply this “reduction of surface area” from the boardroom down to your calendar.

1. Macro Focus: The Corporate Strategy

The Principle: Strategic Abandonment

In a company, the enemy of focus isn’t the bad idea. Bad ideas are typically self-correcting because the data screams “stop.”

The real enemy is the “good” idea. The project that is profitable enough. The feature that customers like somewhat. These are the “Zombie Projects.” They are dangerous because they consume engineering cycles, marketing budget, and leadership headspace, yet they prevent you from putting sufficient pressure on the one or two initiatives that actually drive the business.

This is Strategic Drag. Every “good” project you keep alive slows down the velocity of your “great” projects.

The Framework: Adopt a “Hell Yeah or No” policy. If a new initiative is just “reasonable,” kill it. “Reasonable” is the enemy of “Exceptional.” If you have ten priorities, you actually have zero. You need the courage to starve the profitable “good” to feed the potential “great.”

2. Operational Focus: The Digital Rebirth

The Principle: Subtraction Before Digitization

This physics equation applies just as ruthlessly to how a company operates. Most legacy corporations are clogged with “process sediment”—layers of analog habits, manual workarounds, and spreadsheet safety nets that have accumulated over decades.

When these companies try to modernize, they often fail because they try to add digital workflows on top of analog mindsets. They implement the new ERP system but keep the manual approval chain “just in case.” They build the data lake but keep the Excel trackers. This splits the organization’s focus between the past and the future.

The Framework: You cannot iterate your way from a horse to a car; you have to abandon the horse. For a process to be truly digitized, the analog version must be stopped, not just deprecated.

  • Burn the Boats: If you are digitizing an operation, you must remove the ability to revert to the manual way.
  • The Rebirth: Treat operations like products. Don’t just pave the cow path. Wipe the table clean (just like Jobs did) and ask, “If we started this company today with zero history, how would this workflow exist?” Build that, and delete the rest.

3. Meso Focus: Career & Ambition

The Principle: The Competence Trap

As we grow in our careers, we accumulate skills. We become capable of doing many things well. This leads to the Competence Trap: Just because you can do something doesn’t mean you should.

There is a mental model often attributed to Warren Buffett regarding this. He purportedly asked his pilot to list his top 25 goals, then circle the top 5. The pilot assumed he should work on the top 5 and do the other 20 in his spare time.

Buffett corrected him: The bottom 20 are the Avoid-At-All-Costs list. They are the things you care about just enough to distract you from the few things that matter.

The Framework: Audit your commitments. Identify the things you are doing simply because you are capable of doing them, not because they drive your long-term value. If it’s not in your top 5, it is a distraction. Cut it loose.

4. Micro Focus: The Daily Task

The Principle: The Context Switching Penalty

This is where the battle is won or lost on a Tuesday morning.

We live in an economy designed to fracture our attention. We deceive ourselves into thinking we can multitask, but the science says otherwise. Researchers call it “Attention Residue.” When you switch from a strategic document to check an email, a significant portion of your CPU (your brain) remains allocated to the email even after you switch back.

This is latency. If you switch contexts every 15 minutes, you are effectively operating with a fraction of your IQ. You are introducing lag into your own processing power.

The Framework: You cannot rely on willpower; you must rely on systems.

  • Defensive Design: Remove the option to be distracted. Physical separation from the phone. Turn off notifications.
  • The Deep Work Block: Schedule a 4-hour block for your hardest task. Treat it like a flight—you are buckled in, the door is closed, and there is no WiFi. This is how you generate density.

Final Thought

Focus is painful.

It hurts to kill a product line that people worked on. It hurts to force a team to abandon a familiar manual process.

But that pain is the feeling of the surface area shrinking. It is the feeling of you becoming the needle rather than the palm.

If you want to execute, stop adding. Start subtracting.

The Great Corporate Efficiency Reckoning

Introduction: Welcome to the Reckoning

We’ve reached a moment where corporate transformation is no longer optional—it’s being forced. AI isn’t just another technology wave; it’s an existential efficiency reckoning. Companies that aren’t ruthlessly exposing and eliminating inefficiencies are already behind.

There’s no hiding from this one.

The Myth of Smooth Transformation

Executives often talk about digital transformation like it’s a one-time project. It’s not.

Transformation is constant. The companies who survive are the ones who operate under continuous reinvention. But here’s the painful truth: most companies don’t change until they’re forced to. The veneer of stability has covered up shockingly deep inefficiencies.

Most workflows inside organizations aren’t even documented. Ask an ops leader to produce a process map and watch them sweat. As Peter Drucker said: “You can’t manage what you can’t measure.” But what if you can’t even find the work?

The AI Catalyst — and the Hard Truth

AI hasn’t created inefficiency. It’s simply exposed it.

We’re seeing companies push out CTOs because they “aren’t doing enough with AI.” But the problem isn’t just technical leadership—it’s organizational culture. AI isn’t an engineering project. It’s an “all hands on deck” transformation.

“You can’t outsource transformation—it’s cultural, not contractual.”

AI has thrown every department—from finance to field ops—into the blender. And now executives are realizing the painful truth: they haven’t built for this era.

Efficiency Beyond AI

Ironically, the quest to implement AI has exposed the many areas where companies are simply… bloated.

  • Work expands to fill the time and headcount you give it.
  • Revenue teams buried in manual tasks.
  • Ops teams drowning in redundancy.
  • Leaders who’d rather tweak PowerPoints than face structural inefficiency.

Sometime you have to cut until it bleeds. Then you fix the wound.

The companies that resist necessary cuts won’t survive.

Cut to Transform — or Get Cut

Look no further than Twitter’s transformation into X under Elon Musk. In 2022, Musk cut nearly 50% of the workforce in just weeks—an unprecedented move that shocked the tech world. Critics called it reckless, but it forced radical simplification: collapsing dozens of teams, canceling bloated contracts, and removing layers of slow decision-making. Whether you agree with the method or not, the rapid downsizing catalyzed operational reinvention and aggressive experimentation—at a pace impossible under the old structure.

In your own organization, the same holds true. It’s not about being heartless. It’s about being real. The days of “everyone has a job forever” are gone. And here’s the hidden risk: cultures that protect themselves from this kind of forced reinvention—clinging to the idea of being one big “family”—are often worse off. A family tolerates inefficiency and preserves comfort. A world-class team embraces accountability and expects constant growth. One survives disruption, the other gets disrupted. Meanwhile, cultures that cling to the comfort of the past—celebrating legacy wins instead of inventing their future—won’t just fall behind; they’ll be blindsided. Progress demands letting go of nostalgia and facing the work with fresh eyes.

The Agentification Era

We’re entering a phase where systems manage systems and humans orchestrate outcomes rather than manually triggering workflows. This is the era of agentification—a shift where autonomous or semi-autonomous agents are entrusted with achieving outcomes, not just performing tasks.

Frameworks like LangChainGemini Enterprise, and the emerging OpenAI Agent environments are infrastructure-level systems. They allow developers to chain models, tools, and APIs into cohesive, self-directed workflows capable of context-driven execution. In practical terms, this means a pricing agent could access CRM data, query inventory, consult policy libraries, infer customer sentiment from voice transcripts, and autonomously adjust pricing—all while handling exceptions, logging actions, and escalating when needed.

Agent-based systems are not merely “automated scripts”. They are dynamic reasoning engines managing both state and decision trees over time. But these promises come with significant complexity: memory management, context windows, hallucination control, permissions when tools perform actions (especially destructive ones), and transparent auditability for governance.

“The next great company won’t just automate workflows—it will orchestrate layers of intelligence across every function.”

A crucial challenge remains: from pilot to production. While demos often succeed in controlled environments, the real world introduces unpredictable data, unique edge cases, and higher stakes. Companies often stop at “AI-powered mock-ups” because operationalizing agent workflows requires not only technical sophistication but organizational readiness—clear workflows, defined permission tiers, and a system for rapid correction.

Most companies will never scale beyond prototypes—not because of model limitations, but because they lack the operational discipline to productionize intelligence safely and repeatably.

Where Efficiency Is Already Winning

  • Engineering productivity: Teams are achieving 2–5x development velocity using AI copilots, large-context-aware IDE extensions, and context caching. One of my own technical leads recently noted: “My problem isn’t writing code anymore—it’s code reviewing all the code that’s being generated.” Toolchains like GitHub Copilot, TabNine, and custom agents integrated into developer workflows are producing high volumes of synthetic code, shifting the bottleneck to validation, testing, and deployment pipelines. This is no longer a human-speed system; it’s systems managing systems, with automated PR reviews, AI-driven linting, and even self-healing deployment scripts.
  • Voice & call center automation: Modern real-time voice agents now run on hybrid ASR/NLP stacks powered by models like Whisper, Deepgram, or AssemblyAI. These systems detect intent, sentiment, and compliance in milliseconds, and many now integrate directly with ticketing and ERP systems. This allows for fully automated routing, booking, and billing resolution, driving 30–60% reductions in cost and a massive boost in consistency.
  • Finance automation: AI-enabled reconciliation engines and anomaly detection models are turning manual month-end close processes into continuous close operations. Systems ingest GL data in real time, align it against prior period models, and auto-escalate exceptions. What once required multi-week manual preparation is now accomplished in hours, with precision.

These examples underscore a simple truth: efficiency doesn’t always require full platform rewrites—it often involves augmenting existing systems with intelligence layers. Companies achieving the biggest gains are systematically inserting AI into workflow steps where context is well-structured and feedback cycles are measurable, allowing rapid iteration and ROI tracking.

Are Jobs Actually Going Away?

Some jobs are disappearing—but most aren’t. According to recent estimates from the Wharton School, only about 1% of jobs could be fully automated today—but more than 25% of U.S. roles include tasks where 90–99% of the work could be automated by AI. That’s not just a shift—it’s a restructuring of human work.

McKinsey estimates that up to 30% of current work hours could be automated by 2030 due to generative AI, with productivity boosts concentrated in knowledge work, customer service, finance, and coding-intensive roles, yet only 5% of occupations are expected to be fully replaced.

A recent BCG report suggests AI can support or automate as much as 80% of the tasks within corporate functions like operations and finance, with teams reclaiming between 26–36% of their time when used effectively.

Humans won’t disappear—but they will need to adapt. The Gallup Workplace Index notes that usage of AI tools at work has nearly doubled in the last two years, from 21% to 40%, signaling broad adoption—not displacement. The real shift is in how work is structured and what skills are needed.

Bubble or Not?

Are we in a hype cycle? Maybe.

Yes, It’s a Bubble: The exuberance has led to inflated expectations and a flood of pilots that never make it to production. As one analyst put it, “AI has become the new currency of hype—and hype has a cost.” For every success story, there are failed proofs of concept, governance blind spots, and companies rushing in without readiness.

No, This Is Real: If this is a bubble, it’s a bubble filled with fundamentals—real productivity gains, measurable cost reductions, and transformative potential across sectors. One CTO remarked, “The hype isn’t the point—the results are.” The companies winning right now aren’t just riding hype; they’re building lasting operational muscle.

But here’s what’s real: this efficiency wave isn’t optional. And this feels fundamentally different from the Dot-Com 1.0 boom I lived through as an engineer at Scient—a company that once embodied the innovation narrative and now exists only as a cautionary tale. Back then, we were building for a promise, often without the means to fulfill it. Today, companies are already seeing results before they’ve even finished the slide decks. This isn’t speculative. It’s happening in real time—and the companies who fail to adapt won’t just fade quietly—they’ll be overtaken.

“Once you’ve watched an AI model build a product roadmap in 5 minutes, you can’t go back to waiting 6 months for slide decks.”

Closing: What’s Next for Corporate Efficiency

Leaders must stop treating AI like an IT problem. It’s an operating system shift.

The future belongs to companies who act on inefficiency—not those who fear the discomfort of change. The winners will treat efficiency as a core value.

We’re not just implementing AI. We’re rewriting how companies work.

Welcome to the reckoning.

Recursion: We Have to Go Deeper

Opening Scene: Inception & the Dream Layers

You’re in a van. It’s falling. But somehow… you are not.

Gravity tilts sideways. Time stretches like taffy. And as the van free-falls, you suddenly find yourself in a hotel hallway — where everyone is also falling, but standing up straight, because physics took the night off.

Then — snap — you’re in a snowy fortress. Guns. Ski masks. Zero explanation.

You are multiple layers deep into the same mission. Each layer depends on the one above. Each layer runs slower than the one below.

This is recursion in IMAX — brought to you by director Christopher Nolan, the reigning champion of layered storytelling (and one of my all‑time favorite filmmakers).

A state inside a copy of that state. A system inside a smaller version of the same system. A repeated descent… until you find the exit condition.

Because without a way back — without a base case — you’re not a hacker or an architect or a programmer. You’re lost. In Limbo. Running a function that calls itself forever.

Movie poster for 'Inception', featuring a man in a suit standing in a flooded urban environment, with skyscrapers in the background. The title 'Inception' and credits are displayed prominently.

Dinner in Boston with My Daughter

Recently, I was in Boston having dinner with my daughter — a Computational Biology major (which means she studies things that evolve recursively in nature and then complains when recursion shows up in class).

She’s brilliant, curious, and probably smarter than me already. It’s fine. I’m fine. But… recursion? Recursion has been her intellectual villain — like a video game boss whose health respawns every time you think you’ve finally won.

I told her, “It’s just like Inception — a dream within a dream.” She laughed, rolled her eyes, and informed me that my analogy was both helpful and deeply unhelpful at the same time. But the truth is: recursion requires a different mental model — one where you trust the call stack and hope you remembered your base case.

Parenting is similar.

What Recursion Actually Is

At its heart, recursion is when a function solves a problem by solving a smaller version of itself. Like zooming into a photo… then zooming again… and again — each layer revealing the same picture but slightly smaller, until eventually you hit pixels and either succeed… or question all your life choices.

Every recursive solution needs two things:

  • Recursive case — keep going deeper
  • Base case — stop before everything breaks
function solve(problem):
    if problem is tiny enough:
        return solution  // base case
    else:
        return solve(smaller problem)

Elegant. Terrifying. Very Nolan.

Why Recursion Feels Hard (and How Iteration Tries to Save Us)

Humans love straight lines. Step 1 → Step 2 → Step 3 → victory royale.

Recursion hands you a spinning top and says: “Just solve the problem… by solving the problem… inside the problem.”

Your brain becomes a call stack:

  • “Which version of me is running?”
  • “Is this the dream or the dream inside the dream?”
  • “Wait… did I already solve this?”

One missing base case and — boom: Stack overflow. Your program panics harder than Ariadne when the Paris street folds in half.

This is where iteration strolls in like the chill coworker:

  • Recursion: Elegant. Reads like poetry. The Leonardo DiCaprio of code.
  • Iteration: Practical. Gets things done. Drives a minivan.

In fact, recursion and iteration can solve most of the same problems. But recursion does it with flair — and sometimes significantly worse performance if you’re not careful about stack depth and repeated work.

Example: Fibonacci.

# Recursive Fibonacci
function fib(n):
    if n <= 1:
        return n
    return fib(n-1) + fib(n-2)

Run with n = 40 and you may have time to:

  • Make coffee
  • Drink that coffee
  • Earn a PhD in Computational Biology
  • Explain recursion again over dinner

Why? Because it repeats the same work exponentially.

Now iteration:

# Iterative Fibonacci
function fib(n):
    a = 0
    b = 1
    for i in range(2, n+1):
        temp = a + b
        a = b
        b = temp
    return b

Same output. Runs faster than Dom Cobb when the van is five seconds from splashdown. No duplicate work. No stack explosions.

Performance efficiency matters — and recursion does not always care about your CPU budget.

Still… for problems that nest, branch, or fracture into self‑similar layers, recursion is the genius iteration calls when it’s out of ideas.

Where Recursion Secretly Rules Software

If you’ve made it this far, congratulations — you’ve followed the call stack down a few layers without losing track of who you are. My daughter would be proud.

And yes… we have to go a bit deeper.

Even if production engineers avoid it like unbounded memory leaks, recursion is the hidden backbone of:

  • File systems (folders inside folders inside folders)
  • UI frameworks like React (components inside components)
  • Sorting algorithms (QuickSort’s entire personality is “let’s descend”)
  • Tree and graph search (the internet is basically recursion with ads)

Recursion is the invisible stage crew of computing — unseen but essential.

Bonus nerd moment: Computational biology uses recursion constantly — phylogenetic trees, protein folding… nature recurses.

Recursion in Modern AI (Yes, Even This Post)

Recursion isn’t just a CS 101 rite of passage — it’s powering many of the smartest systems we build today:

  • Neural networks backpropagate through layers recursively.
  • Game‑playing AIs explore huge game trees with recursive search.
  • Recursive neural networks understand language hierarchies.
  • Graph algorithms break complex problems into smaller ones… repeatedly.

Even large language models — hello 👋 — reason over nested patterns that behave recursively, even if there’s no literal call().

Basically, recursion graduated, joined a research lab, and now publishes papers. Iteration still fixes production bugs.

Real Talk: Why We Don’t Use It Everywhere

Production systems hate drama. Recursion can bring:

  • Memory overhead (every function call needs space)
  • Unpredictable depth (sometimes life decides to go four dreams down)
  • Debugging sessions that feel like you’re chasing your own tail
  • That “oh no” moment when your stack detonates in prod

Bring a totem: unit tests, depth limits, and logs — so you always know which layer you’re in. Loops are the steady minivan — recursion is the flashy sports car you only drive when the weather is perfect.

Recursion in Pop Culture & Reality

If recursion feels abstract or like something only math people obsess over… it already found you.

Recursion is everywhere:

  • Mirrors facing mirrors 🎭
  • A photo of someone holding a photo of themselves 📸
  • Russian nesting dolls 🪆
  • Your company org chart (boss → boss → board → shareholders → chaos)

Parenting might actually be recursion: Teach kid → kid grows up → teaches their kid → reboot universe.

The Recursive Ending

Back in Boston, my daughter asked: “So recursion is just… doing the same thing, but smaller, until you can stop?”

I nodded. She spun an imaginary top and said, “Wake me when we hit the base case.”

That’s recursion: Define the base case. Trust the descent. Don’t lose your totem.

If this keeps going, we’re in Limbo.

return "We have to go deeper."

The World Models Within Us

From Asimov to AGI: The Rise of Predictive Minds

“A robot may not injure a human being, or, through inaction, allow a human being to come to harm.” — Isaac Asimov

Isaac Asimov, one of my favorite authors, is best known for his Robot series — books that deeply influenced how I think about technology and ethics. The series wasn’t just about robots; it was about the emergence of intelligence — about what happens when a machine begins not just to obey, but to understand. His characters like Dr. Susan Calvin grappled with robots who broke the Three Laws not out of rebellion, but out of deeper reasoning — they’d built internal models of the world complex enough to predict human consequences.

That, in essence, is the story of world models in modern AI. Machines are learning not just to process data, but to imagine futures.

The Foundation of World Models

While most consumer-facing AI tools we see in the news or use at home or work — like ChatGPT, image generators, or voice assistants — excel at pattern recognition, true world models go a step further. They don’t just respond to input; they build internal representations of how the world works, allowing them to reason about context, dynamics, and cause-effect relationships. This makes them capable of simulating and anticipating outcomes rather than simply reproducing patterns from past data. A world model is a latent representation of reality that allows an intelligent system to simulate, plan, and predict. In technical terms, it’s what allows agents like MuZero or GPT-based systems to operate beyond mere pattern recognition. Instead of memorizing, they model — forming compressed internal maps of how the world behaves.

If a neural network is like a camera capturing pixels, a world model is like a mapmaker building an atlas. It abstracts, generalizes, and anticipates — and crucially, it allows for planning. A good world model can take incomplete information and still simulate likely futures. It’s the engine behind self-driving cars navigating uncertainty, generative agents predicting user intent, and language models inferring context.

This idea originated in reinforcement learning research (Ha & Schmidhuber’s World Models, DeepMind’s MuZero, OpenAI’s Sora). But the same principle now underpins multimodal systems like GeminiGato, and GPT-5, which combine perception, reasoning, and action under one unified architecture. These systems don’t just respond — they simulate, project, and plan.

From Logic to Learning: How We Got Here

The lineage of AI traces a fascinating arc. It began with symbolic AI, when we tried to encode knowledge into explicit rules — much like Asimov’s Three Laws. Those systems could reason, but only within the narrow confines of their logic. Then came the revolution of statistical learning, where we replaced rules with probabilities and let data teach the machine. That gave us vision systems, speech recognition, and language models — but not true understanding.

Now we’re entering the age of world modeling — a synthesis between symbolic precision and statistical intuition. These new systems learn not only to recognize patterns, but to simulate how the world changes over time. It’s a step closer to how humans think. When you reach for a coffee mug, you don’t calculate every possible trajectory — your brain’s world model has already predicted the motion.

World models are the missing bridge between perception and agency. They turn observation into imagination and allow intelligent agents to act not just reactively, but intentionally.

Why World Models Matter

Every intelligent system — human or machine — must answer the same question: What will happen if I act? That is the essence of intelligence — the ability to simulate the future and choose accordingly.

In AI, this predictive capacity is fundamental:

  • In AGI research, world models enable self-supervised learning, planning, and reasoning across tasks.
  • In enterprise applications, they power agents that can simulate business outcomes before executing decisions.
  • In human cognition, predictive coding lets our brains anticipate the world milliseconds before our senses confirm it.

World models, then, are not just a technical concept. They’re a philosophy of intelligence — a recognition that foresight, not memory, is the truest marker of understanding. They allow systems to bridge the gap between data and decision, between reaction and anticipation.

The Business Parallel

In business and technology alike, we build world models to make sense of complexity. Whether we call them strategy frameworks, simulation tools, or digital twins, these systems help us forecast outcomes, reduce uncertainty, and align execution. A well-constructed model of operations or markets acts as a simplified reflection of reality, letting leaders test scenarios before taking action.

Organizations that refine their world models — capturing customer behavior, market dynamics, and operational constraints — become more adaptive. Pricing, scheduling, and resource allocation evolve from guesswork into predictive, data-informed decisions. The fidelity of a company’s model of its world determines how effectively it can act within it.

Every system — human, corporate, or artificial — competes on the fidelity of its world model. The sharper the map, the smarter the motion.

The Rise of Embodied Intelligence

Recent advancements in robotics have brought world models from simulation to the physical world. Robots are no longer confined to factory floors or research labs—they’re navigating complex, unstructured environments using predictive world modeling to anticipate, adapt, and learn.

Take Tesla’s OptimusBoston Dynamics’ Atlas, or Figure AI’s humanoid robots—each of these systems relies on internal models to understand and predict physical interactions. They don’t just follow pre-programmed paths; they build internal maps of how objects, gravity, and force behave. When a robot walks across uneven terrain or picks up a delicate object, it’s using its world model to simulate potential outcomes before acting.

Even in household robotics, we’re seeing the impact. Modern robot vacuums and warehouse automation bots use spatial and behavioral world models to improve navigation and decision-making over time.

This convergence of embodied AI and world modeling is what gives machines agency in the real world. The same predictive reasoning that allows a language model to plan a paragraph now lets a robot plan a movement. As the fidelity of these physical world models increases, robots will move from reactive tools to proactive collaborators—learning continuously through experience.

The Limits and the Next Frontier

All models are simplifications, and even the most advanced world model is still a shadow of reality. The world shifts faster than any simulation can adapt. World models hallucinate, overfit, and misinterpret context — much like humans with bias and intuition. But even imperfect models can be powerful tools for alignment: between perception and truth, intent and impact.

Researchers like LeCun, Schmidhuber, and the teams at DeepMind are exploring architectures (JEPA, Dreamer, Genie) that enable machines to learn continuously from their environment, refining their world models the way children do — through play, prediction, and correction. It’s a vision of AI that doesn’t just compute but grows.

The next frontier is not just bigger models, but better models of the world — ones that learn cause and effect, adapt to change, and stay grounded in reality.

Asimov’s Mirror

Asimov’s fiction gave us robots wrestling with morality; our reality gives us agents wrestling with understanding. His positronic brains modeled the world to serve human ends. Ours must now do the same — but with humility.

Asimov’s Robot stories ultimately asked whether understanding the world made robots more human — or humans more predictable. That same question echoes today. The more our machines model us, the more we must model ourselves.

“His robots learned to imagine the consequences of their actions. We’re finally teaching ours to do the same.”

Every great civilization — and every great AI — begins with a world model.

Transform or Be Transformed: The AI-Native Enterprise and the Law of Acceleration (Part II)

If the last decade was defined by the rise of the software-defined enterprise, this next one will be shaped by the emergence of the AI-defined organization. The difference isn’t just speed — it’s sentience. Systems are no longer passive enablers of process; they are active participants in learning, optimization, and decision-making.

The collapse of traditional procurement and platform cycles naturally leads here: the rise of the AI-native enterprise. As businesses shed their dependence on slow-moving vendors and rigid systems, they’re embracing architectures that evolve as fast as their ideas. The boundary between buying and building is dissolving; adaptation itself becomes the product.

This shift redefines enterprise architecture and culture alike. In software-defined organizations, technology followed business logic; in AI-defined ones, business logic emerges from data. Strategy becomes dynamic, culture becomes computational, and architecture becomes an extension of organizational intelligence. Success will depend on how well teams can blend human creativity with machine reasoning — designing systems that not only serve the business but learn from it.

A glimpse of this future can already be seen in companies like Netflix, Tesla, and Amazon. Netflix’s recommendation engine doesn’t just reflect viewer preferences; it continuously redefines content strategy. Tesla’s vehicles act as data nodes in a distributed learning system, refining autonomy with every mile. Amazon’s logistics and product algorithms anticipate demand before it’s expressed. Each of these models shows what happens when architecture, data, and intelligence converge — when organizations operate less like hierarchies and more like adaptive, self-learning organisms.

But autonomy comes with new risks. As systems learn and act faster than human governance cycles, the potential for misalignment grows. Ethical design, transparency, and bias mitigation become central disciplines, not afterthoughts. The same feedback loops that make these systems powerful can also amplify unintended consequences. The AI-native enterprise must balance acceleration with accountability — building not just intelligent systems, but responsible ones. Ultimately, that balance begins with leadership: those who understand that technical innovation without ethical stewardship isn’t transformation — it’s turbulence.

Tomorrow’s enterprise won’t be defined by org charts. It will be defined by architecture.

  • Systems built on unified data layers.
  • Business logic expressed through AI agents and reusable microservices.
  • Product roadmaps driven by simulation, not gut instinct.
  • Creativity amplified by intelligent copilots.

The companies that thrive won’t distinguish between IT, Product, and Data. They’ll operate as one neural system — adaptive, aware, and accelerating.

This is the future of enterprise architecture: from monoliths to microservices, from APIs to embeddings, from codebases to knowledge graphs.


The New Law of Acceleration

The world doesn’t change one company at a time anymore. It changes all at once.

As I wrote in my earlier blog post, Reflections on AI: The Law of Accelerating Returns, technology doesn’t just move faster—it compounds. Each generation of tools accelerates the creation of the next. The result is an ecosystem that evolves too quickly for traditional business rhythms to keep up. That same principle underpins this entire essay — Transform or Be Transformed is the organizational expression of the law of accelerating returns. If technology evolves exponentially, so must leadership, culture, and execution. The winners will be those who learn to compound adaptation as fast as the machines compound capability.

The speed of transformation is no longer linear — it’s exponential. The half-life of strategy has collapsed. Your roadmap is only as relevant as your next model release.

The leaders who survive will be the ones who integrate faster than their org charts can ossify.

AI isn’t coming for your job. It’s coming for your silos.

Transform or be transformed.

Because in the age of accelerating returns, transformation isn’t a project — it’s a permanent state of being.