ChatGPT Health Identified Respiratory Failure. Then It Said Wait.

Published: 19 Mar 2026 · 01:00 AM AEDT

Abstract

What's really happening inside AI agents when they give you the wrong answer? The common story is that smarter models mean safer agents — but the reality is that reasoning traces and final outputs often run as two separate processes.

Highlights

  • In this video, I share the inside scoop on why AI agents fail in production and how to build evals that actually catch it.
  • Why agents perform worst precisely where the stakes are highest.
  • How reasoning traces routinely contradict an agent's final recommendation.
  • What factorial stress testing reveals that standard benchmarks completely miss.
  • Where to build the four-layer architecture that keeps agents honest in production.
  • Operators who ignore this now will face it later — through customer harm, regulatory pressure, or an insurance policy they can't obtain.

References & Links

Anthropic Didn't Build a New Browser. They Did Something Smarter.

Published: 18 Mar 2026 · 01:00 AM AEDT

Abstract

Claude's Chrome extension isn't a chatbot sitting in the sidebar—it's a workflow recorder that can run entire browser routines on autopilot. Nate shows how scheduling those recordings turns customer-service fights, inbox triage, and multitab research into repeatable background jobs.

Highlights

  • Let Claude fight customer-service battles and negotiate credits while you stay out of the queue.
  • Record a multi-step browser workflow once, schedule it, and keep it running without touching the keyboard.
  • Gmail, Calendar, and Drive awareness means Claude can triage your inbox and surface the items that actually need you.
  • Group tabs plus structured exports turn scattered research into spreadsheets, briefs, or reports automatically.
  • Debugging tips, data limits, and security reminders so the extension doesn't outrun your governance.

References & Links

Claude Code Wiped 2.5 Years of Data. The Engineer Who Built It Couldn't Stop It.

Published: 17 Mar 2026 · 01:00 AM AEDT

Abstract

Vibe coding got a lot of founders to MVP, but agents behave like unsupervised contractors once they touch production work. Nate walks through the five management skills that keep Claude Code from undoing months of shipping.

Highlights

  • Version control and save points are survival skills—without them you can't unwind a bad agent run.
  • Know when to restart an agent (and when to rebuild its context) before the window collapses your instructions.
  • Standing orders, guardrails, and rules files beat heroic prompting when an agent wakes up mid-task.
  • Make small, sandboxed bets so a runaway edit can't torch the entire product.
  • Treat agents like powerful interns: scoped tasks, persistent briefs, and human review keep them from wiping production again.

References & Links

She quit, picked up AI, and shipped in 30 days what her team planned for Q3.

Published: 16 Mar 2026 · 05:00 AM AEDT

Abstract

Solo founders aren't mythical outliers—they're just operating without the meetings, approvals, and coordination drag that suffocates the same people inside larger teams. Nate explains how AI agents cut overhead so top talent can finally ship.

Highlights

  • AI agents delete coordination overhead; they don't just replace headcount, they give builders back their day.
  • Taste without conviction keeps high performers stuck—AI accelerates those who are willing to act on their instincts.
  • "Speed of control" beats "span of control": the faster leaders remove blockers, the faster AI compounding shows up.
  • When companies refuse to clear the path, ambitious people leave to solo-found simply because it's the only place they can move.
  • Recognize, protect, and unburden the 25% of talent that's already operating at 4× output with AI.

References & Links

AI Made Every Company 10x More Productive. The Ones Cutting Headcount Are Telling on Themselves.

Published: 15 Mar 2026 · 02:01 AM AEDT

Abstract

What's really happening when Whoop announces it's hiring 600 people while the media narrative focuses entirely on job displacement? The common story is about how many fewer people companies need—but the reality is more interesting when execution costs drop by an order of magnitude and the pie itself expands.

Highlights

  • In this video, I share the inside scoop on six unlocks that give you a picture of what the future actually looks like:
    • Why iteration cycles compressing from months to days changes the mechanics of strategy
    • How hundreds of millions of domain experts become builders when the translation layer disappears
    • What happens when quality software becomes the default, not a premium
    • Where the market for ambition explodes when CFO math flips on experiments For anyone wrestling with the people challenges of AI, the hardest work ahead isn't technical—it's figuring out what upskilling looks like when the job isn't do the same thing faster.

References & Links

One Simple System Gave All My AI Tools a Memory. Here's How.

Published: 14 Mar 2026 · 01:01 AM AEDT

Abstract

What's really happening when thousands of people build an agent-readable database but can only interact with it through a chat window keyhole? The common story is that the MCP server is the whole system—but the reality is more interesting when you add a human door alongside the agent door.

Highlights

  • In this video, I share the inside scoop on how to give your Open Brain hands and feet through visual interfaces you build and deploy for free:
    • Why the table becomes a shared surface that both you and your agent see
    • How to build a visual layer with Claude and host it on Vercel for nothing
    • What household knowledge, professional relationships, and job hunts look like as dashboards
    • Where time bridging and cross-category reasoning earn their keep

References & Links

4,000 People Lost Their Jobs At Block. Dorsey Blamed AI. Here's What Actually Happened.

Published: 13 Mar 2026 · 01:01 AM AEDT

Abstract

What's really happening when the average knowledge worker spends 60% of their time on meetings and documents that exist only to coordinate with other humans? The common story is that AI automates tasks within your existing org—but the reality is more interesting when the coordination layer evaporates entirely.

Highlights

  • In this video, I share the inside scoop on why AI is revealing the job was never the real job:
    • Why PRDs, sprint planning, and status updates exist because the execution layer is human
    • How agent harnesses delete the need for handoffs, not just automate the handoffs themselves
    • What survives when coordination roles disappear: vision, architecture, genuine care, systems design
    • Where the two qualities that matter most are agency and ramp

References & Links

4 AI Labs Built the Same System Without Talking to Each Other (And Nobody's Discussing Why)

Published: 12 Mar 2026 · 01:00 AM AEDT

Abstract

What's really happening with AI capabilities at work — and why the "jagged AI" frame is now obsolete? The common story is that AI is brilliant at some things and broken at others — but the reality is that jaggedness was never about intelligence; it was about how we were deploying it.

Highlights

  • In this video, I share the inside scoop on why AI agents in proper harnesses are smoothing the capability frontier for real work:
    • Why the jagged AI frontier was always a deployment problem
    • How multi-agent coordination unlocks long-horizon knowledge work
    • What Cursor's math breakthrough reveals about AI generalization
    • Where meta-skills like sniff-checking become your competitive edge The organizations and individuals who learn to decompose work, delegate to AI agents, and verify outputs will extend their leverage — those who don't will find the shift happening to them anyway.

References & Links

Stop accepting AI output that "looks right." The other 17% is everything and nobody is ready for it.

Published: 11 Mar 2026 · 01:00 AM AEDT

Abstract

What's really happening when frontier models beat professionals with 14 years of experience 70% of the time but the output still doesn't survive contact with anyone who actually understands the domain? The common story is about prompting and workflow design—but the reality is more interesting when rejection creates institutional knowledge that did not exist before.

Highlights

  • In this video, I share the inside scoop on why learning to say no is the missing skill in the judgment and taste category:
    • Why your rejections are more valuable than your prompts
    • How recognition, articulation, and encoding break down into learnable dimensions
    • What Epic Systems teaches about scaling taste through thousands of encoded workflows
    • Where the structural gap in the AI tool ecosystem leaves every rejection on the floor For anyone watching AI flood organizations with output, the frontier of AI value is identical to the frontier of your organization's taste.

References & Links

Claude Blackmailed Its Developers. Here's Why the System Hasn't Collapsed Yet.

Published: 10 Mar 2026 · 01:01 AM AEDT

Abstract

What's really happening with AI safety in 2026? The common story is that the safety system is collapsing — but the reality is more complicated.

Highlights

  • In this video, I share the inside scoop on why the AI risk picture is both worse and more resilient than the headlines suggest: Why frontier AI agents scheme even after anti-scheming training
    • How competitive dynamics create emergent safety properties no lab planned
    • What "intent engineering" is and why it beats prompt engineering for AI agents
    • Where the real vulnerability lives — and why it's you, not the models The risks from large language models and autonomous AI agents are accelerating, but so are the structural forces holding the system together — and closing the gap between what you tell an agent and what you actually mean is the most leveraged safety skill you can build right now.

References & Links

45 People, $200M Revenue. The Question Nobody's Asking About AI and Your Team Size.

Published: 09 Mar 2026 · 05:00 AM AEDT

Abstract

What's really happening with AI and team size in your organization? The common story is that AI makes teams more productive so you can cut headcount — but the reality is more complicated.

Highlights

  • In this video, I share the inside scoop on why the five-person strike team is the structural unit of the AI era:
    • Why AI raised coordination costs by the same order as output
    • How scouts and strike teams map to different AI-era missions
    • What correctness-first thinking means for how you hire and build
    • Where the real opportunity is — expanding ambition, not shrinking headcount AI agents and LLMs didn't break your meetings problem — they amplified a team size problem you already had, and the leaders who restructure around small, high-judgment teams will build the defining companies of this decade.

References & Links

GPT-5.4 Let Mickey Mouse Into a Production Database. Nobody Noticed. (What This Means For Your Work)

Published: 08 Mar 2026 · 03:00 AM AEDT

Abstract

What's really happening when OpenAI engineers accidentally leak ChatGPT 5.4's existence but the model isn't even the interesting part? The common story is about the next capability jump—but the reality is more interesting when the company that first makes trillion-token organizational context genuinely usable becomes the new enterprise data platform.

Highlights

  • In this video, I share the inside scoop on why the four-part compound bet determines whether this justifies an $840 billion valuation:
    • Why intelligence and context are multiplicative—and weak reasoning with long context is actively harmful
    • How retrieval at enterprise scale breaks RAG in ways nobody's benchmarking
    • What memory that doesn't rot requires when organizational knowledge continuously evolves
    • Where Anthropic's organic context accumulation through Claude Code might beat OpenAI's infrastructure play For builders watching the enterprise stack get restructured, the lock-in from synthesized understanding is deeper than anything enterprise software has ever seen.

References & Links

Claude Code vs Codex: The Decision That Compounds Every Week You Delay That Nobody Is Talking About

Published: 07 Mar 2026 · 02:00 AM AEDT

Abstract

What's really happening inside AI coding tools that nobody's comparing? The common story is that Claude vs.

Highlights

  • ChatGPT is a model competition — but the reality is that the model is the least important part.
  • In this video, I share the inside scoop on why the AI harness matters more than the model: - Why the same Claude model scored 78% vs.
  • 42% on identical benchmarks
    • How Claude Code and Codex embody opposite philosophies of AI
    • collaboration
    • What harness lock-in actually costs teams who switch tools later
    • Where non-technical leaders are making the wrong procurement decisions The teams getting this right aren't choosing the smartest AI agent — they're choosing the architecture that matches how they work, and that decision compounds every quarter.

References & Links

OpenAI Leaked GPT-5.4. It's a Distraction. (The AI Lock-In No One Is Talking About)

Published: 06 Mar 2026 · 02:00 AM AEDT

Abstract

What's really happening when OpenAI engineers accidentally leak ChatGPT 5.4's existence but the model isn't even the interesting part? The common story is about the next capability jump—but the reality is more interesting when the company that first makes trillion-token organizational context genuinely usable becomes the new enterprise data platform.

Highlights

  • In this video, I share the inside scoop on why the four-part compound bet determines whether this justifies an $840 billion valuation:
    • Why intelligence and context are multiplicative—and weak reasoning with long context is actively harmful
    • How retrieval at enterprise scale breaks RAG in ways nobody's benchmarking
    • What memory that doesn't rot requires when organizational knowledge continuously evolves
    • Where Anthropic's organic context accumulation through Claude Code might beat OpenAI's infrastructure play For builders watching the enterprise stack get restructured, the lock-in from synthesized understanding is deeper than anything enterprise software has ever seen.

References & Links