All Episodes
19 episodes and counting. New episodes daily.
78 Bills, 27 States, and the White House Wants Them All Gone
Twenty-seven states have introduced 78 chatbot safety bills in just two months. Oregon passed the first one this week. Florida's Senate voted 35-2 for an AI Bill of Rights — only for the House to kill it under White House pressure. And on March 11, two federal deadlines could trigger the first lawsuits against state AI laws. In this episode: The wave of state AI legislation sweeping America in 2026 and why it's overwhelmingly bipartisan How Trump's December executive order created an AI Litigation Task Force to challenge state laws — and conditioned $42 billion in broadband funding on compliance The child safety carve-out that could be a lifeline for most state chatbot bills Colorado's AI Act as the likely first target for federal legal challenge The $125 million AI industry spending war between pro- and anti-regulation super PACs ahead of the midterms
Safety Guardrails vs. Government Access: Anthropic's Impossible Choice
The Pentagon blacklisted Anthropic as a national security risk. Hours later, the military used Claude to target strikes in Iran. A leaked internal memo calling OpenAI's deal "safety theater" made everything worse. Full breakdown in today's episode.
AIUC-one: The SOC Two for AI Agents
Enterprises are handing AI agents access to their most sensitive systems, but until now, there was no standardized way to verify those agents are safe. AIUC-one changes that. In this episode: What AIUC-one is and how it works as the SOC 2 equivalent for AI agents The six domains it covers, from prompt injection defense to hallucination detection Why JPMorgan, Anthropic, Google, Cisco, MITRE, and Stanford are all behind it How the Q1 2026 update introduced capability-based scoping and new evidence categories What this means for enterprise procurement, security teams, and AI builders The big takeaway: AIUC-one solves the trust gap holding back enterprise AI adoption, and the companies that get certified first will have a real competitive edge. New episodes every weekday. Share this with your security or procurement team.
Grok 4.20 multi-agent inference works at production scale
xAI just shipped something fundamentally different. Grok 4.20 doesn't use one model to answer your questions. It deploys four specialized AI agents that think in parallel, debate each other in real time, and synthesize a unified answer before you see a single word. In this episode: How the four-agent architecture works: Grok (Captain), Harper (researcher), Benjamin (logician), and Lucas (contrarian) The hallucination results: a sixty-five percent reduction, from twelve percent down to four point two percent Alpha Arena and ForecastBench: where Grok 4.20 outperformed GPT-5 and Gemini The real criticisms: latency, new failure modes, and the social media fact-checking problem Why this might reshape how every lab builds AI over the next year The big takeaway: whether Grok 4.20 wins the model race or not, xAI just proved that teams of models can outperform individual geniuses at production scale. That changes the game. New episodes every weekday. Share this with someone keeping up with AI.
Lockdown Mode: When AI Security Means Disabling AI Features
Microsoft just discovered that thirty-one companies are hiding prompt injections inside ordinary "Summarize with AI" buttons, poisoning your AI assistant's memory to manipulate future recommendations. The tools to do this are open source, documented, and work across ChatGPT, Copilot, Claude, Perplexity, and Grok. In this episode: How AI Recommendation Poisoning works and why Microsoft compares it to the SEO wars Why prompt injection is the number one AI security threat and structurally unfixable in current architectures The EchoLeak zero-click attack, three hundred thousand stolen ChatGPT credentials, and the massive readiness gap in agentic AI deployment OpenAI's new Lockdown Mode: what it disables, why that matters, and the security-versus-capability tradeoff every organization now faces The big takeaway: defending AI systems is going to be a long, iterative war, and the choices organizations make right now about security versus capability will define the next era of AI deployment. New episodes every weekday. Share this with your security team.
Cursor Gave AI Agents Their Own Computers
Cursor just announced cloud agents that change the game for AI-assisted coding. These agents don't just write code in your editor — they spin up their own virtual machines, build and test the software, and deliver merge-ready pull requests with video recordings of themselves using the finished product.In this episode:- How Cursor's cloud agents work: isolated VMs, parallel execution, and self-validating output- The AI coding tool war by the numbers: Cursor at twenty-nine billion valuation versus Claude Code, Codex, and Copilot- Why this signals the shift from AI assistance to AI autonomy in software development- The uncomfortable question: if agents write, test, and demo the code, what's the developer's role?The big takeaway: the AI coding market is moving from autocomplete to autonomous agent fleets, and every developer tool will need to match this model within months.New episodes every weekday. Share this with a developer keeping up with AI tools.
The Swarm, The Solver, and The Coder
Three Chinese AI labs just released models that are rewriting the leaderboards. Moonshot AI's Kimi K2.5 can spin up a hundred agents working in parallel and scored 74.9% on BrowseComp, seventeen points ahead of GPT-5.2. Alibaba's Qwen3-Max-Thinking hit 58.3 on Humanity's Last Exam with perfect scores on AIME 2025. And Zhipu AI's GLM-5 matches Claude Opus 4.6 on SWE-bench Verified at a fraction of the cost. All three are open source. We break down what each one does, why it matters, and what it means for developers and builders. Sources: Moonshot AI (kimi.com), Alibaba Qwen (huggingface.co/Qwen), Zhipu AI (zhipuai.cn), TechCrunch, InfoQ, RAND Corporation.
Inside the AI Microscope — How Researchers Are Finally Learning Why AI Lies and Cheats
For the first time, researchers can peer inside AI models and see not just what they say, but what they're actually thinking. It's called mechanistic interpretability, and MIT Technology Review just named it one of the ten breakthrough technologies of twenty twenty-six. In this episode: how Anthropic built an AI microscope using sparse autoencoders, what they found inside Claude — including features tied to deception, sycophancy, and a collection of absorbed internet personas — and how OpenAI used related techniques to catch one of its own reasoning models cheating on coding tests, in its own words, in real time. Plus: the race to scale this research before AI models outpace our ability to understand them, and the growing divide between Anthropic's ambitious twenty twenty-seven interpretability goals and Google DeepMind's more pragmatic approach.
The Three Sixty Billion Dollar AI Summit
India just hosted the largest AI investment event in history. Here's what was pledged, who showed up, and whether this actually helps the people it's supposed to.
OpenAI's hire of OpenClaw creator Peter Steinberger
OpenClaw went from one-hour side project to nearly two hundred thousand GitHub stars in ninety days. Then OpenAI hired its creator. The story behind how a trademark dispute may have handed OpenAI their most important agent hire of the year. New episode out now.
Seedance 2.0: Hollywood's Worst Nightmare Is Here
ByteDance's new AI video model went viral in 72 hours, triggered cease-and-desist letters from Disney and Paramount, and may have just changed the creative economy forever.
Sixteen Agents, One Compiler, Two Weeks Anthropic's new Agent Teams
Sixteen AI agents built a C compiler in two weeks for twenty thousand dollars. Anthropic's new Agent Teams feature lets Claude agents coordinate like an actual engineering team. We go deep on what this means for the future of coding.
Shadow AI: The Governance Gap Nobody's Talking About
Two hundred and twenty-three AI security incidents per month. That's the average enterprise. And forty-nine percent of employees are using AI tools without approval. Shadow AI is the governance gap nobody is talking about. Full deep dive in today's episode of AI Signals.
MCP Goes to the Linux Foundation
Anthropic donated MCP to the Linux Foundation, and OpenAI, Google, and Microsoft all signed on as backers. Ninety-seven million downloads a month. But security researchers are raising red flags. Full deep dive in today's episode of AI Signals.
Perplexity's Model Council: Why One AI Isn't Enough Anymore
Perplexity just shipped Model Council. It runs Claude, GPT, and Gemini on your question at once, then shows you where they agree and where they don't. This might be the beginning of the end for single-model answers. Full breakdown in today's episode.
Anthropic Promised No Ads. But There's a Hidden Escape Hatch
Anthropic aired a Super Bowl commercial pledging Claude will never have ads. Then we read the fine print. There's a caveat that changes everything about this promise.
The Half-Trillion Dollar Bet
Four companies. $650 billion. One year. Amazon, Google, Microsoft, and Meta are making the largest corporate investment bet in history on AI infrastructure. Google's CEO says even $185B won't be enough. But 95% of enterprises see zero AI returns. New episode breaks it all down.
AI Just Drove a Rover on Mars. Here's How.
NASA used Anthropic's Claude to plan a drive on Mars. Nearly 1,500 feet of Martian terrain, waypoints written in Rover Markup Language, 500,000+ variables checked. The first AI-planned drive on another planet. And it worked. Full story in today's episode.
Software Stocks - Bloodbath
Anthropic and OpenAI go head-to-head with back-to-back launches of Claude Opus 4.6 and GPT-5.3 Codex, ushering in the era of AI agent teams and self-building models. Meanwhile, the fallout hits Wall Street hard — Claude Cowork's new industry plugins trigger a $285 billion selloff across software, legal tech, and financial data stocks, with Thomson Reuters posting its worst day on record. Apple quietly strikes a billion-dollar deal to bring Google's Gemini into Siri, raising privacy questions. UK regulators launch formal investigations into xAI's Grok over deepfake failures that competitors easily avoid. And in the funding arena, over $60 billion pours into AI in January alone as Chinese labs like DeepSeek and Alibaba close the gap with Western frontier models. This was one of the biggest weeks in AI history — here's what it all means.