PodcastEconomiaChain of Thought | AI Agents, Infrastructure & Engineering

Chain of Thought | AI Agents, Infrastructure & Engineering

Conor Bronsdon
Chain of Thought | AI Agents, Infrastructure & Engineering
Ultimo episodio

54 episodi

  • Chain of Thought | AI Agents, Infrastructure & Engineering

    Context Poisoning Is Killing Your AI Agents: How to Stop It

    25/03/2026 | 44 min
    Michel Tricot co-founded Airbyte, the open source data integration platform with 600+ free connectors that hit a $1.5 billion valuation. Now he's building the company's next product: an agent engine, currently in public beta. His thesis is that agents don't fail because models are bad. They fail because the data feeding them is wrong: context poisoning is killing them.
    Michel demos this live. A simple Gong query through raw API calls burned 30,000 extra tokens and took three minutes. The same query through Airbyte's context store ran in one minute and used a fraction of the context window. Conor and Michel dig into why RAG alone won't cut it, what a "context engineer" actually does, how Airbyte tracks entities across Salesforce, Zendesk, and Gong without embeddings, and whether the SaaS apocalypse playing out in public markets is overblown.
    Chapters:
    0:00 Intro
    0:20 Meet Michel Tricot, CEO of Airbyte
    2:27 Data Got Us to the Information Age. Context Gets Us to Intelligence.
    4:48 How Context Poisoning Breaks Agents
    7:49 Why Airbyte Customers Stopped Loading Into Warehouses
    10:12 Live Demo: Context Store vs Raw API Calls
    10:38 What Does a Context Engineer Actually Do?
    14:14 RAG Isn't Dead, But How We Build It Will Die
    16:41 30K Wasted Tokens Without Proper Context
    22:22 Cross-System Joins: Zendesk, Gong, and Salesforce
    26:12 The Open Source Agent Connector SDK
    29:45 The SaaS Apocalypse Is Overblown
    36:09 From Data Pipes to Agent Infrastructure
    38:51 What Agents Need to Get Right by Summer
    40:48 Memory Is Just Another Form of Context
    43:07 Outro
    About the Guest:
    Michel Tricot is the CEO and co-founder of Airbyte, the open source data integration platform used by thousands of companies to move data between systems. Before Airbyte, he led data ingestion and distribution engineering at LiveRamp. Airbyte raised at a $1.5 billion valuation and offers 600+ free connectors. The company recently launched the public beta of its agent engine, which includes a context store, agent connector SDK, and MCP integration.
    Guest Links:
    Airbyte
    Michel on LinkedIn
    Agent Blueprint (Substack)
    Agent Connector SDK (GitHub)
    Show Links:
    Chain of Thought Podcast
    Newsletter
    Conor on LinkedIn
    Conor on X/Twitter
    Thanks to our presenting sponsor Galileo. Download their free 165-page guide to mastering multi-agent systems at galileo.ai.
  • Chain of Thought | AI Agents, Infrastructure & Engineering

    I Started r/AI_Agents and Now I'm Launching a VC Fund

    10/03/2026 | 44 min
    Yujian Tang started the r/AI_Agents subreddit in April 2023. For the first year, it barely moved. Then it hit 9,000 members, he went on vacation, came back to 36,000, and now it's approaching 300,000. In this episode, Yujian talks about how that community grew alongside his event business (Seattle Startup Summit, 900+ attendees last year), his two failed startups, and why he just filed paperwork to launch his own venture fund.

    Conor and Yujian dig into the mechanics of starting a fund from scratch (Delaware PO boxes, EIN numbers, lawyers), why AI startup valuations have doubled in the last two years, whether a one-person unicorn is realistic, and what failed founders learn that successful ones sometimes miss.

    Chapters:
    (0:00) Cold Open: The Subreddit Growth Explosion
    (0:21) Intro and Meet Yujian Tang
    (1:06) From AI Research to Community Building
    (7:26) Where AI Applications Are Headed
    (10:03) The AI Bubble and a Valuation Reset
    (10:39) Getting Deal Flow Through Community Events
    (14:02) Filing the Fund: The Boring Side of VC
    (16:04) How r/AI_Agents Went from Crickets to 300K
    (18:39) Building an Accidental Empire
    (26:37) What Two Failed Startups Taught Him
    (29:52) Why Pre-Seed Valuations Are Out of Control
    (37:37) The One-Person Unicorn Debate
    (39:50) Seattle Startup Summit 2026
    (42:17) What Chain of Thought Should Cover Next
    (43:25) Outro

    About the Guest:
    Yujian Tang is the founder of Seattle Startup Summit, the largest startup event in the Pacific Northwest. He created the r/AI_Agents subreddit (now nearly 300K members), runs hackathons and developer events across Seattle and the Bay Area, and is launching an early-stage AI venture fund.

    Guest Links:
    Seattle Startup Summit: seattlestartupsummit.com
    Reddit: reddit.com/r/AI_Agents

    Show Links:
    Chain of Thought Podcast: https://chainofthought.show
    Newsletter: https://newsletter.chainofthought.show/LinkedIn: https://www.linkedin.com/in/conorbronsdon/X/Twitter: https://x.com/ConorBronsdon

    Sponsor: Thanks to Galileo. Download their free 165-page guide to mastering multi-agent systems at galileo.ai/mastering-multi-agent-systems
  • Chain of Thought | AI Agents, Infrastructure & Engineering

    I Built an AI Coworker That Runs 90% of My Day

    04/03/2026 | 1 h 2 min
    Sterling Chin stopped thinking of AI as a tool and started treating it like a junior employee. Onboarded it with context, corrected its mistakes, and gave it writing rules.
    Forty days later, MARVIN was handling 90% of his workday.
    In this episode of Chain of Thought, Sterling (Applied AI Engineer and Senior Developer Advocate at Postman) walks through live demos of MARVIN, his personal AI assistant built on Claude Code. From pulling meeting transcripts and updating Jira tickets to drafting blog posts and managing his calendar, MARVIN runs as a full-time AI chief of staff.
    We cover:
    How MARVIN bookends Sterling's workday from first login to the end of the day
    Personality, sub-agents, and writing rules that make MARVIN an effective co-worker
    Automating meeting notes to Jira tickets
    Why DIY assistants outperform big tech alternatives
    How Sterling onboarded 12+ colleagues at Postman, including non-technical knowledge workers
    What the compute crunch means for open source AI
    Connect with Sterling:
    LinkedIn: https://www.linkedin.com/in/sterlingchin/
    Twitter/X: https://x.com/SilverJaw82
    MARVIN Template: https://github.com/SterlingChin/marvin-template

    Connect with Conor:
    Newsletter:⁠ ⁠https://conorbronsdon.substack.com/
    Twitter/X:⁠ https://x.com/ConorBronsdon⁠
    LinkedIn:⁠ https://www.linkedin.com/in/conorbronsdon
    YouTube:⁠⁠ https://www.youtube.com/@ConorBronsdon⁠⁠

    🔗 More episodes:⁠⁠ https://chainofthought.show⁠⁠

    Timestamps:
    (0:00) Intro
    (0:28) Meet Sterling Chin and the MARVIN AI Assistant
    (9:10) Live Demo: How MARVIN Bookends Your Workday
    (16:04) Personality, Sub-Agents, and Writing Rules
    (22:00) Automating Meeting Notes to Jira Tickets
    (29:30) Why DIY AI Assistants Outperform Big Tech
    (40:55) Treat Your AI Like a Junior Employee
    (46:41) How to Get Started with MARVIN
    (55:36) The Compute Crunch and Open Source Future

    Thanks to Galileo — download their free 165-page guide to mastering multi-agent systems at galileo.ai/mastering-multi-agent-systems
  • Chain of Thought | AI Agents, Infrastructure & Engineering

    How Intercom Cut $250K/Month by Ditching GPT for Qwen

    26/02/2026 | 53 min
    Intercom was spending $250K/month on a single summarization task using GPT. Then they replaced it with a fine-tuned 14B parameter Qwen model and saved almost all of it. In this episode, Intercom's Chief AI Officer, Fergal Reid, walks through exactly how they made that call, where their approach has changed over time, and how all of their efforts built their Fin customer service agent.
    Fergal breaks down how Fin went from 30% to nearly 70% resolution rate and why most of those gains came from surrounding systems (custom re-rankers, retrieval models, query canonicalization), not the core frontier LLM. He explains why higher latency counterintuitively increases resolution rates, how they built a custom re-ranker that outperformed Cohere using ModernBERT, and why he believes vertically integrated AI products will win in the long term.
    If you're deciding between fine-tuning open-weight models and using frontier APIs in production, you won't find a more detailed decision process walkthrough.
    🔗 Connect with Fergal: 
    Twitter/X: https://x.com/fergal_reid

    LinkedIn: https://www.linkedin.com/in/fergalreid/

    Fin: https://fin.ai/

    🔗 Connect with Conor:
    YouTube: https://www.youtube.com/@ConorBronsdon

    Newsletter: https://conorbronsdon.substack.com/

    Twitter/X: https://x.com/ConorBronsdon

    LinkedIn: https://www.linkedin.com/in/conorbronsdon/

    🔗 More episodes: https://chainofthought.showCHAPTERS
    0:00 Intro
    0:46 Why Intercom Completely Reversed Their Fine-Tuning Position
    8:00 The $250K/Month Summarization Task (Query Canonicalization)
    11:25 Training Infrastructure: H200s, LoRA to Full SFT, and GRPO
    14:09 Why Qwen Models Specifically Work for Production
    18:03 Goodhart's Law: When Benchmarks Lie
    19:47 A/B Testing AI in Production: Soft vs. Hard Resolutions
    25:09 The Latency Paradox: Why Slower Responses Get More Resolutions
    26:33 Why Per-Customer Prompt Branching Is Technical Debt
    28:51 Sponsor: Galileo
    29:36 Hiring Scientists, Not Just Engineers
    32:15 Context Engineering: Intercom's Full RAG Pipeline
    35:35 Customer Agent, Voice, and What's Next for Fin
    39:30 Vertical Integration: Can App Companies Outrun the Labs?
    47:45 When Engineers Laughed at Claude Code
    52:23 Closing Thoughts
    TAGSFergal Reid, Intercom, Fin AI agent, open-weight models, Qwen models, fine-tuning LLMs, post-training, RAG pipeline, customer service AI, GRPO reinforcement learning, A/B testing AI, Claude Code, vertical AI integration, inference cost optimization, context engineering, AI agents, ModernBERT reranker, scaling AI teams, Conor Bronsdon, Chain of Thought
  • Chain of Thought | AI Agents, Infrastructure & Engineering

    How Block Deployed AI Agents to 12,000 Employees in 8 Weeks w/ MCP | Angie Jones

    21/01/2026 | 50 min
    How do you deploy AI agents to 12,000 employees in just 8 weeks? How do you do it safely? Angie Jones, VP of Engineering for AI Tools and Enablement at Block, joins the show to share exactly how her team pulled it off.

    Block (the company behind Square and Cash App) became an early adopter of Model Context Protocol (MCP) and built Goose, their open-source AI agent that's now a reference implementation for the Agentic AI Foundation. Angie shares the challenges they faced, the security guardrails they built, and why letting employees choose their own models was critical to adoption.

    We also dive into vibe coding (including Angie's experience watching Jack Dorsey vibe code a feature in 2 hours), how non-engineers are building their own tools, and what MCP unlocks when you connect multiple systems together.

    Chapters:
    00:00 Introduction
    02:02 How Block deployed AI agents to 12,000 employees
    05:04 Challenges with MCP adoption and security at scale
    07:10 Why Block supports multiple AI models (Claude, GPT, Gemini)
    08:40 Open source models and local LLM usage
    09:58 Measuring velocity gains across the organization
    10:49 Vibe coding: Benefits, risks & Jack Dorsey's 2-hour feature build
    13:46 Block's contributions to the MCP protocol
    14:38 MCP in action: Incident management + GitHub workflow demo
    15:52 Addressing MCP criticism and security concerns
    18:41 The Agentic AI Foundation announcement (Block, Anthropic, OpenAI, Google, Microsoft)
    21:46 AI democratization: Non-engineers building MCP servers
    24:11 How to get started with MCP and prompting tips
    25:42 Security guardrails for enterprise AI deployment
    29:25 Tool annotations and human-in-the-loop controls
    30:22 OAuth and authentication in Goose
    32:11 Use cases: Engineering, data analysis, fraud detection
    35:22 Goose in Slack: Bug detection and PR creation in 5 minutes
    38:05 Goose vs Claude Code: Open source, model-agnostic philosophy
    38:17 Live Demo: Council of Minds MCP server (9-persona debate)
    45:52 What's next for Goose: IDE support, ACP, and the $100K contributor grant
    47:57 Where to get started with Goose

    Connect with Angie on LinkedIn: https://www.linkedin.com/in/angiejones/
    Angie's Website: https://angiejones.tech/
    Follow Angie on X: https://x.com/techgirl1908
    Goose GitHub: https://github.com/block/goose

    Connect with Conor on LinkedIn: https://www.linkedin.com/in/conorbronsdon/
    Follow Conor on X: https://x.com/conorbronsdon
    Modular: https://www.modular.com/

    Presented By: Galileo AI
    Download Galileo's Mastering Multi-Agent Systems for free here: https://galileo.ai/mastering-multi-agent-systems

    Topics Covered:
    - How Block deployed Goose to all 12,000 employees
    - Building enterprise security guardrails for AI agents
    - Model Context Protocol (MCP) deep dive
    - Vibe coding benefits and risks
    - The Agentic AI Foundation (Block, Anthropic, OpenAI, Google, Microsoft, AWS)
    - MCP sampling and the Council of Minds demo
    - OAuth authentication for MCP servers
    - Goose vs Claude Code and other AI coding tools
    - Non-engineers building AI tools
    - Fraud detection with AI agents
    - Goose in Slack for real-time bug fixing

Altri podcast di Economia

Su Chain of Thought | AI Agents, Infrastructure & Engineering

AI is reshaping infrastructure, strategy, and entire industries. Host Conor Bronsdon talks to the engineers, founders, and researchers building breakthrough AI systems about what it actually takes to ship AI in production, where the opportunities lie, and how leaders should think about the strategic bets ahead. Chain of Thought translates technical depth into actionable insights for builders and decision-makers. New episodes bi-weekly. Conor Bronsdon is an angel investor in AI and dev tools, Head of Technical Ecosystem at Modular, and previously led growth at AI startups Galileo and LinearB.
Sito web del podcast

Ascolta Chain of Thought | AI Agents, Infrastructure & Engineering, STORIE DI BRAND e molti altri podcast da tutto il mondo con l’applicazione di radio.it

Scarica l'app gratuita radio.it

  • Salva le radio e i podcast favoriti
  • Streaming via Wi-Fi o Bluetooth
  • Supporta Carplay & Android Auto
  • Molte altre funzioni dell'app

Chain of Thought | AI Agents, Infrastructure & Engineering: Podcast correlati