When LLMs write code to accomplish a task, that code has to actually run somewhere. And right now, the options aren't great. Spin up a sandboxed container and you're paying a full second of cold start overhead plus the complexity of another service. Let the LLM loose on your actual machine and... well, you'd better be watching.
On this episode, I sit down with Samuel Colvin, creator of Pydantic, now at 10 billion downloads, to explore Monty, a Python interpreter written from scratch in Rust, purpose-built to run LLM-generated code. It starts in microseconds, is completely sandboxed by design, and can even serialize its entire state to a database and resume later. We dig into why this deliberately limited interpreter might be exactly what the AI agent era needs.
If you've built documentation in the Python ecosystem, chances are you've used Martin Donath's work. His Material for MKDocs powers docs for FastAPI, uv, AWS, OpenAI, and tens of thousands of other projects. But when MKDocs 2.0 took a direction that would break Material and 300 ecosystem plugins, Martin went back to the drawing board. The result is Zensical: A new static site generator with a Rust core, differential builds in milliseconds instead of minutes, and a migration path designed to bring the whole community along.
When you type a question into ChatGPT, the model only has what you typed to work with. But tools like Claude Code can plan, iterate, test, and recover from mistakes. They work more like we do. The difference is the agent harness: Planning tools, file system access, sub-agents, and carefully crafted system prompts that turn a raw LLM into something genuinely capable.
Sydney Runkle is back on Talk Python representing LangChain and their new open source library, Deep Agents: A framework for building your own deep agents with plain Python functions, middleware hooks, and MCP support. This is how the magic works under the hood.
When you pip install a package with compiled code, the wheel you get is built for CPU features from 2009. Want newer optimizations like AVX2? Your installer has no way to ask for them. GPU support? You're on your own configuring special index URLs. The result is fat binaries, nearly gigabyte-sized wheels, and install pages that read like puzzle books. A coalition from NVIDIA, Astral, and QuantSight has been working on Wheel Next: A set of PEPs that let packages declare what hardware they need and let installers like uv pick the right build automatically. Just uv pip install torch and it works. I sit down with Jonathan Dekhtiar from NVIDIA, Ralf Gommers from QuantSight and the NumPy and SciPy teams, and Charlie Marsh, founder of Astral and creator of uv, to dig into all of it.
Add Talk Python to your favorite podcast player and never miss an episode.
Subscribe NowGet episode updates, Python news, and exclusive content delivered to your inbox.
Subscribe to NewsletterOnline courses from Talk Python. Real courses with real coding helping you build production-ready skills.
Looking to level up your entire team's Python skills? Talk Python Training offers bulk discounts for organizations, plus an admin dashboard to track team progress and completion.
Learn about team plans
Learn how to build, deploy, and scale production-ready Python applications. From best practices to real-world patterns, this book covers everything you need to take your Python projects from prototype to production.
Search across 542+ episodes to find exactly what you're looking for.
Search episodesDo you enjoy Talk Python To Me? We created a second podcast to be the perfect counterpart to the long-form interview format.
Python Bytes delivers Python headlines directly to your earbuds in just 15-20 minutes per week. The topical items of the Python ecosystem, condensed and ready to go.
Talk Python to Me is a weekly podcast covering a wide array of Python topics as well as many related technologies. Each episode features a casual 1-hour conversation with industry experts.
Have feedback for the show? We'd love to hear from you.
We are always on the look out for exciting topics to bring to you on Talk Python To Me. Do you have a great idea for a show? Know a dynamic Python personality who we should be speaking to?