
In this episode, Philip Kiely, head of AI education at Baseten, joins us to unpack the fast-evolving discipline of inference engineering. We explore why inference has become the stickiest and most critical workload in AI, how it blends GPU programming, applied research, and large-scale distributed systems, and where the line sits between inference and model serving. Philip shares how research-to-production can move in hours, not months, and why understanding “the knobs” of inference—batching, quantization, speculation, and KV cache reuse—lets teams design better products and SLAs. We trace the inference maturity journey from closed APIs to dedicated deployments and in-house platforms, discuss GPU lifecycles, and survey today’s runtime landscape, including vLLM, SGLang, and TensorRT LLM. Finally, we look ahead to agents and multimodality, making the case for specialized, workload-specific runtimes when performance and efficiency matter most. The complete show notes for this episode can be found at https://twimlai.com/go/766.
AI Summary coming soon
Sign up to get notified when the full AI-powered summary is ready.
Free forever for up to 3 podcasts. No credit card required.

How Capital One Delivers Multi-Agent Systems with Rashmi Shetty - #765

The Race to Production-Grade Diffusion LLMs with Stefano Ermon - #764

Agent Swarms and Knowledge Graphs for Autonomous Software Development with Siddhant Pardeshi - #763

AI Trends 2026: OpenClaw Agents, Reasoning LLMs, and More with Sebastian Raschka - #762
Free AI-powered recaps of The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence) and your other favorite podcasts, delivered to your inbox.
Free forever for up to 3 podcasts. No credit card required.