#serverless-gpu
8 episodes
#2548: Static vs Server-Side: What Actually Happens When You Deploy
The moment you see content appear instantly on production and realize it wasn't pre-built — that's when architecture gets interesting.
#2303: Optimizing Podcast Pipelines: TTS Costs and Batch Processing
How batch processing and smart queue management can slash TTS costs for episodic podcast production.
#1927: Workers vs. Servers: The 2026 Compute Showdown
Is the persistent server dead? We compare Cloudflare Workers, GitHub Actions, and VPS options for modern app architecture.
#1926: How We Built a 2,000-Episode AI Podcast Engine
We pulled back the curtain on the tech stack behind our 1,858th episode. From Gemini to LangGraph, here’s how we automate quality.
#1820: Renting vs. Owning GPUs: The Break-Even Math
Is it cheaper to rent serverless GPUs or buy your own hardware? We break down the math on utilization, depreciation, and hidden costs.
#1778: Audio Is the New "Read Later" Graveyard
Why listening to AI conversations beats reading dense PDFs, and how serverless GPUs make it cheap.
#1491: Inside the Machine: Podcasting with AI Agents in 2026
Peek behind the curtain of a 2026 AI podcast, from agentic workflows to maintaining production during global conflict.
#346: GPU Scaling: The "Go Wide or Go Tall" Dilemma
Should you use a fleet of cheap GPUs or one powerhouse? Learn the math behind serverless GPU costs, cold starts, and batching efficiency.