← All Tags

#transformers

17 episodes

#2066: The Transformer Trinity: Why Three Architectures Rule AI

Why did decoder-only models like GPT dominate AI, while encoders and encoder-decoders still hold critical niches?

transformersai-modelslarge-language-models

#2062: How Transformers Learn Word Order: From Sine Waves to RoPE

Transformers can’t see word order by default. Here’s how positional encoding fixes that—from sine waves to RoPE and massive context windows.

transformerscontext-windowlarge-language-models

#2061: How Attention Variants Keep LLMs From Collapsing

Attention is the engine of modern AI, but it’s also a memory hog. Here’s how MQA, GQA, and MLA evolved to fix it.

transformersai-modelsattention-mechanisms

#2056: How Music Models Turn Sound Into Language

A look at how AI music models use audio tokens, transformers, and diffusion to turn text into songs.

audio-processingtransformersgenerative-ai

#1799: The Original AI Blueprints: BERT & CLIP

Before GPT, two models changed everything. Discover how BERT and CLIP taught machines to read and see the world.

transformersai-historycomputer-vision

#1679: Chinese AI Is Built Different—Here's How

DeepSeek and MiMo are topping developer charts, but they're not just cheaper clones. Here's why their design philosophy is fundamentally different.

ai-modelstransformerslocal-ai

#1666: Multi-Agent AI: One Model, Four Brains

Grok 4.20’s native multi-agent architecture cuts token costs by 75% and enables real-time cross-agent reasoning.

ai-agentstransformersrag

#1633: Agent Interview: MiniMax M two point seven

We grill MiniMax M2.7 to see if a model built for "virtual companions" can actually handle high-level comedy and complex character logic.

ai-agentsai-reasoningtransformers

#1632: Agent Interview: DeepSeek V three point two

We interview DeepSeek V3 to see if this open-weight powerhouse can handle weird podcast prompts better than big tech’s flagship models.

ai-agentsopen-source-aitransformers

#1604: The $3 Billion Stealth Giant: AI21 Labs & Nvidia

Why is Nvidia eyeing a $3B deal for AI21 Labs? Discover the tech behind the "OpenAI of Israel" and their revolutionary hybrid architecture.

large-language-modelsstate-space-modelstransformers

#1547: Why AI Stopped Reading and Started Seeing Everything

From sequential bottlenecks to parallel powerhouses, discover how the Transformer architecture revolutionized how machines process the world.

transformersai-historyparallel-computing

#1108: Beyond the Emoji: How Hugging Face Conquered AI

Discover how a quirky chatbot company became the central nervous system of AI, hosting millions of models and standardizing the entire industry.

open-sourceai-modelstransformers

#135: Is OCR Dead? How Vision AI Is Redefining Text Extraction

Are specialized OCR tools obsolete? Herman and Corn explore how Vision Language Models are revolutionizing the way we turn images into data.

ocrvision-aivlmoptical-character-recognitiontext-extraction

#126: The Spotlight Effect: Understanding AI Attention Mechanisms

Why do AI models "lose the plot" after a few thousand words? Discover the mechanics of attention and the innovations solving context window limits.

attention-mechanismcontext-windowquadratic-scalingmambaring-attention

#26: Personalizing Whisper: The Voice Typing Revolution

Voice typing is changing everything. Join us as we explore the revolution of personalizing Whisper!

speech-recognitionfine-tuningtransformers

#19: AI Images: The Jigsaw Beneath the Magic

Beyond the prompt, discover the intricate 'jigsaw puzzle' of AI image generation. Uncover the magic's true workings.

transformersdiffusion-modelslatent-space

#12: The AI Breakthrough: Transformers & The Perfect Storm

AI's everywhere. How did chatbots, art, and video all emerge so suddenly? The secret lies in Transformers and a perfect storm.

transformersfine-tuninggpu-acceleration