Published onFebruary 9, 2026|Views: 23|37 min readUnderstanding DeepSeek's Multi-Head Latent Attention (MLA)llmattentiontransformersdeepseekmlakv-cacheinferenceOn bottlenecks in attention, kv caching, long-context decoding, attention variants, and how DeepSeek MLA came to be. Part 1 of the FlashMLA blog series.