Workflow
Attention heads
icon
Search documents
X @Avi Chawla
Avi Chawla· 2026-04-19 21:09
Mixture of Experts (MoEs), explained visually:(learn how they work below) https://t.co/HG4CN5gGBdAvi Chawla (@_avichawla):You're in an ML Engineer interview at MistralAI.The interviewer asks:"We need an LLM that excels across code, math & creative writing. How do you achieve multi-domain performance?"You: "I'll increase the number of attention heads."Interview over.Here's what you missed: ...