Workflow
How DeepSeek used distillation to train its artificial intelligence model, and what it means for companies such as OpenAI
GOOGLAlphabet(GOOGL) CNBC·2025-02-21 13:00

Core Insights - The emergence of DeepSeek has triggered significant market reactions, particularly a selloff in tech and semiconductor sectors, due to its introduction of cost-effective and efficient AI models compared to American counterparts [1] - The technique of distillation in AI development is gaining traction, allowing smaller teams to create advanced models quickly and with minimal resources, thus reshaping the competitive landscape in the AI industry [2][4] Group 1: Distillation Technique - Distillation is a method for extracting knowledge from larger AI models to create smaller, specialized models, enabling rapid development with fewer resources [2][3] - This technique allows smaller teams to train models that are nearly as capable as those developed by larger companies, but in a more efficient manner [3] Group 2: Market Dynamics and Competition - The distillation technique is expected to foster increased competition in the large language model (LLM) space, as it democratizes access to advanced AI capabilities [4] - Researchers have demonstrated the effectiveness of distillation, with examples such as recreating OpenAI's reasoning model for 450in19hoursandanothermodelinjust26minutesforunder450 in 19 hours and another model in just 26 minutes for under 50 [5] Group 3: Open Source Movement - DeepSeek has highlighted the disruptive potential of distillation and has contributed to the rise of an open-source movement in AI, emphasizing transparency and accessibility as key drivers of innovation [6] - OpenAI has acknowledged the need to adapt its strategy towards open-source in response to the competitive pressures introduced by distillation and the success of open-source projects [7]