Workflow
阿里开源首个“混合推理模型”:集成“快思考”、“慢思考”能力

Core Insights - Alibaba has open-sourced its new generation model Qwen3, which integrates "fast thinking" and "slow thinking" capabilities, significantly reducing deployment costs compared to other large models like Deepseek [1] - The Qwen3 model employs a "Mixture of Experts (MoE)" architecture, allowing it to mimic human problem-solving by providing multi-step deep thinking for complex issues and quick responses for simpler queries, thus saving computational resources [3] - Alibaba is focusing on building its AI strategy around the Qwen series, with plans to invest over 380 billion RMB in cloud and AI hardware infrastructure over the next three years, surpassing the total investment of the past decade [4] Industry Context - Following the release of Deepseek's low-cost high-performance R1 model, domestic tech companies in China, including Baidu and iFlytek, are rapidly launching a series of cost-effective AI model services [3] - Alibaba's Qwen series has surpassed the US Llama in terms of open-source model downloads, with over 300 million downloads and more than 100,000 derivative models [4] - On the same day Alibaba announced Qwen3, OpenAI released several updates to ChatGPT, enhancing its shopping features and optimizing for various consumer categories, indicating a competitive landscape in AI model development [4]