Workflow
Qwen3
icon
Search documents
林俊旸离职后首次发声,复盘千问的弯路,指出AI的新路
36氪· 2026-03-27 11:12
Core Insights - The article discusses the transition from "Reasoning Thinking" to "Agentic Thinking" in AI, emphasizing the need for models to adapt and interact with their environments rather than just providing static answers [4][14][73] - Lin Junyang acknowledges that the previous approaches did not fully succeed, indicating a need for improvement in AI model integration and performance [7][30] Group 1: Transition in AI Thinking - The past two years have defined the mission of Reasoning Thinking, with significant advancements in training models for reasoning capabilities [11][13] - The emergence of Agentic Thinking is seen as the next step, focusing on continuous interaction with the environment and adjusting plans based on real-world feedback [14][49] - Key differences between Reasoning Thinking and Agentic Thinking include the ability to decide when to act, manage tool selection dynamically, and maintain coherence across multiple interactions [11][50] Group 2: Infrastructure and Environment Design - The rise of reasoning models highlights the importance of robust infrastructure and the need for scalable feedback signals in reinforcement learning [16][21] - As the focus shifts to Agentic Thinking, the design of the environment becomes crucial, emphasizing stability, authenticity, and the ability to generate diverse trajectories [59][60] - The integration of tools and the environment into the training process is essential for developing effective AI systems, moving beyond traditional model training [56][71] Group 3: Future Directions and Challenges - The future of AI is expected to revolve around training intelligent agents rather than just models, with a focus on system-level training that includes both the model and its environment [71][73] - The definition of "good thinking" is evolving, prioritizing the ability to maintain effective action under real-world constraints rather than merely producing lengthy reasoning outputs [75] - Competitive advantages in the Agentic Thinking era will stem from better environmental design, tighter training-reasoning coupling, and effective orchestration of multiple agents [77]
林俊旸离职后首次发声:复盘千问的弯路,指出AI的新路
创业邦· 2026-03-27 07:18
Core Insights - The article discusses the transition from "Reasoning Thinking" to "Agentic Thinking" in AI, emphasizing the need for models to not only think but also act effectively in real-world environments [5][20][27] Group 1: Transition in AI Thinking - Lin Junyang reflects on the shortcomings of the Qwen team's ambitious goal to merge thinking and instruct modes into a single model, highlighting that true success lies in a continuous spectrum of reasoning efforts rather than a forced combination [5][10] - The emergence of models like OpenAI's o1 and DeepSeek-R1 has demonstrated that reasoning capabilities can be trained and scaled, leading to a critical understanding in the industry about the necessity of strong, scalable feedback signals for reinforcement learning [8][9] Group 2: Key Differences in Thinking Models - Agentic Thinking differs from Reasoning Thinking in that it requires models to continuously switch between thinking and acting, manage tool selection dynamically, and adapt to environmental feedback [6][22] - The focus has shifted from merely extending reasoning time to ensuring that models can think in a way that maintains effective action, thus redefining the evaluation criteria for AI models [20][27] Group 3: Infrastructure and Environment Design - The infrastructure for reinforcement learning must evolve to support the complexities of Agentic Thinking, necessitating a decoupling of training and reasoning processes to avoid inefficiencies [19][21] - The quality of the environment in which models operate is becoming a critical factor, with emphasis on stability, authenticity, and diversity of states, marking a shift from data diversity to environment quality [23][27] Group 4: Future Directions - The article predicts that Agentic Thinking will become the mainstream cognitive approach, potentially replacing traditional static reasoning methods, as systems become more capable of interacting with their environments [24][25] - The rise of harness engineering is highlighted, where the organization of multiple agents will play a crucial role in enhancing core intelligence and operational efficiency [25][27]
林俊旸离职后首度发声:万字复盘,大模型下一站「智能体式思考」
机器之心· 2026-03-27 00:10
Core Insights - The article discusses the evolution of large language models over the past two years, particularly focusing on the transition from "reasoning" thinking to "agentic" thinking in AI development [3][29]. Group 1: Evolution of Large Models - The emergence of models like OpenAI's o1 and DeepSeek's R1 has taught the industry about the importance of deterministic, stable, and scalable feedback signals for expanding reinforcement learning in language models [6][7]. - The shift from expanding pre-training scale to expanding post-training scale for reasoning is highlighted as a significant transformation in model development [7]. Group 2: Integration of Thinking and Instruction - The Qwen team envisioned a system that merges "thinking" and "instruction" modes, allowing adjustable reasoning intensity based on user prompts and context [9][10]. - The challenge lies in the fundamentally different data distributions and behavior goals required for these two modes, making it difficult to achieve effective integration [10][11]. - Maintaining separation between "thinking" and "instruction" modes is seen as a more attractive option for practical applications, allowing teams to focus on specific training challenges [11][12]. Group 3: Anthropic's Approach - Anthropic's Claude 3.7 and Claude 4 models emphasize integrated reasoning capabilities and user-controllable "thinking budgets," aiming to enhance practical task performance [14][15]. - The development trajectory of Anthropic reflects a rigorous approach, shaping the thinking process based on specific workloads rather than generating verbose outputs [16]. Group 4: Agentic Thinking - Agentic thinking sets a different optimization goal, focusing on the model's ability to make progress through interaction with the environment rather than just internal reasoning quality [17][18]. - The transition to agentic reinforcement learning requires a more complex infrastructure, integrating various components like tool servers and APIs into the training framework [19][20]. Group 5: Future Directions - The next frontier is expected to be agentic thinking, which may replace static reasoning models by enabling systems to perform searches, simulations, and code execution in a robust manner [23][24]. - Challenges such as "reward hacking" and ensuring effective interaction with external tools will be critical in the development of these systems [25][26]. - The evolution from training models to training entire agent systems is anticipated, emphasizing the importance of environment design and coordination among multiple agents [27][30].
林俊旸离职后首次发声!复盘千问的弯路,指出AI的新路
量子位· 2026-03-26 16:01
Core Insights - The article discusses the transition from "Reasoning Thinking" to "Agentic Thinking" in AI, emphasizing the need for models to adapt and interact with their environments for effective decision-making [2][12][73] - It reflects on the shortcomings of the Qwen team's ambitious goal to merge thinking and instruction modes into a single model, acknowledging that not everything was executed correctly [5][36] Group 1: Transition in AI Thinking - The past two years have redefined how models are evaluated and the expectations placed on them, moving towards a focus on interaction with the environment [15][73] - The emergence of models like OpenAI's o1 and DeepSeek-R1 has demonstrated that reasoning capabilities can be trained and scaled, highlighting the importance of strong, scalable feedback signals [9][23][27] - The industry is now focused on enhancing reasoning time, training stronger rewards, and controlling reasoning intensity [11][21] Group 2: Agentic Thinking - Agentic Thinking is defined as thinking for action, continuously adjusting plans based on environmental interactions [12][54] - The key difference between Agentic Thinking and Reasoning Thinking is summarized as moving from "thinking longer" to "thinking for action" [13][54] - Future competitiveness will rely not only on better models but also on improved environment design, harness engineering, and orchestration among multiple agents [13][71] Group 3: Challenges in Merging Thinking and Instruction - The ideal system should unify thinking and instruction modes, allowing for adjustable reasoning intensity based on context [30][31] - The difficulty lies in the fundamental differences in data distribution and behavioral objectives between the two modes, which can lead to mediocre performance if not carefully managed [36][38] - Many organizations are exploring different approaches, with some advocating for integrated models while others prefer to keep instruction and thinking separate for better focus on each mode's unique challenges [39][40][42] Group 4: Infrastructure and Environment Design - The transition to Agentic Thinking necessitates a shift in infrastructure, as the classic reasoning RL setup is insufficient for interactive tasks [56][61] - The environment becomes a critical component of the training system, requiring a focus on quality, stability, and diversity [61][62] - The next frontier in AI development will involve creating more usable thinking processes that prioritize effective action over lengthy reasoning [62][69] Group 5: Future Directions - The article concludes that the shift from reasoning to agentic thinking changes the definition of "good thinking" to maintaining effective action under real-world constraints [75][76] - Competitive advantages in the agentic era will stem from better environment design, tighter training-reasoning coupling, and effective orchestration of multiple agents [76]
阿里巴巴:推出玄铁 C950 AI 芯片
2026-03-26 13:20
Summary of Alibaba Group Holding Conference Call Company Overview - **Company**: Alibaba Group Holding (BABA.N, BABA UN) - **Industry**: China Internet and Other Services - **Date of Call**: March 24, 2026 Key Developments - **Launch of XuanTie C950 AI Chip**: Alibaba introduced its next-generation AI chip, the XuanTie C950, which is a 5-nanometer processor based on open-source RISC-V architecture. This chip is reported to perform over 3 times faster than its predecessor and supports large models such as Qwen3 and DeepSeek V3 [2][3]. Core Insights - **Full AI Stack Ownership**: Alibaba is viewed as owning the complete AI stack, which includes in-house chips (T-Head), cloud infrastructure (AliCloud), state-of-the-art open-weight models (Qwen), and consumption-centric applications (Qwen apps). This vertical integration is expected to reduce reliance on third-party suppliers, enable application-specific designs, and support rapid capacity expansion during demand spikes [3][4]. - **Financial Performance of T-Head**: For the first time, management disclosed operational and financial achievements of T-Head, including: - Cumulative shipment of over 470,000 units - Revenue exceeding RMB 10 billion - More than 60% of the mix serving external AliCloud customers - Potential for a spin-off or separate listing, although no specific timeline was provided [9]. Valuation and Market Position - **Valuation of T-Head**: The T-Head division is valued between US$28 billion to US$86 billion, translating to approximately US$22 per share. This is part of a sum-of-the-parts (SOTP) valuation of US$245 at the midpoint [3]. - **Stock Rating**: The stock is rated as a "Top Pick" with a price target of US$180, indicating a potential upside of 43% from the closing price of US$126.06 on March 23, 2026 [5][9]. Risks and Considerations - **Upside Risks**: - Improved core e-commerce monetization could drive earnings growth - Accelerated enterprise digitalization may boost cloud revenue - Increased demand for AI could enhance cloud revenue [12]. - **Downside Risks**: - Intense competition in the market - Higher-than-expected reinvestment costs - Weaker consumer spending amid a slower post-COVID recovery - Regulatory scrutiny of internet platforms [12]. Conclusion Alibaba's advancements in AI technology, particularly with the launch of the XuanTie C950 chip, position the company favorably within the competitive landscape of the internet services industry. The company's integrated approach to AI and cloud services, along with strong financial metrics from its T-Head division, supports a positive outlook despite potential market risks.
不是所有token都平等,谷歌提出真·深度思考:思维链长≠深度推理
3 6 Ke· 2026-02-25 12:23
Core Insights - Google's research challenges the long-held belief that longer reasoning chains in large models lead to better inference quality, introducing a new metric called Deep Thinking Ratio (DTR) to assess true cognitive depth rather than mere token count [1][3][9]. Group 1: Research Findings - The study found a negative correlation of -0.54 between token length and accuracy across various models, indicating that longer reasoning chains can lead to misdirection and overthinking [3][5]. - DTR measures the proportion of "deep thinking tokens" in a generated sequence, with a higher ratio indicating a focus on core reasoning rather than unnecessary content [8][10]. Group 2: Implementation of DTR - Google introduced the Think@n strategy, which allows models like GPT-OSS and DeepSeek-R1 to maintain accuracy while halving computational costs by filtering out low-quality samples early in the reasoning process [2][12]. - In tests, the Think@n strategy achieved an accuracy of 94.7% for GPT-OSS-120B-medium on the AIME 2025 dataset, surpassing traditional methods, while reducing token consumption from 355.6k to 181.9k [12][13]. Group 3: Implications for Model Development - The findings suggest a shift in focus for model developers from merely increasing token length to enhancing the quality of reasoning, emphasizing the importance of deep cognitive processing [1][19]. - The research highlights the potential for significant cost savings and efficiency improvements in model inference through the application of DTR and the Think@n strategy [9][12].
DeepSeek、月之暗面、MiniMax被点“非法提取”,它们做错了吗? | 电厂
Xin Lang Cai Jing· 2026-02-25 10:47
Core Viewpoint - Anthropic has accused three Chinese AI companies—DeepSeek, Moonshot, and MiniMax—of illicitly extracting data from its model Claude, marking the second controversy involving domestic models within three months [1][9]. Group 1: Allegations and Responses - Anthropic claims that the three Chinese companies used approximately 24,000 fraudulent accounts to interact with Claude over 16 million times, using these interactions to enhance their own models [1][4]. - The accused companies have remained silent regarding the allegations, with no public response from DeepSeek, MiniMax, or Moonshot [1]. - Anthropic's statement highlighted that the interaction patterns with Claude were abnormal, indicating intentional extraction of Claude's unique capabilities [7]. Group 2: Technical Aspects of Distillation - The technique used by the accused companies is known as "distillation," which allows models to learn from a "teacher model" like Claude by interacting with it [4][6]. - Distillation is a common method for rapidly evolving models, enabling smaller models to approximate the performance of larger ones with less data [6]. - Major AI companies, including OpenAI and Google, have included clauses in their usage agreements prohibiting distillation, reflecting a growing concern over intellectual property [9]. Group 3: Legal and Ethical Considerations - The ongoing debate over model distillation raises questions about legal definitions, including contract law, copyright law, and unfair competition [10]. - Both Chinese and American companies utilize vast amounts of internet data for training, leading to discussions about authorization and ethical use of such data [10]. - The narrative surrounding "Chinese companies distilling American models" has become a one-sided discourse, with the potential for a prolonged public relations battle [10]. Group 4: Open Source vs. Closed Source Models - Many leading Chinese models operate under open-source licenses that permit distillation, contrasting with the closed-source models that prohibit such practices [10][13]. - For instance, DeepSeek's models are released under the MIT license, allowing for academic and commercial use, while other models like MiniMax and Qwen3 follow the Apache 2.0 license [10]. - The controversy over distillation also highlights the ongoing debate between open-source and closed-source development paths in the AI industry [13].
Rokid Glasses支持OpenClaw及私有大模型自定义接入
Bei Jing Shang Bao· 2026-02-11 12:53
Core Insights - Rokid has launched the "Customizable Intelligent Agent" feature on its Lingzhu platform, marking a significant shift in user control over AI glasses [1] Group 1: Product Development - The new feature is described as not merely a simple iteration but as the beginning of returning the definition of AI glasses to the users [1] - Users can now connect Rokid Glasses to any desired backend through a standard SSE (Server-Sent Events) interface [1] Group 2: Market Positioning - The integration allows compatibility with popular platforms such as OpenClaw and private deployments like DeepSeek R1, Qwen3, and Kimi K2.5 [1]
传阿里巴巴新一代模型Qwen3.5发布在即
Zhi Tong Cai Jing· 2026-02-09 07:21
Core Insights - The latest development in the AI open-source community is the emergence of Qwen3.5 integrated into Transformers, indicating that Alibaba's new generation base model is nearing release [1] - Qwen3.5 is expected to utilize a novel hybrid attention mechanism and may be a VLM model capable of visual understanding, with potential open-source offerings of at least 2 billion dense models and 35 billion MoE models [1] - Previous reports suggested that Qwen3.5 would be open-sourced during the Spring Festival [1] - On April 29, 2025, Alibaba released the new Qwen3 model, which became the strongest open-source model globally, featuring a "hybrid reasoning model" that integrates "fast thinking" and "slow thinking," significantly reducing computational power consumption [1]
传阿里巴巴(09988)新一代模型Qwen3.5发布在即
智通财经网· 2026-02-09 07:21
Core Viewpoint - The latest development in the AI sector involves the anticipated release of Alibaba's Qwen3.5 model, which is expected to be integrated into the Transformers framework, indicating significant advancements in open-source AI technology [1] Group 1: Model Features - Qwen3.5 utilizes a new hybrid attention mechanism, potentially making it a vision-language model (VLM) capable of native visual understanding [1] - The model is expected to include at least a 2 billion dense model and a 35 billion MoE model, showcasing its scalability and complexity [1] Group 2: Release Timeline - Reports suggest that Qwen3.5 will be open-sourced during the Spring Festival period, indicating a strategic timing for its release [1] - Alibaba previously launched the new generation Qwen3 model on April 29, 2025, which was recognized as the strongest open-source model globally [1] Group 3: Technological Innovation - Qwen3 is noted as the first "hybrid reasoning model" in China, integrating "fast thinking" and "slow thinking" within a single framework, significantly reducing computational resource consumption [1]