Workflow
人工智能开源
icon
Search documents
智谱GLM-5正式开源,GLM Coding Plan同步升级
Zhi Tong Cai Jing· 2026-02-12 00:30
为了惠及更广泛的开发者社区,智谱(02513)即日起在Hugging Face与ModelScope平台同步开源GLM-5模 型权重,遵循MIT License。 与此同时,GLM Coding Plan服务也同步升级,为用户带来Agentic Engineering体验。新能力包括官方适 配OpenClaw,支持快速配置Agent工作流;并为Pro/Max用户限量赠送AutoGLM*OpenClaw,支持将云端 AI助手接入飞书等平台,实现智能自动化任务执行。 目前,GLM-5模型已上线智谱开发平台BigModel.cn、Z.ai及智谱清言(App/网页端),可供用户体验。需 要注意的是,由于算力限制,当前仅Max用户套餐中包含GLM-5调用额度,官方表示将逐步扩大开放范 围。 ...
LeCun创业0产品估值247亿,回应谢赛宁入伙
量子位· 2026-01-23 07:44
Group 1 - The core viewpoint of the article is that Yann LeCun, after leaving Meta, is launching a new company called Advanced Machine Intelligence (AMI), focusing on world models rather than large language models (LLMs) for achieving human-level intelligence [9][17][20] - LeCun criticizes Meta's product development decisions, stating that while research is acceptable, product execution has been poor, particularly under Mark Zuckerberg's leadership [2][3][15] - AMI aims to be an open-source platform, contrasting with the recent trend in Silicon Valley towards closed-source models, which LeCun believes is a misguided approach [11][13][16] Group 2 - The company will initially focus on research and development, specifically on world models, which LeCun argues are essential for building intelligent systems [17][19] - LeCun emphasizes that LLMs are not equivalent to AI and that understanding the real world is crucial for achieving human-like intelligence, which LLMs struggle to do [21][22][23] - AMI is seeking to raise €30 million (approximately 247 billion RMB) in funding, with an initial goal of €3.5 million for early financing, aiming for a total of €5 million in the first round [45][46][50] Group 3 - The company has already attracted interest from potential investors, including Cathay Innovation and Hiro Capital, indicating a shift in venture capital investment logic towards valuing founders over products [52][53][54] - LeCun is actively recruiting talent, including former Meta executives, to strengthen AMI's capabilities [40][42] - The ultimate goal of AMI is to become a leading supplier of intelligent systems, with a focus on practical applications of world models and planning capabilities [38][39]
上海发文明确:到2027年,打造1-2个具有国际影响力的开源社区
Yang Shi Wang· 2025-12-25 08:19
Core Viewpoint - The implementation plan for strengthening the open-source system in Shanghai aims to accelerate the development of an independent open-source ecosystem, positioning Shanghai as a national service hub and an internationally competitive center for open-source innovation by 2027 [1][4]. Group 1: Objectives - The primary goal is to establish Shanghai as a global open-source city within 3-5 years, fostering a robust open-source ecosystem that supports national service and international competitiveness [4]. - By 2027, the plan targets the creation of a comprehensive open-source system with complete technical support, a well-rounded service system, a reasonable talent structure, and a thriving open-source industry [6][8]. - The initiative aims to cultivate 100 commercial open-source enterprises and incubate over 200 high-quality open-source projects, gathering more than 3 million open-source developers [8]. Group 2: Implementation Strategies - The plan includes five major enhancement projects to build foundational open-source capabilities, such as lowering participation barriers for using open-source and establishing an international AI open-source community [9]. - It emphasizes the importance of creating a supportive ecosystem for open-source services, enhancing operational services for open-source projects, and promoting commercial capabilities [9][10]. - The establishment of a "dual launch" mechanism for open-source projects is proposed to encourage platforms to either launch or synchronize globally, along with annual selections of outstanding projects for support and promotion [11]. Group 3: Talent Development - The initiative focuses on nurturing developers who embrace open-source by collaborating with universities to set up training courses and practical projects related to open-source [15]. - It encourages the inclusion of open-source contributions in student evaluations and assessments, as well as the establishment of open-source departments in enterprises to attract critical talent [15][16]. Group 4: Cultural Promotion - The plan aims to promote open-source culture and facilitate international exchanges by hosting global open-source events and supporting the establishment of branches of renowned open-source organizations in Shanghai [18]. - It seeks to enhance the visibility of open-source culture through high-level conferences and market-oriented cultural products and services [19]. Group 5: Governance and Risk Management - The initiative emphasizes the need for coordinated governance and departmental collaboration to ensure the effective implementation of open-source tasks, with regular evaluations of progress [20].
昔日开源霸主承认蒸馏阿里千问,世界进入中国AI时间
3 6 Ke· 2025-12-11 11:39
Core Insights - The article highlights a significant shift in the global open-source AI landscape, with Meta's latest AI model "Avocado" utilizing Alibaba's Qwen open-source model for distillation training, marking a pragmatic turn for Meta amid the rise of Eastern models [2][5][10]. Group 1: Market Dynamics - Alibaba's stock price rose over 4% following the news of Meta's use of its Qwen model, indicating market confidence in Alibaba's technology [2]. - The number of Qwen derivative models surpassed 180,000, and its global download count exceeded 700 million by October 2025, both metrics surpassing Meta's Llama model [5][12]. - Alibaba has open-sourced over 300 models since 2023, covering a wide range of parameters and supporting 119 languages and dialects, enhancing its global developer appeal [5][12]. Group 2: Industry Trends - The trend of Silicon Valley integrating Chinese open-source AI technologies is evident, with major companies like Nvidia and Microsoft adopting Qwen-based solutions for complex multi-language and multi-modal data scenarios [6][12]. - Airbnb's CEO noted a 30-40% cost reduction when using Qwen models compared to other solutions, highlighting the competitive advantage of Chinese models [6][12]. - Academic institutions and innovative labs in Silicon Valley are increasingly adopting Qwen-based technologies, further validating its effectiveness [7][12]. Group 3: Technological Advancements - Alibaba's "full-stack AI" strategy is recognized for its comprehensive integration of technology, from chips to applications, supporting rapid model iteration and commercialization [10][14]. - The Qwen model's optimization for multi-language support provides a significant advantage in non-English markets, facilitating its global expansion [12][14]. - The Qwen application achieved over 30 million monthly active users within 23 days of its public launch, showcasing its rapid adoption and alignment with market needs [13][12]. Group 4: Future Outlook - The article suggests a structural reshaping of the global AI open-source ecosystem, with Meta's shift reflecting deeper changes in technological capabilities and competitive dynamics [12][14]. - The collaboration across cultural and geographical boundaries in AI innovation is expected to accelerate, with the best technological solutions emerging from the most open ecosystems [14].
中国模型厂商开辟“开源战场”,顶层设计再添一把火
Di Yi Cai Jing· 2025-08-29 06:58
Core Insights - Open source is not merely a technical means but is becoming a key mechanism to drive the AI ecosystem and industry implementation [1][5] Group 1: Government Initiatives - The State Council has issued opinions to deepen the implementation of the "AI+" action, emphasizing the enhancement of foundational model capabilities and the promotion of a thriving open-source ecosystem [2] - The development of a robust open-source community is seen as essential for gathering global developers and forming a vast network for technical exchange and innovation, which can enhance the influence of Chinese models on the world stage [2] Group 2: Industry Performance - Chinese model vendors are experiencing a dual scenario of "open source for a leapfrog advantage" and "technical breakthroughs facing bottlenecks," achieving significant rankings on international lists while still confronting challenges in commercializing their innovations [2][4] - Chinese open-source models, such as those from Alibaba and DeepSeek, are now comparable in performance to top proprietary models, marking a significant advancement in the global AI landscape [4] Group 3: Competitive Landscape - The competitive landscape is characterized by a dichotomy where leading firms either choose not to open-source or only open-source non-core models, with proprietary models used to maintain competitive advantages [8] - Despite the advantages of open-source models, such as customization and cost savings, there remains a performance gap of 9 to 12 months compared to leading proprietary models, influencing enterprise preferences [8] Group 4: Future Outlook - Open source is expected to evolve beyond a technical tool to become a critical mechanism for driving AI ecosystem and industry implementation, facilitating innovation and collaboration [5][9] - The relationship between open-source and proprietary models is seen as complementary, with open-source potentially leading breakthroughs in industry customization and multi-modal innovation [9] Group 5: Commercialization Challenges - Chinese open-source projects face challenges in commercializing effectively compared to their overseas counterparts, primarily due to different market environments and user habits [9] - The path to overcoming these challenges lies in globalization, where establishing a developer base in international communities can enhance customer willingness to pay and brand recognition [9]
马斯克开源Grok 2.5:中国公司才是xAI最大对手
3 6 Ke· 2025-08-24 23:25
Core Points - xAI has officially open-sourced Grok 2.5, with Grok 3 expected to be released in six months [1] - Elon Musk had previously indicated that it was time to open-source Grok, which was initially expected to happen the following week [3] - Despite the delay, the sentiment remains positive as many believe that it is better late than never [4] Summary by Sections Open Source Release - Grok can now be downloaded from HuggingFace, consisting of 42 files with a total size of approximately 500GB [5] - The official recommendation is to use SGLang to run Grok 2, with specific steps provided for downloading and setting up the model [6][7] Technical Specifications - The model requires 8 GPUs, each with over 40GB of memory, to operate effectively [8][17] - The initial setup involves downloading weight files and launching the inference server using SGLang [8] Performance Metrics - Grok 2 has shown competitive performance in various academic benchmarks, including GPQA, MMLU, and MATH, with scores that rival leading models [13] - In the LMSYS ranking, Grok 2 surpassed Claude and GPT-4 in overall Elo scores [10] Community Feedback - There are mixed reactions regarding the open-source model, particularly concerning the lack of clarity on the model's parameter weights, which are speculated to be 269 billion for the MoE model [15] - The open-source license has also drawn criticism, as it does not align with more common licenses like MIT or Apache 2.0 used by other models [15] Additional Features - Alongside the open-source release, new features have been added to the Grok APP, focusing on AI video generation [19]
马斯克开源Grok-2,称“中国公司将是最强大的对手”
Mei Ri Jing Ji Xin Wen· 2025-08-24 11:08
Core Insights - Tesla CEO Elon Musk announced the open-sourcing of xAI's best model, Grok-2.5, which is actually Grok-2, and stated that Grok-3 will be open-sourced in approximately six months [1] - Musk expressed confidence that xAI will soon surpass all companies except Google, and eventually surpass Google as well [1] - Musk highlighted that Chinese companies will be the strongest competitors due to their greater access to electricity and strong capabilities in hardware development [1]
刚刚,马斯克开源Grok 2.5:中国公司才是xAI最大对手
Sou Hu Cai Jing· 2025-08-24 01:29
Core Viewpoint - Elon Musk's xAI has officially open-sourced Grok 2.5, with Grok 3 expected to be released in six months, generating significant attention in the AI community [1][2]. Group 1: Open Source Release - Grok 2.5 is now available for download on HuggingFace, consisting of 42 files totaling approximately 500GB [3][4]. - The model requires a minimum of 8 GPUs, each with over 40GB of memory, to operate effectively [4][10]. Group 2: Model Performance - Grok 2 has surpassed Claude and GPT-4 in overall Elo scores on the LMSYS leaderboard, indicating competitive performance [4]. - In various academic benchmark tests, Grok 2 has shown strong results in areas such as graduate-level scientific knowledge (GPQA), general knowledge (MMLU, MMLU-Pro), and mathematics (MATH) [7][8]. Group 3: Community Feedback - While the open-sourcing of Grok has been positively received, there are criticisms regarding the lack of clarity on model parameters and the open-source licensing terms [9]. - The community has noted that the open-source model's parameters are speculated to be around 269 billion in a MoE configuration, but this has not been explicitly confirmed by xAI [9].
三位90后,估值700亿
创业家· 2025-08-11 10:09
Core Viewpoint - Mistral AI, founded by three young graduates, is raising $1 billion in a new funding round, reaching a valuation of $10 billion, reflecting a nearly 50-fold increase in just two years [4][8]. Group 1: Company Overview - Mistral AI was established by three 90s graduates who previously worked at top AI companies and returned to France to seize the AI opportunity [8]. - The company launched its first open-source model, Mistral 7B, which outperformed competitors in several benchmarks, quickly gaining attention in the developer community [8][9]. - Mistral aims to lead the generative AI wave through open-source initiatives, contrasting with closed models from competitors like OpenAI [8][9]. Group 2: Funding and Valuation - Mistral AI completed a record seed round of $113 million shortly after its founding, achieving a valuation of over $260 million [12]. - By the end of 2023, Mistral raised $415 million in Series A funding, led by a16z, increasing its valuation to $2 billion [13]. - The company’s valuation skyrocketed to $6 billion after a $640 million Series B round, with major investors including Microsoft and Nvidia [14]. - Currently, Mistral is negotiating a $1 billion funding round, which could elevate its valuation to approximately $10 billion [14]. Group 3: Competitive Landscape - The AI landscape is becoming increasingly competitive, with the emergence of DeepSeek as a significant player, prompting Mistral to accelerate its product development and commercialization efforts [9]. - Mistral has launched several products, including the chatbot Le Chat, which achieved high download rates in France but struggled internationally [9]. - The company is actively pursuing partnerships with industry giants like Nvidia to enhance its market position [9]. Group 4: Young Entrepreneurs in AI - The AI sector is witnessing a surge of young entrepreneurs, with several companies founded by 90s graduates achieving significant funding and rapid growth [16][17]. - Companies like Perplexity and Genesis AI have also seen remarkable valuations, highlighting the trend of young innovators in the AI space [16][17]. - This new generation of entrepreneurs is characterized by their global perspective and technical expertise, positioning them well to capitalize on AI opportunities [18].
OpenAI时隔6年再度开源,两款推理模型,o4-mini级,手机和笔记本能跑
3 6 Ke· 2025-08-06 03:23
Core Insights - OpenAI has released two long-anticipated open-source models: gpt-oss-120b and gpt-oss-20b, both utilizing the MoE architecture for efficient deployment [1][2][3] - The gpt-oss-120b model can run efficiently on a single 80GB GPU, while the gpt-oss-20b requires only 16GB of memory for edge devices, providing local model options for AI applications [1][2][3] - The models have shown competitive performance in benchmark tests, with gpt-oss-120b performing similarly to OpenAI's o4-mini and gpt-oss-20b comparable to o3-mini [1][2][3] Model Specifications - gpt-oss-120b has 117 billion total parameters, activating 5.1 billion parameters per token, while gpt-oss-20b has 21 billion total parameters with 3.6 billion active parameters per token [29][30] - Both models support a context length of up to 128k tokens and utilize advanced attention mechanisms to enhance efficiency [29][30] Performance and Compatibility - The gpt-oss-120b model has achieved a record inference speed of over 3000 tokens per second, while gpt-oss-20b can run on mobile devices, although some experts question the feasibility of this claim [10][45][22] - At least 14 deployment platforms, including Azure and Hugging Face, have already integrated support for these models, indicating strong industry adoption [9][10] Community and Industry Response - While many users celebrate the release, there are concerns regarding the lack of transparency in the training process and data sources, limiting the open-source community's ability to fully leverage the models [9][27][29] - OpenAI's decision to open-source these models is seen as a response to previous criticisms regarding its openness, potentially influencing more developers and companies to adopt these technologies [47]