Workflow
人工智能开源
icon
Search documents
LeCun创业0产品估值247亿,回应谢赛宁入伙
量子位· 2026-01-23 07:44
Group 1 - The core viewpoint of the article is that Yann LeCun, after leaving Meta, is launching a new company called Advanced Machine Intelligence (AMI), focusing on world models rather than large language models (LLMs) for achieving human-level intelligence [9][17][20] - LeCun criticizes Meta's product development decisions, stating that while research is acceptable, product execution has been poor, particularly under Mark Zuckerberg's leadership [2][3][15] - AMI aims to be an open-source platform, contrasting with the recent trend in Silicon Valley towards closed-source models, which LeCun believes is a misguided approach [11][13][16] Group 2 - The company will initially focus on research and development, specifically on world models, which LeCun argues are essential for building intelligent systems [17][19] - LeCun emphasizes that LLMs are not equivalent to AI and that understanding the real world is crucial for achieving human-like intelligence, which LLMs struggle to do [21][22][23] - AMI is seeking to raise €30 million (approximately 247 billion RMB) in funding, with an initial goal of €3.5 million for early financing, aiming for a total of €5 million in the first round [45][46][50] Group 3 - The company has already attracted interest from potential investors, including Cathay Innovation and Hiro Capital, indicating a shift in venture capital investment logic towards valuing founders over products [52][53][54] - LeCun is actively recruiting talent, including former Meta executives, to strengthen AMI's capabilities [40][42] - The ultimate goal of AMI is to become a leading supplier of intelligent systems, with a focus on practical applications of world models and planning capabilities [38][39]
上海发文明确:到2027年,打造1-2个具有国际影响力的开源社区
Yang Shi Wang· 2025-12-25 08:19
12月25日,智通财经记者从上海市经信委获悉,《上海市加强开源体系建设实施方案》(以下简称实施方 案)于近日正式印发。 《上海市加强开源体系建设实施方案》主要目标是加快推进自主开源体系建设,释放开源发展潜能,将上 海建设成为服务全国、具有国际竞争力的开源创新策源地和发展高地。到2027年,上海初步建成技术支撑 齐全、服务体系完善、人才结构合理、开源产业繁荣的开源体系,打造1-2个具有国际影响力的开源社区, 培育100家开源商业化企业,孵化200个以上优质开源项目,集聚开源开发者超300万人。到2030年,开源 在引领产业创新发展、提升产业链供应链韧性方面作用凸显,主动建设开源、使用开源、贡献开源的氛围 基本形成。 什么是开源体系建设? 开源体系是以"开放共享、众智众享"为理念,涵 盖技术、社区、治理、应用的一体化生态系统。 上海建设开源体系以"引领产业发展"为根本遵 循,通过统筹技术、人才、企业等核心要素,打 造技术创新策源、全球开发者集聚、龙头企业培 育的可持续开源创新生态。 8 深度思考 ← 01 主要目标 用3-5年时间,将上海建设成为服务全国、具有 国际竞争力的开源创新策源地和发展高地,打 造全球开源 ...
昔日开源霸主承认蒸馏阿里千问,世界进入中国AI时间
3 6 Ke· 2025-12-11 11:39
全球人工智能开源领域出现标志性事件。12月10日,彭博社的一则报道震动了硅谷——曾经的全球开源霸主Meta公司,其最新AI模型"牛油果"项目正在秘 密使用中国阿里巴巴的千问Qwen开源模型进行蒸馏训练。 这一消息传出后,阿里巴巴股价盘前应声上涨超过4%。 就在一年前,Meta创始人扎克伯格还在公开场合呼吁构建以美国模型为核心的开源生态。如今,面对自身Llama系列增长乏力与东方模型的强势崛起, Meta的这次选择更像一次务实的转身。 开源世界的权力转移早有征兆。2024年8月,Qwen的衍生模型数量首次超越长期霸主Llama;2025年10月,其全球下载量也完成了反超。英伟达CEO黄仁 勋在近期一场演讲中的观察点明了这一趋势:"开源已变得极其重要,而中国在这一领域正展现出领先态势。" 当昔日的规则制定者开始向曾经的追赶者汲取养分,一场深刻的行业变局已经拉开序幕。 一场无声的倒戈:开源王座易主 根据彭博社报道,扎克伯格密切关注的TBD实验室团队,在"牛油果"模型的训练中,蒸馏了包括谷歌Gemma、OpenAI的gpt-oss以及阿里巴巴Qwen在内的 多方开源模型。 这一技术选择背后是Meta自身开源战略面临的 ...
中国模型厂商开辟“开源战场”,顶层设计再添一把火
Di Yi Cai Jing· 2025-08-29 06:58
Core Insights - Open source is not merely a technical means but is becoming a key mechanism to drive the AI ecosystem and industry implementation [1][5] Group 1: Government Initiatives - The State Council has issued opinions to deepen the implementation of the "AI+" action, emphasizing the enhancement of foundational model capabilities and the promotion of a thriving open-source ecosystem [2] - The development of a robust open-source community is seen as essential for gathering global developers and forming a vast network for technical exchange and innovation, which can enhance the influence of Chinese models on the world stage [2] Group 2: Industry Performance - Chinese model vendors are experiencing a dual scenario of "open source for a leapfrog advantage" and "technical breakthroughs facing bottlenecks," achieving significant rankings on international lists while still confronting challenges in commercializing their innovations [2][4] - Chinese open-source models, such as those from Alibaba and DeepSeek, are now comparable in performance to top proprietary models, marking a significant advancement in the global AI landscape [4] Group 3: Competitive Landscape - The competitive landscape is characterized by a dichotomy where leading firms either choose not to open-source or only open-source non-core models, with proprietary models used to maintain competitive advantages [8] - Despite the advantages of open-source models, such as customization and cost savings, there remains a performance gap of 9 to 12 months compared to leading proprietary models, influencing enterprise preferences [8] Group 4: Future Outlook - Open source is expected to evolve beyond a technical tool to become a critical mechanism for driving AI ecosystem and industry implementation, facilitating innovation and collaboration [5][9] - The relationship between open-source and proprietary models is seen as complementary, with open-source potentially leading breakthroughs in industry customization and multi-modal innovation [9] Group 5: Commercialization Challenges - Chinese open-source projects face challenges in commercializing effectively compared to their overseas counterparts, primarily due to different market environments and user habits [9] - The path to overcoming these challenges lies in globalization, where establishing a developer base in international communities can enhance customer willingness to pay and brand recognition [9]
马斯克开源Grok 2.5:中国公司才是xAI最大对手
3 6 Ke· 2025-08-24 23:25
Core Points - xAI has officially open-sourced Grok 2.5, with Grok 3 expected to be released in six months [1] - Elon Musk had previously indicated that it was time to open-source Grok, which was initially expected to happen the following week [3] - Despite the delay, the sentiment remains positive as many believe that it is better late than never [4] Summary by Sections Open Source Release - Grok can now be downloaded from HuggingFace, consisting of 42 files with a total size of approximately 500GB [5] - The official recommendation is to use SGLang to run Grok 2, with specific steps provided for downloading and setting up the model [6][7] Technical Specifications - The model requires 8 GPUs, each with over 40GB of memory, to operate effectively [8][17] - The initial setup involves downloading weight files and launching the inference server using SGLang [8] Performance Metrics - Grok 2 has shown competitive performance in various academic benchmarks, including GPQA, MMLU, and MATH, with scores that rival leading models [13] - In the LMSYS ranking, Grok 2 surpassed Claude and GPT-4 in overall Elo scores [10] Community Feedback - There are mixed reactions regarding the open-source model, particularly concerning the lack of clarity on the model's parameter weights, which are speculated to be 269 billion for the MoE model [15] - The open-source license has also drawn criticism, as it does not align with more common licenses like MIT or Apache 2.0 used by other models [15] Additional Features - Alongside the open-source release, new features have been added to the Grok APP, focusing on AI video generation [19]
马斯克开源Grok-2,称“中国公司将是最强大的对手”
Mei Ri Jing Ji Xin Wen· 2025-08-24 11:08
Core Insights - Tesla CEO Elon Musk announced the open-sourcing of xAI's best model, Grok-2.5, which is actually Grok-2, and stated that Grok-3 will be open-sourced in approximately six months [1] - Musk expressed confidence that xAI will soon surpass all companies except Google, and eventually surpass Google as well [1] - Musk highlighted that Chinese companies will be the strongest competitors due to their greater access to electricity and strong capabilities in hardware development [1]
刚刚,马斯克开源Grok 2.5:中国公司才是xAI最大对手
Sou Hu Cai Jing· 2025-08-24 01:29
Core Viewpoint - Elon Musk's xAI has officially open-sourced Grok 2.5, with Grok 3 expected to be released in six months, generating significant attention in the AI community [1][2]. Group 1: Open Source Release - Grok 2.5 is now available for download on HuggingFace, consisting of 42 files totaling approximately 500GB [3][4]. - The model requires a minimum of 8 GPUs, each with over 40GB of memory, to operate effectively [4][10]. Group 2: Model Performance - Grok 2 has surpassed Claude and GPT-4 in overall Elo scores on the LMSYS leaderboard, indicating competitive performance [4]. - In various academic benchmark tests, Grok 2 has shown strong results in areas such as graduate-level scientific knowledge (GPQA), general knowledge (MMLU, MMLU-Pro), and mathematics (MATH) [7][8]. Group 3: Community Feedback - While the open-sourcing of Grok has been positively received, there are criticisms regarding the lack of clarity on model parameters and the open-source licensing terms [9]. - The community has noted that the open-source model's parameters are speculated to be around 269 billion in a MoE configuration, but this has not been explicitly confirmed by xAI [9].
三位90后,估值700亿
创业家· 2025-08-11 10:09
Core Viewpoint - Mistral AI, founded by three young graduates, is raising $1 billion in a new funding round, reaching a valuation of $10 billion, reflecting a nearly 50-fold increase in just two years [4][8]. Group 1: Company Overview - Mistral AI was established by three 90s graduates who previously worked at top AI companies and returned to France to seize the AI opportunity [8]. - The company launched its first open-source model, Mistral 7B, which outperformed competitors in several benchmarks, quickly gaining attention in the developer community [8][9]. - Mistral aims to lead the generative AI wave through open-source initiatives, contrasting with closed models from competitors like OpenAI [8][9]. Group 2: Funding and Valuation - Mistral AI completed a record seed round of $113 million shortly after its founding, achieving a valuation of over $260 million [12]. - By the end of 2023, Mistral raised $415 million in Series A funding, led by a16z, increasing its valuation to $2 billion [13]. - The company’s valuation skyrocketed to $6 billion after a $640 million Series B round, with major investors including Microsoft and Nvidia [14]. - Currently, Mistral is negotiating a $1 billion funding round, which could elevate its valuation to approximately $10 billion [14]. Group 3: Competitive Landscape - The AI landscape is becoming increasingly competitive, with the emergence of DeepSeek as a significant player, prompting Mistral to accelerate its product development and commercialization efforts [9]. - Mistral has launched several products, including the chatbot Le Chat, which achieved high download rates in France but struggled internationally [9]. - The company is actively pursuing partnerships with industry giants like Nvidia to enhance its market position [9]. Group 4: Young Entrepreneurs in AI - The AI sector is witnessing a surge of young entrepreneurs, with several companies founded by 90s graduates achieving significant funding and rapid growth [16][17]. - Companies like Perplexity and Genesis AI have also seen remarkable valuations, highlighting the trend of young innovators in the AI space [16][17]. - This new generation of entrepreneurs is characterized by their global perspective and technical expertise, positioning them well to capitalize on AI opportunities [18].
OpenAI时隔6年再度开源,两款推理模型,o4-mini级,手机和笔记本能跑
3 6 Ke· 2025-08-06 03:23
Core Insights - OpenAI has released two long-anticipated open-source models: gpt-oss-120b and gpt-oss-20b, both utilizing the MoE architecture for efficient deployment [1][2][3] - The gpt-oss-120b model can run efficiently on a single 80GB GPU, while the gpt-oss-20b requires only 16GB of memory for edge devices, providing local model options for AI applications [1][2][3] - The models have shown competitive performance in benchmark tests, with gpt-oss-120b performing similarly to OpenAI's o4-mini and gpt-oss-20b comparable to o3-mini [1][2][3] Model Specifications - gpt-oss-120b has 117 billion total parameters, activating 5.1 billion parameters per token, while gpt-oss-20b has 21 billion total parameters with 3.6 billion active parameters per token [29][30] - Both models support a context length of up to 128k tokens and utilize advanced attention mechanisms to enhance efficiency [29][30] Performance and Compatibility - The gpt-oss-120b model has achieved a record inference speed of over 3000 tokens per second, while gpt-oss-20b can run on mobile devices, although some experts question the feasibility of this claim [10][45][22] - At least 14 deployment platforms, including Azure and Hugging Face, have already integrated support for these models, indicating strong industry adoption [9][10] Community and Industry Response - While many users celebrate the release, there are concerns regarding the lack of transparency in the training process and data sources, limiting the open-source community's ability to fully leverage the models [9][27][29] - OpenAI's decision to open-source these models is seen as a response to previous criticisms regarding its openness, potentially influencing more developers and companies to adopt these technologies [47]
扎克伯格发文正式告别“默认开源”!网友:只剩中国 DeepSeek、通义和 Mistral 还在撑场面
AI前线· 2025-08-02 05:33
Core Viewpoint - Meta is shifting its AI model release strategy to better promote the development of "personal superintelligence," emphasizing the need for careful management of associated risks and selective open-sourcing of content [3][5][11]. Group 1: Shift in Open-Source Strategy - Mark Zuckerberg's recent statements indicate a significant change in Meta's approach to open-source AI, moving from being a "radical open-source advocate" to a more cautious stance on which models to open-source [6][8]. - The company previously viewed its Llama open-source model series as a key competitive advantage against rivals like OpenAI and Google DeepMind, but this perspective is evolving [5][9]. - Meta is unlikely to open-source its most advanced models in the future, which could lead to increased expectations for companies that remain committed to open-source AI, particularly in China [10][11]. Group 2: Investment and Development Focus - Meta has committed $14.3 billion to invest in Scale AI and restructure its AI department into "Meta Superintelligence Labs," indicating a strong focus on developing closed-source models [11][12]. - The company is reallocating resources from testing the latest Llama model to concentrate on developing a closed-source model, reflecting a strategic pivot in its AI commercialization approach [12][14]. - Meta's primary revenue source remains internet advertising, allowing it to approach AI development differently than competitors reliant on selling access to AI models [11]. Group 3: Future of Personal Superintelligence - Zuckerberg envisions "personal superintelligence" as a means for individuals to achieve their personal goals through AI, with plans to integrate this concept into products like augmented reality glasses and virtual reality headsets [14]. - The company aims to create personal devices that can understand users' contexts, positioning these devices as the primary computing tools for individuals [14].