Workflow
Ministral 3
icon
Search documents
AI三国杀:OpenAI狂卷,DeepSeek封神,却被Mistral偷了家?
3 6 Ke· 2025-12-03 11:55
Core Insights - Mistral has launched two significant products: the Mistral Large 3 model and the Ministral 3 series, both of which are open-source, multimodal, and designed for practical applications [1][3]. Mistral Large 3 - Mistral Large 3 features a MoE architecture with 41 billion active parameters and 675 billion total parameters, showcasing advanced image understanding and multilingual capabilities, ranking 6th among open-source models [3][6]. - It has achieved a high ELO score, placing it in the top tier of open-source models, comparable to Kimi K2 and slightly behind DeepSeek v3.2 [6][10]. - The model performs on par with larger models like DeepSeek 37B and Kimi K2 127B across various foundational tasks, indicating its competitive strength [8][10]. - Mistral has partnered with NVIDIA to enhance the model's stability and performance by optimizing the underlying inference pathways, making it faster and more cost-effective [10][16]. Ministral 3 Series - The Ministral 3 series includes models of 3B, 8B, and 14B sizes, all capable of running on various devices, including laptops and drones, and optimized for performance [11][18]. - The instruct versions of the Ministral 3 models show significant improvements in performance, with scores of 31 (14B), 28 (8B), and 22 (3B), surpassing the previous generation [11][29]. - The 14B version of Ministral has demonstrated superior performance in reasoning tasks, outperforming competitors like Qwen 14B in multiple benchmarks [25][28]. Strategic Positioning - Mistral aims to address enterprise needs by providing customizable AI solutions that are cost-effective and reliable, contrasting with the high costs associated with proprietary models from competitors like OpenAI and Google [29][33]. - The company is evolving into a platform that not only offers models but also integrates various functionalities such as code execution and structured reasoning through its Mistral Agents API [33][37]. - Mistral's approach reflects a shift towards a more decentralized AI model, emphasizing accessibility and usability across different devices and environments, which could reshape the global AI landscape [37][39].
刚刚,「欧洲的DeepSeek」发布Mistral 3系列模型,全线回归Apache 2.0
机器之心· 2025-12-03 00:06
Core Viewpoint - Mistral AI has launched the Mistral 3 series of open models, which are positioned as high-performance, cost-effective alternatives in the AI model landscape, particularly in response to competition from DeepSeek [2][4][28]. Model Details - The Mistral 3 series includes multiple models: Mistral 3 (14B, 8B, 3B) with base, instruction-tuned, and reasoning versions [5][19]. - Mistral Large 3, a state-of-the-art open model, features a total parameter count of 675 billion and 41 billion active parameters, trained on 3000 NVIDIA H200 GPUs [7][5]. Performance and Benchmarking - Mistral Large 3 ranks second in the OSS non-inference model category on the LMArena leaderboard, indicating it is one of the best-performing open models available [14]. - The model demonstrates strong performance in general prompt tasks and excels in image understanding and multilingual dialogue [7][14]. Collaboration and Optimization - Mistral has partnered with vLLM and Red Hat to enhance accessibility and efficiency for developers using Mistral Large 3, utilizing optimized checkpoints for better performance [17][18]. - The collaboration with NVIDIA focuses on advanced optimization techniques, ensuring that Mistral models leverage high-bandwidth memory for demanding workloads [17][18]. Cost-Effectiveness - Mistral claims that its models offer the best cost-performance ratio among open-source models, with instruction models performing comparably or better than competitors while generating tokens at a significantly lower rate [22][28]. Availability and Customization - Mistral 3 models are available on various platforms including Mistral AI Studio, Amazon Bedrock, and Azure Foundry, among others [25]. - The company also offers custom model training services to organizations seeking tailored AI solutions for specific tasks or environments [27].
英伟达官宣新合作成就:Mistral开源模型提速,任意规模均提高效率和精度
Hua Er Jie Jian Wen· 2025-12-02 20:03
Core Insights - Nvidia has announced a significant breakthrough in collaboration with French AI startup Mistral AI, achieving substantial improvements in performance, efficiency, and deployment flexibility through the use of Nvidia's latest chip technology [1] - The Mistral Large 3 model has achieved a tenfold performance increase compared to the previous H200 chip, translating to better user experience, lower response costs, and higher energy efficiency [1][2] - Mistral AI's new model family includes a large frontier model and nine smaller models, marking a new phase in open-source AI and bridging the gap between research breakthroughs and practical applications [1][6] Performance Breakthrough - Mistral Large 3 is a mixture of experts (MoE) model with 67.5 billion total parameters and 41 billion active parameters, featuring a context window of 256,000 tokens [2] - The model utilizes Wide Expert Parallelism, NVFP4 low-precision inference, and the Dynamo distributed inference framework to achieve best-in-class performance on Nvidia's GB200 NVL72 system [4] Model Compatibility and Deployment - The Mistral Large 3 model is compatible with major inference frameworks such as TensorRT-LLM, SGLang, and vLLM, allowing developers to deploy the model flexibly across various Nvidia GPUs [5] - The Ministral 3 series includes nine high-performance models optimized for edge devices, supporting visual functions and multi-language capabilities [6] Commercialization Efforts - Mistral AI is accelerating its commercialization efforts, having secured agreements with major companies, including HSBC, for model access in various applications [7] - The company has signed contracts worth hundreds of millions of dollars and is collaborating on projects in robotics and AI with organizations like the Singapore Ministry of Home Affairs and Stellantis [7] Accessibility of Models - Mistral Large 3 and Ministral-14B-Instruct are now available to developers through Nvidia's API directory and preview API, with all models accessible for download from Hugging Face [8]