人工智能模型

Search documents
9月30日国际晨讯 | 现货黄金价格升破3840美元再创新高 美国关键经济数据或延迟发布
Sou Hu Cai Jing· 2025-09-30 01:09
Market Review - Spot gold prices have surpassed $3840 per ounce, reaching a new high [6] - On September 29, US major stock indices experienced slight gains, with the Dow Jones up 0.15% at 46316.07 points, S&P 500 up 0.26% at 6661.21 points, and Nasdaq up 0.48% at 22591.15 points [6] - European stock indices also saw minor increases, with Germany's DAX up 0.02% at 23745.06 points, France's CAC40 up 0.13% at 7880.87 points, and the UK's FTSE 100 up 0.16% at 9299.84 points [6] International Macro - US President Trump met with congressional leaders on September 29 to discuss avoiding a government shutdown, with significant disagreements noted by Senate Democratic leader Chuck Schumer [7] - The US federal government is set to run out of funding by midnight on September 30, risking a government shutdown if no agreement is reached on funding legislation [7] - The Bureau of Labor Statistics (BLS) has announced that it will halt data collection and not release planned reports, including the monthly non-farm payroll report, if funding is interrupted [7] Corporate News - On September 29, DeepSeek-V3.2-Exp model was officially released and open-sourced on the Hugging Face platform, introducing a sparse attention mechanism for improved efficiency in training and inference of long texts [8] - OpenAI plans to release the new Sora 2 video generator as a standalone application, which may generate videos containing copyrighted content unless rights holders opt out [8] Institutional Views - Goldman Sachs analysts predict that global stock markets are likely to continue rising until the end of the year, supported by strong US economic performance and favorable stock valuations [9] - The team led by Christian Mueller-Glissmann has upgraded the global stock market allocation rating to "overweight" for the next three months, suggesting that stocks typically perform well in the later stages of economic slowdown with strong policy support [9]
阿里巴巴正式推出 Qwen3-Max
Mei Ri Jing Ji Xin Wen· 2025-09-24 03:06
每经AI快讯,9月24日,据"通义千问Qwen"微信公众号消息,阿里巴巴推出 Qwen3-Max —— 迄今为止 规模最大、能力最强的模型。目前,Qwen3-Max-Instruct 的预览版在 LMArena 文本排行榜上位列第 三,超越了 GPT-5-Chat。正式版本在代码能力和智能体(agent)能力方面进一步提升,在涵盖知识、 推理、编程、指令遵循、人类偏好对齐、智能体任务和多语言理解的全面基准测试中均达到业界领先水 平。 ...
DeepSeek宣布涨价!适配下一代国产芯片,概念股飙升
2 1 Shi Ji Jing Ji Bao Dao· 2025-08-21 09:36
Group 1 - DeepSeek officially announced the release of version 3.1 on August 21, featuring significant upgrades including a hybrid reasoning architecture and improved response efficiency [1] - The new version utilizes UE8M0FP8Scale parameter precision and has made substantial adjustments to the tokenizer and chat template, showing clear differences from version 3 [1] - DeepSeek has adjusted its pricing for API interface calls, with input prices increasing to 0.5 yuan per million tokens for cache hits and 4 yuan for cache misses, while output prices rose to 12 yuan per million tokens [2] Group 2 - The foundational model of DeepSeek V3.1 underwent extensive retraining, adding a total of 840 billion tokens, and both the foundational and post-training models are available on Huggingface and Modao [4] - Following the announcement, shares of Daily Interactive (300766) surged, closing at 47.98 yuan per share with a daily increase of 13.62% [4] - Daily Interactive, established in 2010, provides data intelligence products and services, and there were rumors of its ownership stake in DeepSeek through its subsidiary, although the company later clarified it does not hold any equity in DeepSeek or its associated companies [7]
您猜怎么着?Grok 4进决赛,大模型对抗赛Gemini全军覆没,马斯克「装」起来了
3 6 Ke· 2025-08-07 07:05
Group 1 - The core event is the ongoing AI chess tournament where models like Gemini 2.5 Pro, Grok 4, o3, and o4-mini are competing, with Grok 4 and o3 advancing to the finals after intense matches [2][5][31] - Grok 4 faced a challenging match against Gemini 2.5 Pro, resulting in a tie that was only resolved through a special tiebreaker, showcasing the competitive nature of the tournament [16][25][28] - o3 demonstrated exceptional performance, achieving a perfect accuracy score of 100 in one of its matches, indicating its strong reasoning capabilities [10][12] Group 2 - The tournament's structure includes initial rounds where models like o4-mini and o3 both achieved 4-0 victories, highlighting their dominance in the early stages [7][31] - The matches have been characterized by a mix of expected outcomes and surprising twists, particularly in the close contest between Grok 4 and Gemini 2.5 Pro [16][24] - The final match will feature Grok 4 against o3, with predictions favoring Gemini 2.5 Pro and Grok 4 as potential winners based on public voting [31][32]
亚马逊云科技上线Anthropic新一代Claude模型
Sou Hu Cai Jing· 2025-08-06 10:12
Claude Opus 4.1在Agentic搜索与研究、内容创作以及记忆和上下文管理方面表现出色,能够综合提炼深 刻见解、产出高质量内容并实现高效总结。Claude Opus 4.1可直接替代Opus 4。Anthropic表示,Claude Opus 4.1是其迄今为止最智能的模型,且在编码和Agent领域处于行业领先地位。另外,Opus 4.1具备出 色的长期任务处理能力和复杂问题解决能力。Claude Sonnet 4则以高效性著称,适用于各类项目。 Claude Sonnet 4在编码和推理能力上,均超越了其前代产品Claude Sonnet 3.7,并在性能与成本优化之间 实现了平衡,非常适合处理大批量应用场景。两款模型均配备20万token的上下文窗口,能让用户处理 和生成大篇幅内容(如文档分析和研究工作),且始终保持内容质量与连贯性。 (全球TMT2025年8月6日讯)亚马逊云科技宣布,Anthropic最新一代模型Claude Opus 4.1与Claude Sonnet 4,现已在Amazon Bedrock全面上线。这批新型双模推理模型支持在即时响应与"深度思考"模式 间动态切换,能够在 ...
阿里通义千问推出新模型Qwen3-30B-A3B-Thinking-2507
news flash· 2025-07-30 23:30
Core Insights - The article introduces a new reasoning model named Qwen3-30B-A3B-Thinking-2507, which is described as more intelligent, agile, and versatile compared to its predecessor Qwen3-30-A3B released on April 29 [1] - The new model shows significant improvements in reasoning capabilities, general capabilities, and context length [1] - Qwen3-30B-A3B-Thinking-2507 is now open-sourced on the Modao community and HuggingFace platforms [1]
Qwen全面升级非思考模型,3B激活、256K长文、性能直逼GPT-4o
量子位· 2025-07-30 09:44
Core Viewpoint - The article highlights the rapid advancements and performance improvements of the Qwen3-30B-A3B-Instruct-2507 model, emphasizing its capabilities in reasoning, long text processing, and overall utility compared to previous models [2][4][7]. Model Performance Enhancements - The new model Qwen3-30B-A3B-Instruct-2507 shows significant improvements in reasoning ability (AIME25) by 183.8% and capability (Arena-Hard v2) by 178.2% compared to its predecessor [4]. - The long text processing capability has been enhanced from 128K to 256K, allowing for better handling of extensive documents [4][11]. - The model demonstrates superior performance in multi-language knowledge coverage, text quality for subjective and open tasks, code generation, mathematical calculations, and tool usage [5][7]. Model Characteristics - Qwen3-30B-A3B-Instruct-2507 operates entirely in a non-thinking mode, focusing on stable output and consistency, making it suitable for complex human-machine interaction applications [7]. - The model's architecture supports a context window of 256K, enabling it to retain and understand large amounts of input information while maintaining semantic coherence [11]. Model Series Overview - The Qwen series has released multiple models in a short time, showcasing a variety of configurations and capabilities tailored for different scenarios and hardware resources [12][18]. - The naming convention of the models is straightforward, reflecting their parameters and versions, which aids in understanding their specifications [14][17]. Conclusion - The Qwen3 series is positioned as a comprehensive model matrix, catering to diverse needs from research to application, and is ready to address various demands in the AI landscape [19].
一句话克隆 ChatGPT Agent?智谱GLM-4.5首测:零配置,全功能|内有福利
歸藏的AI工具箱· 2025-07-28 15:20
Core Insights - The article discusses the release of GLM-4.5 by Zhipu, highlighting its strong performance in reasoning, coding, and agent capabilities, with a total parameter count of 335 billion and an activation parameter count of 32 billion [1] - GLM-4.5 is noted for its cost-effectiveness, priced at 0.8 yuan per million tokens for input and 2 yuan per million tokens for output, with a high-speed output rate exceeding 100 tokens per second [1] Performance and Features - GLM-4.5 demonstrates superior coding abilities, even with fewer total parameters compared to competitors, and excels in mixed reasoning, providing excellent results even with short prompts [2] - The model integrates various agent capabilities within a single API, allowing for seamless product development and the creation of a simplified ChatGPT-like agent [3][25] - It is compatible with Claude Code, enabling users to replace Claude Code models easily [5] Use Cases and Applications - The model successfully completes coding tasks without complex instructions, such as generating a Gmail page or a 3D abstract art piece, showcasing its ability to understand and execute detailed requirements [7][9] - GLM-4.5 can create comprehensive components like a calendar manager and an OKR management tool, fulfilling all specified requirements without bugs [11][13][14] - The model also generates high-fidelity e-commerce web pages, including detailed checkout processes, demonstrating its capability in UI/UX design [17][19][20] Integration and Accessibility - GLM-4.5 supports integration with various tools and APIs, including a search tool for generating dynamic web pages based on real-time data, such as event information for WAIC [27][28] - The model is available for a subscription fee of 50 yuan for unlimited usage, making it accessible for developers and non-developers alike [34] Strategic Positioning - The article emphasizes that GLM-4.5 represents a strategic advantage by integrating multiple functionalities into a single model, contrasting with competitors that have developed fragmented solutions [35][36] - This integration approach allows users to streamline their workflows, reducing the need for multiple models and simplifying the process of cross-model orchestration [36][37]
3550亿参数!智谱发布GLM-4.5模型,12项基准评测国产最佳
Xin Lang Ke Ji· 2025-07-28 14:32
Core Insights - The article discusses the launch of GLM-4.5, a new flagship model by Zhipu, designed specifically for intelligent agent applications, which is now open-sourced on Hugging Face and ModelScope platforms under the MIT License [2] - GLM-4.5 has achieved state-of-the-art (SOTA) performance in reasoning, coding, and intelligent agent capabilities, ranking third globally among all models and first among domestic and open-source models in 12 key evaluation benchmarks [2] - The model boasts higher parameter efficiency, with a total parameter count of 355 billion, which is half of DeepSeek-R1 and one-third of Kimi-K2, while achieving the best performance-to-parameter ratio in the SWE-bench Verified leaderboard [2][3] Model Architecture - The model utilizes a mixture of experts (MoE) architecture, with GLM-4.5 having a total parameter count of 355 billion and active parameters of 32 billion, while GLM-4.5-Air has 106 billion total parameters and 12 billion active parameters [3] - It is designed for complex reasoning and tool usage, as well as for immediate response in non-thinking modes [3] Pricing and Performance - The API call pricing is set at 0.8 yuan per million tokens for input and 2 yuan per million tokens for output, with a high-speed version capable of processing up to 100 tokens per second [3]
Nature报道:谷歌新模型1秒读懂DNA变异!首次统一基因组全任务,性能碾压现有模型
量子位· 2025-06-26 14:11
Core Viewpoint - Google DeepMind has introduced a groundbreaking biological model, AlphaGenome, which can accurately predict genomic sequence variations in just one second, marking a significant advancement in the field of genomics [3][2]. Group 1: Model Capabilities - AlphaGenome can predict thousands of functional genomic features from DNA sequences up to 1 million base pairs long, assessing variation effects with single-base resolution [4][5]. - The model outperforms existing models across various tasks, providing a powerful tool for deciphering genomic regulatory codes [5][8]. - It is described as a milestone in biology, being the first unified model that integrates a wide range of genomic tasks with high accuracy and performance [7][10]. Group 2: Model Architecture - The architecture of AlphaGenome is inspired by U-Net, processing 1 million base pairs of DNA input sequences through downsampling to generate two types of sequence representations [13]. - It employs convolutional layers for local sequence pattern modeling and Transformer blocks for modeling longer-range dependencies, achieving high-resolution training of complete base pairs [13]. - The model outputs 11 modalities, covering 5,930 human or 1,128 mouse genomic tracks, demonstrating its comprehensive predictive capabilities [13]. Group 3: Training and Performance - AlphaGenome is trained through a two-phase process involving pre-training and distillation, achieving inference times under one second on NVIDIA H100 GPUs [15][16]. - In evaluations across 24 genomic tracks, AlphaGenome maintained a leading position in 22 tasks, showing a 17.4% relative improvement in cell-type-specific LFC predictions compared to existing models [19]. - The model achieved significant enhancements in various tasks, such as a 25.5% improvement in expression QTL direction predictions compared to Borzoi3 [21]. Group 4: Clinical Applications - AlphaGenome can aid researchers in understanding the underlying causes of diseases and discovering new therapeutic targets, exemplified by its application in T-cell acute lymphoblastic leukemia research [29]. - The model's capabilities extend to predicting synthetic DNA designs and assisting in fundamental DNA research, with potential for broader species coverage and improved prediction accuracy in the future [29]. Group 5: Availability - A preview version of AlphaGenome is currently available, with plans for a formal release, inviting users to experience its capabilities [30].