Workflow
大语言模型(LLM)
icon
Search documents
香港证监会原主席梁定邦:智能金融不“唯大模型”论 监管需严保数据真实与风险可控
Xin Lang Cai Jing· 2025-12-20 10:02
Core Insights - The former chairman of the Hong Kong Securities and Futures Commission, Liang Ding-bong, discussed the development of smart finance and artificial intelligence in the Hong Kong-Macau region at the Shenzhen Xiangmi Lake Financial Annual Conference [5][7]. Group 1: Smart Finance Coverage - Smart finance in the Hong Kong-Macau region encompasses five areas: banking, securities, insurance, cross-border finance, and electronic payments [3][7]. - Hong Kong's approach to integrating artificial intelligence into traditional finance involves a multi-layered and multi-architecture technology fusion strategy, rather than solely relying on large language models (LLMs) [3][7]. Group 2: Regulatory Perspective - From a regulatory standpoint, "big data" remains the foundation of fintech applications in Hong Kong, with "large models" being just one component [3][7]. - Since 2019, Hong Kong has incorporated various technologies such as big data analysis, expert systems, and machine learning into its regulatory framework, prioritizing verifiable and traceable underlying real data in core business operations [3][7]. Group 3: Caution in AI Application - Liang Ding-bong cautioned about the "hallucination" risks associated with large models, emphasizing the need for a prudent approach to AI in financial regulation and business scenarios [3][7]. - The application of generative artificial intelligence in front-office customer interaction remains cautious, primarily focusing on back-office risk management and data analysis support roles [3][7]. - Final decision-making should involve risk management committees and risk officers, combining personal experience with multi-dimensional data, rather than relying solely on model outputs [3][7]. Group 4: Commitment to Data Integrity - Hong Kong maintains a highly open attitude towards the development of smart finance, but emphasizes the need to ensure data authenticity and risk control in client-facing and core business areas to guarantee the safety and stability of the financial system [3][7].
Karpathy 2025 年度盘点:o3 是真正拐点,Cursor 证明了应用层比我们想象的要厚
Founder Park· 2025-12-20 08:59
Core Insights - The article emphasizes that 2025 is an exciting year for Large Language Models (LLMs), highlighting their potential and the ongoing evolution in the field [2][3]. - It suggests that the industry has yet to realize even 10% of its potential, indicating vast opportunities for exploration and innovation [4][5]. Paradigm Shifts - The introduction of Reinforcement Learning from Verifiable Rewards (RLVR) is identified as a significant shift in LLM training, expected to become a primary component by 2025 [12]. - RLVR allows LLMs to train in environments where answers can be automatically verified, leading to improved problem-solving capabilities [14][16]. - The article notes that the performance improvements in 2025 will primarily stem from the adoption of RLVR, rather than an increase in model parameters [17]. New Applications - Cursor is highlighted as a new application layer product that demonstrates the potential for LLMs to be tailored for specific verticals, sparking discussions about the future of LLM applications [28][30]. - Claude Code is presented as a groundbreaking product that showcases LLM capabilities in a local environment, emphasizing the shift from cloud-based to local AI applications [34][36]. - Vibe Coding is introduced as a transformative concept that democratizes programming, allowing anyone to create software using natural language [38][40]. Future Models - The Gemini Nano Banana model is described as one of the most significant models of 2025, hinting at the future of LLMs and their integration with graphical user interfaces [42][46]. - The article suggests that LLMs should communicate in preferred formats such as images and visualizations, rather than just text, to enhance user interaction [44].
AI智能体时代中的记忆:形式、功能与动态综述
Xin Lang Cai Jing· 2025-12-17 04:42
Core Insights - Memory is identified as a core capability for agents based on foundational models, facilitating long-term reasoning, continuous adaptation, and effective interaction with complex environments [1][11][15] - The field of agent memory research is rapidly expanding but is becoming increasingly fragmented, with significant differences in motivation, implementation, assumptions, and evaluation schemes [1][11][16] - Traditional classifications of memory, such as long-term and short-term memory, are insufficient to capture the diversity and dynamics of contemporary agent memory systems [1][11][16] Summary by Sections Introduction - Over the past two years, powerful large language models (LLMs) have evolved into robust AI agents, achieving significant progress across various fields such as deep research, software engineering, and scientific discovery [4][14] - There is a growing consensus in academia that agents require capabilities beyond just LLMs, including reasoning, planning, perception, memory, and tool usage [4][14][15] Importance of Memory - Memory is crucial for transforming static LLMs into adaptive agents capable of continuous adaptation through environmental interaction [5][15] - Various applications, including personalized chatbots, recommendation systems, social simulations, and financial investigations, depend on agents' ability to manage historical information actively [5][15] Need for New Classification - The increasing importance of agent memory systems necessitates a new perspective on contemporary agent memory research [6][16] - Existing classification systems are outdated and do not reflect the breadth and complexity of current research, highlighting the need for a coherent classification that unifies emerging concepts [6][16] Framework and Key Questions - The review aims to establish a systematic framework to reconcile existing definitions and connect emerging trends in agent memory [19] - Key questions addressed include the definition of agent memory, its relationship with related concepts, its forms, functions, and dynamics, as well as emerging research frontiers [19] Emerging Research Directions - The review identifies several promising research directions, including automated memory design, integration of reinforcement learning with memory systems, multimodal memory, shared memory in multi-agent systems, and issues of trustworthiness [20][12] Contributions of the Review - The review proposes a multidimensional classification of agent memory from a "form-function-dynamics" perspective, providing a structured view of current developments in the field [20] - It explores the applicability and interaction of different memory forms and functions, offering insights on aligning various memory types with different agent objectives [20] - A comprehensive resource collection, including benchmark tests and open-source frameworks, is compiled to support further exploration of agent memory systems [20]
AI赋能资产配置(三十二):AI如何赋能财经信息“聚合提纯”?
Guoxin Securities· 2025-12-13 13:02
证券研究报告 | 2025年12月13日 AI 赋能资产配置(三十二) AI 如何赋能财经信息"聚合提纯"? 策略研究·策略解读 | 证券分析师: | 王开 | 021-60933132 | wangkai8@guosen.com.cn | 执证编码:S0980521030001 | | --- | --- | --- | --- | --- | | 证券分析师: | 陈凯畅 | 021-60375429 | chenkaichang@guosen.com.cn | 执证编码:S0980523090002 | 事项: ①AI 大模型在金融信息处理领域的应用持续深化,为解决信息过载、分析成本高的行业痛点提供了技术支 撑。基于 LLM 的自动化财经情报工具 Wide-Research-for-Finance 通过整合多源数据与智能分析能力,为个 人投资者及小型研究团队提供了轻量化、高性价比的信息解决方案。②该工具以两阶段处理机制为核心, 先通过标题快速筛选每小时采集的 200+条新闻,再依托 DeepSeek 大模型完成情绪识别、实体提取、事件 分类与影响评估,同步生成结构化报告,支持本地部署与自定义数据源扩展, ...
LLM距离AGI只差一层:斯坦福研究颠覆「模式匹配」观点
机器之心· 2025-12-10 10:30
机器之心报道 编辑:杨文、泽南 有关大语言模型的理论基础,可能要出现一些改变了。 斯坦福发了篇论文,彻底颠覆了「LLM 只是模式匹配器」的传统论调。 它提出的不是扩展技巧或新架构,而是一个让模型真正具备推理能力的「协调层」。 核心观点:AGI 的瓶颈在于协调,而非规模 人工智能界正因围绕大语言模型本质的争论而分裂。一方面,扩展派认为 LLMs 足以实现 AGI;另一方 面,有影响力的批评者认为 LLM「仅仅是模式匹配器」,在结构上不具备推理、规划或组合泛化能力,因 此是死胡同。 作者认为这场争论建立在一个错误的二分法之上,并提出一个颠覆性极强的核心观点: LLM 的失败不是因 为缺乏推理能力,而是因为我们缺少将其模式与目标绑定的系统。 为了解释这一点,作者用了一个捕鱼隐喻。 海洋代表模型庞大的模式库,渔夫不用鱼饵就撒网,收获的只是最常见的鱼类(训练数据中的通用模 式)。批评者谴责这些未锚定的输出,但他们观察到的只是未加诱饵的捕捞所产生的原始统计基线,这不 是系统损坏,而是系统在默认模式下的自然表现。 然而,智能行为不仅仅是撒网,它还涉及下饵和过滤。如果诱饵过于稀疏,它就无法吸引特定、稀有的 鱼,海洋的先验仍然 ...
深大团队让机器人听懂指令精准导航,成功率可达72.5%,推理效率提升40%
3 6 Ke· 2025-12-10 07:00
Core Insights - The article discusses the introduction of a new framework called UNeMo for visual-language navigation (VLN), developed by a team led by Professor Li Jianqiang from Shenzhen University in collaboration with Beijing Institute of Technology and Moscow University [1][3]. Group 1: Framework Overview - UNeMo utilizes a multi-modal world model (MWM) and a hierarchical predictive feedback navigator (HPFN) to enhance the decision-making capabilities of navigation agents by allowing them to predict future visual states based on current visual features and language instructions [6][12]. - The framework addresses the disconnection between language reasoning and visual navigation, which has been a significant challenge in embodied AI [6][8]. Group 2: Performance Metrics - UNeMo achieves a navigation success rate of 72.5% in unseen environments, outperforming the existing method NavGPT2, which has a success rate of 71% [15]. - The framework demonstrates a significant reduction in resource consumption, with GPU memory usage dropping from 27GB to 12GB (a 56% reduction) and inference speed improving by 40% [15]. Group 3: Experimental Validation - In experiments on the R2R dataset, UNeMo shows a balance between lightweight configuration and high-performance decision-making, achieving a path efficiency (SPL) improvement from 60% to 61.3% [15]. - UNeMo exhibits a notable advantage in long-path navigation, with a success rate increase of 5.6% for paths longer than 7 units, compared to a mere 1.2% increase for shorter paths [17]. Group 4: Scalability and Adaptability - The framework has been tested across various navigation baselines and datasets, demonstrating its adaptability and scalability beyond LLM-based systems [20]. - UNeMo's collaborative training architecture allows it to effectively apply to different types of navigation tasks, enhancing its overall utility in practical applications [20].
深大团队让机器人听懂指令精准导航!成功率可达72.5%,推理效率提升40%|AAAI2026
Xin Lang Cai Jing· 2025-12-10 06:52
Core Insights - The UNeMo framework, developed by a team led by Professor Li Jianqiang from Shenzhen University, aims to enhance visual-language navigation (VLN) for robots, allowing them to understand commands and navigate accurately in unknown environments [1][18]. Group 1: Framework and Mechanism - UNeMo utilizes a dual-module architecture combining a Multi-modal World Model (MWM) and a Hierarchical Predictive Feedback Navigator (HPFN) to address the disconnection between visual reasoning and navigation decision-making [5][20]. - The MWM predicts future visual states based on current visual features, language instructions, and potential navigation actions, overcoming limitations of existing methods that only focus on the present [21][22]. - The HPFN employs a two-stage hierarchical mechanism to generate coarse-grained candidate actions and refine them based on MWM predictions, ensuring robust navigation in complex environments [24][26]. Group 2: Performance and Efficiency - UNeMo demonstrates significant improvements in resource efficiency, with GPU memory usage reduced by 56% (from 27GB to 12GB) and inference speed increased by 40% (from 1.1 seconds to 0.7 seconds) compared to mainstream methods [27][28]. - In unseen test environments, UNeMo achieves a navigation success rate (SR) of 72.5%, surpassing NavGPT2's 71% by 1.5 percentage points, and improves path efficiency (SPL) from 60% to 61.3% [28][30]. Group 3: Robustness and Scalability - UNeMo shows a marked advantage in long-path navigation, with SR increasing by 5.6% for paths longer than 7 units, compared to a mere 1.2% increase for shorter paths [30][31]. - The framework's adaptability is validated across various navigation baselines and datasets, proving its scalability beyond LLM-based systems [32][33].
谷歌IMO金牌级Gemini 3深夜上线,华人大神挂帅,OpenAI无力反击
3 6 Ke· 2025-12-05 10:08
Core Insights - Google DeepMind has launched its latest model, Gemini 3 Deep Think, which is touted as the strongest model in the IMO competition history [1][7]. - Gemini 3 Deep Think has demonstrated significant advancements in reasoning capabilities, particularly in solving complex mathematical and scientific problems [7][14]. Performance Metrics - Gemini 3 Deep Think achieved a score of 41% in the Humanity's Last Exam without tool assistance and a record 45.1% in ARC-AGI-2 with code execution [2][7]. - In benchmark tests, Gemini 3 Deep Think outperformed its predecessor, Gemini 3 Pro, showcasing superior performance in various tasks [2][10]. Features and Capabilities - The model utilizes "parallel reasoning" to explore multiple hypotheses simultaneously, enhancing its problem-solving abilities [11][14]. - Gemini 3 Deep Think has been integrated into the Gemini App, available exclusively for Ultra subscribers [5][11]. Team Development - Google DeepMind announced the formation of a new elite team in Singapore, led by scientist Yi Tay, focusing on advanced reasoning and the development of cutting-edge models like Gemini and Gemini Deep Think [18][23]. - The team aims to gather top talent in the AI field, emphasizing the importance of "talent density" in the era of large language models [21][23]. Market Impact - Following the release of Gemini 3 Pro, its web market share has surpassed 15%, indicating a strong competitive position against ChatGPT, which has seen a decline in market share [26][30]. - Gemini's traffic reached 1.351 billion visits, a 14.3% increase from October, while ChatGPT's traffic fell to 5.844 billion visits [30][34].
元保发布第三季度财报:营收达11.58亿元,AI驱动业务成效显著
Ge Long Hui· 2025-12-03 10:07
Group 1: Financial Performance - In Q3 2025, the company reported total revenue of RMB 1.158 billion, representing a year-on-year growth of 33.6% [1] - The net profit for the same period reached RMB 370 million, showing a year-on-year increase of 51.3% [1] - As of September 30, 2025, the company's cash reserves stood at RMB 3.75 billion, indicating a solid financial position [1] Group 2: Technological Advancements - The company's model library expanded to over 4,900 models and 5,500 features, with approximately 400 new models and 750 new features added year-on-year [1] - The introduction of large language models (LLM) has significantly enhanced operational efficiency, with AI-generated code accounting for nearly 50% of the coding process in Q3 [1] - The AI team comprises over 10% of the total workforce, reflecting the company's commitment to technology [1] Group 3: Market Trends and Product Development - The establishment of a national "medical insurance + commercial insurance" settlement center has integrated commercial insurance into China's multi-tiered medical security system [2] - The company is focusing on the inclusive health insurance sector, launching a short-term critical illness insurance product that combines a "one-time payment + multiple reimbursements" model [2] - The customer service center has introduced a "Five Hearts Service" standard to enhance the user experience throughout the entire process [2]
2025年AI智能体在未来产业创新上的前沿应用与发展趋势报告(1)
Sou Hu Cai Jing· 2025-12-02 21:04
Core Insights - The report outlines the evolution of AI from large language models (LLMs) to Agentic AI, emphasizing a shift towards a closed-loop system of perception, decision-making, action, and learning [1][6] - The global Agentic AI market is projected to grow from approximately $5.29 billion in 2024 to $46-47 billion by 2030, with a compound annual growth rate (CAGR) exceeding 40% [15] - Key industry applications include finance, healthcare, education, manufacturing, and collaborative office environments, with a significant transformation expected in organizational operations and employment structures by 2028 [25][28] Industry Trends - The transition from model intelligence to behavioral intelligence marks a significant macro trend in the AI industry, moving towards a focus on closed-loop systems [6] - The report identifies five major evolutionary trends in Agentic AI, including a shift from application-driven to ecosystem-driven models and from single-agent to multi-agent collaboration [29] - The anticipated inflection point for large-scale application of AI agents is 2025, with expectations that 33% of enterprise software will integrate AI agent functionalities by 2028 [23] Market Dynamics - North America is identified as the primary funding pool for Agentic AI, while Europe focuses on privacy compliance and efficiency tools, and China leans towards outbound application services [15] - The report highlights the emergence of ten innovative solutions in Agentic AI technology, including retrieval-augmented generation (RAG) and multi-agent collaboration [30][32] - The expected impact of Agentic AI on traditional industries includes a 40% reduction in operational costs and a 20% increase in revenue by 2028 [25] Employment and Skills - The rise of AI agents is expected to lead to job displacement in repetitive and rule-based roles, while simultaneously creating new positions in AI development, training, and maintenance [28] - There will be a shift in skill requirements, with increased demand for creativity, strategic thinking, and emotional intelligence [28] Technological Innovations - Future breakthroughs in Agentic AI are anticipated in areas such as multi-modal integration, enhanced autonomous decision-making, and improved collaboration capabilities among multiple agents [38] - The report emphasizes the importance of safety and risk governance, proposing strategies for reliability, compliance, and ethical considerations in AI deployment [10][12]