Large Language Model
Search documents
Nature认定的论文综述神器来了
量子位· 2026-02-07 04:22
Core Viewpoint - The article discusses the launch of OpenScholar, an AI system developed by the Allen Institute for AI and the University of Washington, which aims to eliminate the issue of false citations in academic writing by leveraging a vast database of 45 million scientific papers [2][5]. Group 1: OpenScholar's Features - OpenScholar connects to a large database called ScholarStore, which contains full texts and abstracts of 45 million papers, significantly reducing the false citation rate of traditional large language models (LLMs) [9][11]. - The system employs Retrieval-Augmented Generation (RAG) technology to ensure that each knowledge point is backed by a real paper, enhancing the accuracy of citations [12][13]. - OpenScholar's feedback loop allows it to refine its outputs by searching, generating, self-reviewing, and revising, which helps confirm the existence of supporting literature [12][13]. Group 2: Performance Comparison - In a benchmark test called Scholar QABench, OpenScholar-8B outperformed GPT-4o by 5% in correctness and matched human expert citation accuracy [16]. - A double-blind experiment showed that 51% of OpenScholar's answers were rated better than those written by human researchers, with an upgraded version achieving a 70% success rate [18]. - Experts noted that OpenScholar's strengths lie in its comprehensive information coverage, clearer structure, and stronger logical coherence compared to traditional models [19].
3 E Network Initiates Strategic Procurement for Mikkeli AI Data Center Project
Globenewswire· 2026-02-03 12:50
Core Insights - 3 E Network Technology Group Limited has officially initiated the procurement process for critical equipment for its AI Data Center Project in Finland, marking a transition from planning to construction preparation [1][2] Group 1: Project Overview - The AI data center is a strategic cornerstone for the company to connect digital ecosystems and deploy AI infrastructure, with a focus on high-performance computing and large language model training [2] - The procurement process aims to secure top-tier hardware resources early, aligning with the company's construction philosophy of "Green and Low-Carbon, Modular Assembly, and Extreme Energy Efficiency" [2][3] Group 2: Supply Chain and Compliance - The company has begun supply chain selection and technical validation procedures, developing a vendor qualification system focused on engineering adaptability and regulatory compliance [3] - All candidate technical solutions will undergo comprehensive compliance reviews according to Finnish national building standards and environmental permitting requirements [3] Group 3: Infrastructure Focus Areas - The company plans to procure prefabricated structural components that comply with Finnish fire and structural standards to improve construction efficiency [4] - Prioritization of prefabricated power skids and modular UPS systems is intended to support a decoupled power system design for scalable deployment [4] - Evaluation of liquid-cooling-ready coolant distribution units and air-cooling solutions aims to address thermal requirements of next-generation AI chips [4] - High-capacity optical cable systems will be evaluated to support low-latency data transmission for GPU-based computing workloads [4] - The company will prioritize the evaluation and procurement of sensor arrays and edge computing gateways to support the 3 E Intellisight™ Smart Operations Platform [4]
榜单更新!Kimi 2.5表现突出|xbench月报
红杉汇· 2026-02-03 00:04
Core Insights - The article highlights the recent updates in the xbench leaderboard, showcasing the performance of various AI models, particularly emphasizing the Kimi K2.5 model's significant improvements and its ranking among competitors [1][4][10]. Group 1: Model Performance Updates - As of January 2026, Kimi K2.5 achieved an average score of 63.2, marking a notable improvement from its predecessor K2, and ranked 4th on the leaderboard, making it the top model in China [4][5]. - The new benchmarks introduced by xbench include BabyVision for evaluating multimodal understanding and AgentIF-OneDay for assessing complex task instruction adherence [1]. - The leaderboard updates reflect the performance of mainstream large language models (LLMs) available through public APIs, with Kimi K2.5 scoring 36.5 in the BabyVision benchmark, placing it second behind Gemini 3 Pro [8][10]. Group 2: Kimi K2.5 Specifications - Kimi K2.5, released on January 27, 2026, is a next-generation multimodal model that integrates visual understanding, logical reasoning, programming, and agent capabilities [10]. - The model is based on approximately 15 trillion mixed visual and text tokens for continuous pre-training, enabling it to natively understand and process visual information [10]. - Kimi K2.5 employs a mixture of experts (MoE) architecture, with a total parameter count of around 1 trillion, activating approximately 32 billion parameters during inference to maintain high performance and efficiency [10]. Group 3: Competitive Landscape - The leaderboard indicates that Kimi K2.5 is positioned as a strong competitor in the AI model market, with its performance metrics suggesting a competitive edge in terms of cost-effectiveness and speed [4][7]. - The article notes that Kimi K2.5's inference time is significantly reduced to 2-3 minutes per question, enhancing its usability in practical applications [7].
The 1 thing You Need to Watch in Amazon's Earnings
Yahoo Finance· 2026-01-27 17:43
Key Points Amazon Web Services (AWS) is the largest cloud provider in the world. Despite its size, AWS growth reaccelerated in last year's third quarter. AWS boasted a $200 billion backlog of business at quarter's end. 10 stocks we like better than Amazon › Amazon (NASDAQ: AMZN) stock has barely moved over the past year, but come earnings time, and the stock might finally get the juice it needs to jump higher. There are many things the market is going to take note of when the company reports 2025 ...
Neusoft and Cerence AI Sign Strategic Cooperation Agreement to Deliver an AI-Powered Automotive Cockpit Platform
Prnewswire· 2026-01-22 08:30
Core Insights - Neusoft Corporation and Cerence AI have signed a Memorandum of Understanding to collaborate on large language model-based voice AI solutions for the automotive industry [1][2] - The partnership aims to enhance in-cabin user experiences by providing intelligent interaction solutions that meet rising user expectations for natural language communication [2][5] Company Overview - Neusoft Corporation is a leading information technology company founded in 1991, recognized as the first listed software company in China, with a focus on intelligent vehicle connectivity, healthcare, smart cities, and digital transformation [6] - Cerence Inc. is a global leader in AI-powered experiences for automotive and transportation, with over 525 million cars equipped with its technology, emphasizing voice, generative AI, and large language models [7] Collaboration Details - Neusoft will utilize its advanced intelligent cockpit software platform (NAGIC) to integrate Cerence AI's expertise in conversational AI and large language models, aiming for innovative applications in voice interaction [3][4] - The collaboration will leverage Neusoft's global product development network and Cerence AI's technological strengths to expand into global target markets [4] Future Outlook - Neusoft plans to continue its philosophy of open collaboration and shared ecosystem success, working with more technology partners to address market challenges in automotive intelligence and AI [5]
DeepSeek新模型曝光
财联社· 2026-01-21 06:34
Core Viewpoint - DeepSeek is advancing its AI model capabilities with the introduction of MODEL1, which is designed for efficient inference and optimized for various GPU architectures, indicating a strategic focus on enhancing performance and reducing memory usage in AI applications [4][5][6]. Group 1: MODEL1 and FlashMLA - MODEL1 is a newly revealed model architecture within DeepSeek's FlashMLA, which is a software tool optimized for NVIDIA Hopper architecture GPUs, aimed at accelerating large model inference generation [4]. - FlashMLA utilizes a multi-layer attention mechanism (MLA) to minimize memory usage and maximize GPU hardware efficiency, which is crucial for the performance of DeepSeek's models [4][5]. - MODEL1 is expected to be a low-memory consumption model suitable for edge devices and cost-sensitive scenarios, with optimizations for long sequence tasks such as document understanding and code analysis [5]. Group 2: DeepSeek's Model Development - DeepSeek's existing models represent two technical routes: the V series focusing on comprehensive performance and the R series targeting complex reasoning tasks [6]. - The V3 model, launched in December 2024, marked a significant milestone with its efficient MoE architecture, followed by rapid iterations leading to V3.1 and V3.2, which enhance reasoning and agent capabilities [6]. - The R1 model, released in January 2025, excels in solving complex reasoning tasks through reinforcement learning and introduces a "deep thinking" mode, showcasing DeepSeek's commitment to advancing AI capabilities [7]. Group 3: Future Developments - DeepSeek is expected to launch its next flagship AI model, DeepSeek V4, around mid-February 2025, which is anticipated to have enhanced coding capabilities [7]. - Recent technical papers from DeepSeek discuss new training methods and an AI memory module inspired by biology, suggesting that these innovations may be integrated into upcoming models [7].
I Predicted Alphabet Would Be the Best-Performing "Magnificent 7" Stock in 2025. Here Are the Main Reasons Why It Actually Happened.
Yahoo Finance· 2026-01-20 19:50
Key Points Alphabet was able to change investors' perceptions about the impact of AI on its business in 2025. The company looks well positioned to continue to drive growth in 2026. 10 stocks we like better than Alphabet › Heading into 2025, I predicted that the best performer among the so-called "Magnificent Seven" stocks for the year would be Alphabet (NASDAQ: GOOGL) (NASDAQ: GOOG). That proved to be the correct call, as the stock climbed more than 65%, finishing well ahead of second-place perform ...
大模型长脑子了?研究发现LLM中层会自发模拟人脑进化
3 6 Ke· 2026-01-15 01:26
生物智能与人工智能的演化路径截然不同,但它们是否遵循某些共同的计算原理? 最近,来自帝国理工学院、华为诺亚方舟实验室等机构的研究人员发表了一篇新论文。该研究指出,大型语言模型(LLM)在学习过程中会自发演化出 一种协同核心(Synergistic Core)结构,有些类似于生物的大脑。 论文标题:A Brain-like Synergistic Core in LLMs Drives Behaviour and Learning 论文地址:https://arxiv.org/abs/2601.06851 研究团队利用部分信息分解(Partial Information Decomposition, PID)框架,对 Gemma、Llama、Qwen 和 DeepSeek 等模型进行了深度剖析。 他们发现,这些模型的中层表现出极强的协同处理能力,而底层和顶层则更偏向于冗余处理。 协同与冗余:LLM 的内部架构 研究团队将大型语言模型视为分布式信息处理系统,其核心实验设计旨在量化模型内部组件之间交互的本质。为了实现这一目标,研究者选取了 Gemma 3、Llama 3、Qwen 3 8B 以及 DeepSeek ...
幻方量化去年收益率56.6% 为DeepSeek提供超级弹药
2 1 Shi Ji Jing Ji Bao Dao· 2026-01-14 02:15
Core Insights - The article highlights the impressive returns of Fantom Quantitative, which achieved an average return of 56.55% in 2025, ranking second among quantitative private equity firms in China, only behind Lingjun Investment with a return of 73.51% [1] - Fantom Quantitative's average return over the past three years is 85.15%, and 114.35% over the past five years, providing substantial funding support for DeepSeek's large model research [2] - Founded in 2015 by Liang Wenfeng, Fantom Quantitative focuses on AI quantitative trading and has a current management scale exceeding 70 billion yuan, maintaining a leading position in the domestic private quantitative investment sector [2][3] Company Overview - Fantom Quantitative has a team composed of award-winning mathematicians, physicists, and experts in AI, employing interdisciplinary collaboration to tackle challenges in deep learning, big data modeling, and quantitative analysis [2] - The company has been utilizing machine learning for fully automated quantitative trading since 2008 and has expanded rapidly since its inception [2] - Significant investments were made in AI training platforms, with "Firefly No. 1" established in 2019 and "Firefly No. 2" in 2021, leading to the establishment of DeepSeek in July 2023 [3] Financial Performance - Liang Wenfeng holds a majority stake in Fantom Quantitative and has ceased to introduce external funding for the fund, indicating a strong accumulation of capital for supporting large model research [4] - The strong performance of Fantom Quantitative is estimated to have generated over 700 million USD in revenue last year, assuming a 1% management fee and 20% performance fee [4] DeepSeek Developments - DeepSeek's V3 model has a total training cost budget of 5.57 million USD, while competitors like Zhizhu and MiniMax have reported significant R&D expenditures [5] - DeepSeek plans to release its next-generation AI model, DeepSeek V4, around the Lunar New Year, which is expected to surpass current leading models in programming capabilities [5]
2 No-Brainer AI Stocks to Buy Hand Over Fist in 2026
Yahoo Finance· 2026-01-13 22:20
Key Points Amazon's biggest AI advantage could come from using the technology internally. Micron stands to benefit tremendously from a global memory hardware shortage. 10 stocks we like better than Amazon › It's finally 2026, and that means it's been just over three years since OpenAI's ChatGPT gave rise to the generative artificial intelligence (AI) industry. As the hype cycle gets long in the tooth, investors should pivot to more value-oriented picks like Amazon (NASDAQ: AMZN) and Micron Technolog ...