Large language model

Search documents
X @Avi Chawla
Avi Chawla· 2025-06-14 06:30
Transformer vs. Mixture of Experts in LLMs, clearly explained (with visuals): ...
Apple Sets Plans for Delayed Siri Launch in 2026
Bloomberg Technology· 2025-06-13 19:50
It was this week that WWE DC left us hanging on theory and you bring us some clarity. That's exactly right. So they did not talk about Siri features pretty much at all at WDC, other than to say they're still delayed and they'll be arriving in the coming year.So I looked into it and the new launch time frame is likely going to be as part of what will be known as iOS 26.4% and that it's currently scheduled for the spring time next year. So around March, April. And this will deliver the delayed Siri features.T ...
1200行代码逆袭!DeepSeek工程师开源轻量级vLLM,吞吐量逼近原版
机器之心· 2025-06-13 04:31
机器之心报道 机器之心编辑部 开源社区的人应该对 vLLM 不陌生,它是一个由加州大学伯克利分校团队开发的高性能、开源 LLM 推理和服务引擎,核心目标是提升 LLM 的推理速度(吞吐 量)和资源利用率(尤其是内存),同时兼容 Hugging Face 等流行模型库。 简单来说,vLLM 能让 GPT、Mistral、LLaMA 等主流模型系列跑得更快、消耗更少资源,取得这些效果的关键是其创新的注意力机制实现方案 —— PagedAttention。 近日,DeepSeek AI 研究者、深度学习系统工程师俞星凯 从零开始构建了一个轻量级 vLLM 实现 ——Nano-vLLM,将代码简化到了 1200 行 。 | Inference Engine | Output Tokens | Time (s) | Throughput (tokens/s) | | --- | --- | --- | --- | | vLLM | 133,966 | 98.95 | 1353.86 | | Nano-vLLM | 133,966 | 101.90 | 1314.65 | 作者简介 GitHub 地址:https://g ...
Report: Apple Aims to Release AI-Powered Upgrade of Siri in Spring 2026
PYMNTS.com· 2025-06-13 02:02
Core Viewpoint - Apple is aiming to launch an AI-powered upgrade of its Siri voice assistant in spring 2026, after multiple delays from its original target of fall 2024 [1][2][3] Group 1: Release Timeline - The initial announcement for Siri's new features was made in June 2024, with a planned launch in fall 2024 [2][3] - The release date was later pushed to spring 2025 and then postponed to sometime in the coming year [3] - Apple has indicated that the new features could be previewed in the fall before the spring release [2] Group 2: Technical Challenges - Technical challenges have necessitated a complete rebuild of Siri, leading to shifts in management responsibilities within Apple [3] - Apple CEO Tim Cook acknowledged the need for "more time" to enhance Siri's AI capabilities during a quarterly earnings call [4][5] - Employees who recently left the company reported difficulties in updating Siri with advanced large language models (LLMs) for improved responses [5] Group 3: Competitive Landscape - Apple's cautious approach to generative AI contrasts with competitors like Amazon, Google, and Microsoft, which are aggressively adopting LLMs and enterprise-scale AI solutions [6]
Large language models in Xcode
Apple Developer· 2025-06-11 17:01
Xcode can now use large language models such as ChatGPT to provide coding assistance. You can ask general questions about Swift — like “tell me about swift concurrency”. And because of the integration with Xcode, the model can take your code into consideration and answer specific questions about your project, or even make changes on your behalf. ...
大模型能够自发形成“人类思维地图”!Nature子刊重磅研究揭示多模态大模型类脑机制
机器人圈· 2025-06-11 11:43
大模型≠随机鹦鹉!Nature子刊最新研究证明: 大模型内部存在着类似人类对现实世界概念的理解。 LLM能理解现实世界和各种抽象概念吗?还是仅仅在"鹦鹉学舌",纯粹依靠统计概率预测下一个token? 长期以 来,AI社区对这一问题存在很大的分歧。 有一种猜测是,纯粹基于语言的形式(例如训练语料库中token的条件分布)进行训练的语言模型不会获得任何语 义。 相反,它们仅仅是根据从训练数据中收集的表面统计相关性来生成文本,其强大的涌现能力则归因于模型和训练 数据的规模。这部分人将LLM称为"随机鹦鹉"。 但现在研究证明,并非如此! 中国科学院自动化研究所与脑科学与智能技术卓越创新中心的联合团队 在 《Nature Machine Intelligence》 发表 题为 《Human-like object concept representations emerge naturally in multimodal large language models》 的研 究。 | Received: 26 June 2024 | Changde Du@ 12, Kaicheng Fu1-2, Bincheng Wen ...
WWDC25: Introducing the Foundation Models framework
Apple Developer· 2025-06-10 23:01
The FoundationModels framework gives you access to the on-device Large Language Model that powers Apple Intelligence, with a convenient and powerful Swift API. You can use it to enhance existing features in your apps, like providing personalized search suggestions. Or you can create completely new features, like generating an itinerary in a travel app, all on-device. You can even use it to create dialog on-the-fly for characters in a game.It is optimized for generating content, summarizing text, analyzing u ...
Cerence (CRNC) Conference Transcript
2025-06-10 17:30
Summary of Cerence (CRNC) Conference Call - June 10, 2025 Company Overview - Cerence is a global leader in voice AI interaction within the automotive industry, spun off from Nuance Communication in 2019, focusing on automotive software solutions [4][5] - The company claims over 50% penetration in the global automotive market, with technology implemented in over 500 million vehicles [5][6] Key Points Market Position and Growth - Cerence is well-positioned in a growing market for automotive software, with strong relationships with major automotive OEMs [6] - The company has a unique market position with higher margins and less exposure to tariffs compared to other suppliers [8][10] Tariff Impact - As a software company, Cerence is not directly impacted by tariffs, but there are concerns about overall production implications [10][11] - The company anticipates limited production concerns for the upcoming quarter, despite potential tariff impacts [19][20] China Market - Cerence faces challenges penetrating the Chinese market due to strong local competition but maintains relationships with large Chinese OEMs for exports outside of China [12][13] - The company sees potential growth in relationships with Chinese OEMs for their products outside of China [13][15] Revenue and Royalties - Pro forma royalties have been relatively flat over the past year, with expectations for growth tied to new product launches and pricing strategies [20][21] - The company has seen a decline in prepaid license revenue, with a target of around $20 million for the current year [23][24] Pricing Per Unit (PPU) - The PPU metric has shown growth, increasing from $450 to $487 over the trailing twelve months, with expectations for further growth as new products are launched [25][26] - The company aims to increase PPU through higher penetration of its technology in vehicles and the introduction of more valuable AI products [30][31] AI Product Development - Cerence is excited about the upcoming XUI product, which will integrate a large language model for enhanced voice interaction capabilities in vehicles [45][46] - The XUI product aims to provide a unified interface for both embedded and connected features, enhancing user experience [34][60] Competitive Landscape - Competition comes from both big tech companies and smaller competitors, but Cerence believes its proven implementation capabilities give it an advantage [50][51] - There is a reluctance among OEMs to adopt big tech solutions, favoring branded experiences instead [62] Additional Insights - The company is focused on creating win-win situations with OEMs by potentially reducing costs while increasing capabilities [41][43] - Cerence is exploring ways to enhance user interaction through multimodal capabilities, allowing for more natural voice commands [39][40] This summary captures the essential points discussed during the conference call, highlighting Cerence's market position, challenges, and future growth strategies.
2 Social Media Stocks That Are Screaming Buys in June
The Motley Fool· 2025-06-10 08:36
After a rough start to the year, tech stocks have come roaring back. As of this writing, the tech-heavy Nasdaq Composite index has surged by more than 25% over the last two months. So with market sentiment swinging back in favor of tech stocks, let's have a closer look at two social media stocks that I believe are buys right now. Reddit First up is Reddit (RDDT). A relative newcomer as far as social media stocks go, Reddit has only been a public company for a little over a year. However, during that time, i ...
一招缓解LLM偏科!调整训练集组成,“秘方”在此 | 上交大&上海AI Lab等
量子位· 2025-06-10 07:35
IDEAL团队 投稿 量子位 | 公众号 QbitAI 大幅缓解LLM偏科,只需调整SFT训练集的组成。 本来不擅长coding的Llama 3.1-8B,代码能力明显提升。 上海交大&上海AI Lab联合团队提出创新方法 IDEAL ,可显著提升LLM在多种不同领域上的综合性能。 此外,研究还有一些重要发现,比如: 具体来看—— IDEAL方法 问题建模: 首先按照不同 的领域准备高质量的训练数据集: , 并给出对应的用于验证的验证集: 。通过在训练集上面训练模型θ,获得训练集上的最优参数:θ 论文 希望在验证 集上的损失达到最小。为了能够方便的调整训练集,论文引入了对应的变量β ,并将这个优化问题 显示地建模了出来: SFT后LLM部分能力甚至退化 大型语言模型 (LLM) 凭借其强大的理解和逻辑推理能力,在多个领域展现了惊人的能力。除了模型参数量的增大, 高质量的数据是公认的LLM性能提升最关键的影响因素。 当对模型进行监督微调(SFT)时,研究人员发现 LLM在多任务场景下常出现"偏科"现象 ——部分能力突出而部分 能力并未涨进,甚至退化。这种不平衡的现象导致大模型在不同的领域上能力不同,进而影响用户 ...