Workflow
OpenAI
icon
Search documents
大模型究竟是个啥?都有哪些技术领域,面向小白的深度好文!
自动驾驶之心· 2025-08-05 23:32
Core Insights - The article provides a comprehensive overview of large language models (LLMs), their definitions, architectures, capabilities, and notable developments in the field [3][6][12]. Group 1: Definition and Characteristics of LLMs - Large Language Models (LLMs) are deep learning models trained on vast amounts of text data, capable of understanding and generating natural language [3][6]. - Key features of modern LLMs include large-scale parameters (e.g., GPT-3 with 175 billion parameters), Transformer architecture, pre-training followed by fine-tuning, and multi-task adaptability [6][12]. Group 2: LLM Development and Architecture - The Transformer architecture, introduced by Google in 2017, is the foundational technology for LLMs, consisting of an encoder and decoder [9]. - Encoder-only architectures, like BERT, excel in text understanding tasks, while decoder-only architectures, such as GPT, are optimized for text generation [10][11]. Group 3: Core Capabilities of LLMs - LLMs can generate coherent text, assist in coding, answer factual questions, and perform multi-step reasoning [12][13]. - They also excel in text understanding and conversion tasks, such as summarization and sentiment analysis [13]. Group 4: Notable LLMs and Their Features - The GPT series by OpenAI is a key player in LLM development, known for its strong general capabilities and continuous innovation [15][16]. - Meta's Llama series emphasizes open-source development and multi-modal capabilities, significantly impacting the AI community [17][18]. - Alibaba's Qwen series focuses on comprehensive open-source models with strong support for Chinese and multi-language tasks [18]. Group 5: Visual Foundation Models - Visual Foundation Models are essential for processing visual inputs, enabling the connection between visual data and LLMs [25]. - They utilize architectures like Vision Transformers (ViT) and hybrid models combining CNNs and Transformers for various tasks, including image classification and cross-modal understanding [26][27]. Group 6: Speech Large Models - Speech large models are designed to handle various speech-related tasks, leveraging large-scale speech data for training [31]. - They primarily use Transformer architectures to capture long-range dependencies in speech data, facilitating tasks like speech recognition and translation [32][36]. Group 7: Multi-Modal Large Models (MLLMs) - Multi-modal large models can process and understand multiple types of data, such as text, images, and audio, enabling complex interactions [39]. - Their architecture typically includes pre-trained modal encoders, a large language model, and a modal decoder for generating outputs [40]. Group 8: Reasoning Large Models - Reasoning large models enhance the reasoning capabilities of LLMs through optimized prompting and external knowledge integration [43][44]. - They focus on improving the accuracy and controllability of complex tasks without fundamentally altering the model structure [45].
Claude Just Got a Big Update (Opus 4.1)
Matthew Berman· 2025-08-05 23:02
Model Release & Performance - Anthropic 发布了 Claude Opus 4.1%,是对 Claude Opus 4 的升级,尤其在 Agentic 任务、真实世界编码和推理方面 [1] - SWEBench verified 基准测试中,Claude Opus 4.1% 的得分从 Opus 4 的 72.5% 提升至 74.5%,提升了 2 个百分点 [3] - Terminal Bench 基准测试中,Claude Opus 4.1% 的终端使用能力从 39.2% 提升至 43.3%,提升了 4.1 个百分点 [4] - GPQA Diamond(研究生水平推理)基准测试中,Claude Opus 4.1% 的得分从 79.6% 提升至 80.9%,提升了 1.3 个百分点 [4] - Towbench(Agentic 工具使用)基准测试中,Claude Opus 4.1% 在零售方面的得分从 81.4% 提升至 82.4%,提升了 1 个百分点,但在航空方面从 59.6% 下降至 56%,下降了 3.6 个百分点 [5] - 多语言问答基准测试中,Claude Opus 4.1% 的得分从 88.8% 提升至 89.5%,提升了 0.7 个百分点 [5] - Amy 2025 基准测试中,Claude Opus 4.1% 的得分提升了 2.5 个百分点至 78% [5] Competitive Positioning & Future Outlook - 在 SWEBench 和 Terminal Bench 基准测试中,Claude Opus 4.1% 优于 OpenAI 的 GPT-3 和 Gemini 1.5 Pro [5] - 在 GPQA Diamond 和 Agentic 工具使用基准测试中,Claude Opus 4.1% 不及 OpenAI 的 GPT-3 和 Gemini 1.5 Pro [6] - 在高中数学竞赛基准测试中,Claude Opus 4.1% 的得分低于 OpenAI 的 GPT-3 (88.9%) 和 Gemini 1.5 Pro (88%),仅为 78% [6] - Claude 目前被广泛认为是市场上最佳的编码模型,尤其擅长 Agentic 编码和 Agent-driven 开发 [7]
六年来首次!OpenAI发布两款开放权重AI推理模型!奥尔特曼称其为“全球最佳开放模型”
Mei Ri Jing Ji Xin Wen· 2025-08-05 22:57
OpenAI向开源模型迈出重要一步:六年来首次推出开放权重模型。 OpenAI首席执行官山姆·奥尔特曼当地时间8月5日宣布,公司将在未来几天里带来许多新东西,其中周 二迎来一项"小而重磅"的更新——预热已久的开源模型GPT-OSS。 两款模型都以宽松的Apache 2.0许可证发布,企业在商用前无需付费或获得许可。 奥尔特曼在社交媒体表示:gpt-oss是一个重大突破,这是最先进的开放权重推理模型,具有与o4-mini 相当的强大现实世界性能,可以在你自己的电脑(或手机的较小版本)上本地运行。我们相信这是世界 上最好、最实用的开放模型。 简单而言,OpenAI在8月5日共发布两款开放权重AI推理模型。其中参数量达到1170亿的gpt-oss-120b能 力更强,可以由单个英伟达专业数据中心GPU驱动;参数量210亿的gpt-oss-20b模型,则能够在配备 16GB内存的消费级笔记本电脑上运行。 同时,亚马逊宣布将首次向客户提供OpenAI的模型,计划在其Bedrock和SageMaker平台上提供OpenAI 的开放AI权重新模型。这是云计算巨头亚马逊首次提供OpenAI的产品。 gpt-oss-20b和1 ...
最高达250%!特朗普称将在“未来一周左右”宣布对芯片与药品进口征税
Zhi Tong Cai Jing· 2025-08-05 22:35
此举可能对依赖高端芯片的大型数据中心运营商构成重大打击,包括微软(MSFT.US)、OpenAI、 Meta(META.US)和亚马逊(AMZN.US)等公司。上述企业正在大举投资人工智能业务,对先进芯片的需 求尤为迫切。一旦征税,成本可能急剧上升。 美国总统特朗普周二表示,美国政府将在"未来一周左右"宣布对半导体和制药行业的进口产品加征关 税,意在重塑全球贸易格局,并迫使关键产业回流美国本土制造。 特朗普在接受专访时称:"我们将先对药品加征一项初步的小幅关税,但在一年到一年半之内,这项关 税将升至150%,最终达到250%,因为我们希望药品在美国本土生产。" 他还表示,美国政府将"另行宣布"对半导体和芯片的关税措施。美国商务部自今年4月起便对全球半导 体市场展开调查,为潜在的关税行动奠定法律基础。根据相关预测,全球半导体市场的年销售额接近 7000亿美元。 此次涉及药品、金属等关键行业的关税措施,均源自依据《贸易扩展法》第232条进行的国家安全调 查。这类调查周期一般为九个月,较特朗普此前动用紧急权力对特定国家征税具有更强的法律基础。后 者目前正面临法院的法律挑战。 除半导体之外,特朗普还警告制药行业将面临 ...
美股三大指数集体收跌,大型科技股多数走低,中概股涨跌不一
Feng Huang Wang· 2025-08-05 22:13
Market Overview - The US stock market experienced a collective decline on August 5, with the Dow Jones down 0.14% to 44,111.74 points, the S&P 500 down 0.49% to 6,299.19 points, and the Nasdaq Composite down 0.65% to 20,916.55 points [1][2]. Economic Indicators - The ISM reported that the US services PMI for July was 50.1, below market expectations of 51.5 and the previous month's 50.8, indicating near stagnation in service sector growth [2][3]. - Employment indicators fell from 47.2 to 46.4, marking a low point since the COVID-19 pandemic [2]. - The prices for materials and services rose to 69.9, the highest since October 2022 [2]. Policy and Regulatory Environment - The report highlighted challenges for the Federal Reserve, as rising price indices contrast with weakening activity and employment metrics [3]. - New tariff announcements from former President Trump, including a "small tariff" on imported drugs that could rise to 250%, are expected to impact market sentiment [3]. Company Performance - Major tech stocks mostly declined, with Nvidia down 0.97%, Microsoft down 1.47%, and Meta down 1.66%. Amazon was an exception, rising by 0.99% [4][5]. - Palantir's stock surged by 7.85%, reaching a market cap exceeding $400 billion, following a quarterly revenue report that surpassed $1 billion, a 48% year-over-year increase [5]. Chinese Stocks - The Nasdaq Golden Dragon China Index fell by 0.56%, with mixed performances among popular Chinese stocks [7]. Company News - OpenAI, Google, and Anthropic have been approved as AI suppliers for the US government, which is expected to accelerate the adoption of AI tools in federal operations [8][9]. - Coinbase announced plans to issue $2 billion in convertible bonds to raise funds for stock buybacks and debt repayment [10]. - AMD reported a second-quarter net profit of $781 million, a 31% year-over-year decline, with revenues of $7.69 billion, a 32% increase [11]. - Supermicro's fourth-quarter net sales were $5.76 billion, below analyst expectations, with a projected fiscal year 2026 net sales of at least $33 billion [12]. - Lucid reported a second-quarter adjusted loss of $632.1 million, with revenues of $259.4 million, a 29% increase year-over-year [13].
OpenAI Goes OPEN-SOURCE! gpt-oss is HERE!
Matthew Berman· 2025-08-05 22:09
Open AAI has delivered on their promise to release a state-of-the-art open-source model. This is GPTOSS. It comes in two sizes, a 120 billion parameter version and a 20 billion parameter version.These are state-of-the-art openweight language models. Open weight. So, not just open- source, but they are actually releasing the weights to these models.Now, for some benchmarks, here is the code forces competition. Now the 120 billion parameter version with tools scores a 2622. That is compared to 03 a frontier m ...
OpenAI发布ChatGPT世代首个开源模型gpt-oss,4060Ti都能跑得动。
数字生命卡兹克· 2025-08-05 22:08
8月6号,真的今夕是何年了。 一晚上,三个我觉得都蛮大的货。 先是晚上10点,Google发了一个世界模型(但期货),Genie 3。 这个非常的强,我看的热血沸腾,我这两天也会单独写一篇文章,来聊聊这个玩意,真的,作为一个这么多年的游戏和VR玩家,看到Genie 3非常的激 动。 然后就是12点半,Anthropic突然就发布了Claude Opus 4.1,在编程能力上继续进化。 这节奏,感觉就是来狙击OpenAI的。 然后,重头戏来了。 凌晨1点,OpenAI在GPT-2后,在整个ChatGPT世代,官宣发布了他们的第一个开源模型,GPT-oss。 真的,不眠之夜。 直接来聊聊GPT-oss。 很强,非常强,OpenAI终于干了点人事。 也就是说,20B模型的大小就12.8GB ,最低只要16GB内存就能跑,我这个破壁5080的16G卡,也能本地跑的动了20B的gpt-oss了。 GPT-oss一共开源了两款模型,120B和20B,都是MoE,纯文本、非多模态的推理模型, Apache 2.0 许可,也就是最宽松的那种,你随便用 。 | Model | Layers | Total Params | A ...
刚刚,OpenAI开源2个推理模型:笔记本/手机就能跑,性能接近o4-mini
量子位· 2025-08-05 21:09
而这次的名字也是非常的直接,gpt-oss,即Open Source Series,意思就是"开源系列"。 它们的亮点如下: 并且它俩均采用Apache 2.0许可证,允许商用无需付费或授权。 从性能角度来看,gpt-oss已经达到了开源模型里推理性能的第一梯队,但在代码生成和复杂推理任务中仍略逊于闭源模型(如GPT-o3和 o4-mini)。 gpt-oss-120b :1170亿参数(MoE架构,激活参数约51亿),可在单张80GB GPU上运行,性能接近闭源的o4-mini。 gpt-oss-20b :210亿参数(Moe架构,激活参数约36亿),可在16GB内存的消费级设备上运行,性能接近o3-mini。 金磊 发自 凹非寺 量子位 | 公众号 QbitAI 没能等到GPT-5,但OpenAI在深夜却 很突然地open了一下 —— 开源两个推理模型: gpt-oss-120b 和 gpt-oss-20b 。 要知道,上一次OpenAI开源模型还是6年前,也就是2019年的GPT-2。 | | gpt-oss-120b | gpt-oss-20b | OpenAl o3 | OpenAl o4-mini ...
Where Curiosity Meets Talent
Y Combinator· 2025-08-05 20:34
your inherent ability already is like one part of the ven diagram and then the other part is just something weird. It's literally just like where does your interest come from. Like I'm really taken by to what degree both open AAI and SpaceX for instance were you know the genesis came from like interest and a hunch and just like not really any commercial intent and yet coming out the other side that was enough to attract the smartest people in the world, attract capital and then really create the most enduri ...
AMD reports weaker-than-expected earnings even as revenue tops estimates
CNBC· 2025-08-05 20:34
Core Insights - Advanced Micro Devices (AMD) reported fiscal second quarter earnings that missed estimates, leading to a 5% decline in stock price during extended trading [1] - AMD's net income for the quarter was $872 million, or 54 cents per share, a significant increase from $265 million, or 16 cents per share in the same period last year [2] - AMD expects sales of $8.7 billion for the current quarter, exceeding previous earnings expectations of $8.3 billion [1] Financial Performance - AMD's adjusted gross margin was 43%, which would have been 54% without the impact of export control costs [6] - Revenue for AMD's data center segment reached $3.2 billion, up 14% year-over-year [6] - The Client and Gaming segment saw a 69% increase in revenue, driven by strong demand for AMD's latest desktop CPUs, with gaming revenue alone at $1.1 billion, up 73% year-over-year [7] Market Position and Strategy - AMD is the second-largest manufacturer of graphics processing units (GPUs) for artificial intelligence, following Nvidia, which dominates the market [3] - Major AI customers like Meta and OpenAI are increasingly considering AMD as an alternative to Nvidia's expensive chips, particularly for inference applications [3] - AMD announced new AI chips, the MI400, expected to launch next year, with OpenAI's CEO committing to using these GPUs [4] Regulatory Challenges - AMD faced export controls on some AI chips, notably the MI308, which previously cost the company $800 million in the June quarter [5] - The company anticipates resuming shipments of the MI308 after receiving indications of potential waivers from the Trump administration [5]