GPU
Search documents
Arm is releasing its first in-house chip in its 35-year history
TechCrunch· 2026-03-24 19:48
Storied semiconductor and software company Arm Holdings is starting to make its own chips after nearly 36 years of licensing its designs to companies like Nvidia and Apple.At an event Tuesday in San Francisco, the company revealed the Arm AGI CPU, a production-ready chip built for running inference in an AI data center. The U.K.-based company developed the chip using its Arm Neoverse family of CPU IP cores and through a partnership with Meta. Meta is also the chip’s first customer of the Arm AGI CPU, which ...
Quantum Computing Reaches an Inflection Point With NVIDIA NVQLink | GTC 2026
NVIDIA· 2026-03-20 20:54
We're standing here at the NVIDIA NVQLink booth here at GTC in San Jose. NVQLink is really the Rosetta Stone for communicating between quantum hardware and classical supercomputing. And that's crucial because, without GPU supercomputers, we’ll never be able to turn the cubits that we have today into useful quantum computers.There are many areas where we don't expect quantum computers to ever be better than conventional computers, but in some areas they're absolutely transformational. And those are things li ...
Will.i.am showcases new agentic EV dubbed Trinity at GTC.
Yahoo Finance· 2026-03-20 19:59
AI democratizes knowledge. AI is this new renaissance where you put your imagination to work and dream up that doesn't exist yet. I want to solve a problem when it comes to AI and job creation.I wanted to build a product that America doesn't really focus on at the moment. Agentic micromobility. Building the vehicle from the agent up.The agent is not just, you know, helping me navigate getting to and from work. It has actually helped me do my work. Inside the Trinity is a GPU on board.All the inferences are ...
Meta Platforms Just Unveiled Its New AI Chips. Should Nvidia Investors Be Worried?
The Motley Fool· 2026-03-15 11:00
Core Viewpoint - Nvidia, recognized as a leader in artificial intelligence (AI), faces increasing competitive pressures as the industry shifts from training large language models to inference, particularly with Meta Platforms announcing new custom chips [1][9]. Group 1: Nvidia's Competitive Position - Nvidia's dominance in AI may be challenged as large customers begin to adopt custom XPU solutions, raising concerns for investors [1][9]. - Despite the competitive landscape, Nvidia maintains a strong lead in training infrastructure, which is expected to continue growing [10]. Group 2: Meta's Chip Development - Meta Platforms launched four new AI chips: MTIA 300, MTIA 400, MTIA 450, and MTIA 500, designed for various inference workloads [2][3]. - The MTIA 300 is currently in use, while the other models will be rolled out starting in early 2027, focusing on generative AI inference [4]. - Meta employs a modular chip design strategy, allowing for rapid iteration and alignment with evolving AI workloads [4][5]. Group 3: Broadcom's Role and Industry Trends - Broadcom, a key partner for Meta, is witnessing a trend towards XPUs over traditional GPUs, emphasizing the need for specialized chips for AI workloads [6][7]. - Broadcom's CEO highlighted that XPUs offer greater flexibility and performance for specific AI tasks compared to general-purpose GPUs [7]. Group 4: Market Dynamics and Future Outlook - The emergence of new inference chipmakers is not expected to displace Nvidia's traditional training-focused GPUs, as overall AI computing demand continues to grow exponentially [13]. - Meta's ongoing partnership with Nvidia, including a significant multiyear deal for deploying Nvidia chips, indicates that even with new chip designs, Nvidia's infrastructure remains essential [11][12].
这颗GPU,改变了行业
半导体行业观察· 2026-03-02 01:41
Core Insights - GeForce 3, released in February 2001, marked a pivotal moment in GPU history as the first truly programmable GPU supporting DirectX 8.0 pixel and vertex shaders, allowing graphics programmers to write programs that run on the GPU [4][11] - Prior to GeForce 3, most GPUs were fixed-function accelerators, requiring the CPU to handle all special graphics effects, which limited the capabilities of early graphics processing [3][4] - GeForce 3's architecture included a "light-speed memory architecture" that significantly improved effective memory bandwidth, providing advantages at higher resolutions despite similar rasterization performance to its predecessor, GeForce 2 Pro [6][8] Product Evolution - The introduction of GeForce 3 Ti500 in 2001 addressed some performance shortcomings and established the "Ti" suffix, which continues to be used in NVIDIA's graphics cards today [7] - The original Xbox, released in late 2001, solidified GeForce 3's role as a foundational technology for future gaming consoles, showcasing NVIDIA's contributions beyond just graphics to include audio hardware and memory controllers [8] - Although GeForce 3 was not an immediate commercial success, it laid the groundwork for the more popular GeForce 4 series, which enhanced Direct3D support and introduced new features like 3D textures and improved shader capabilities [9][10] Long-term Impact - The evolution from GeForce 3 to the GeForce 8 series, which introduced the Tesla microarchitecture and unified shader design, reflects a significant shift towards fully programmable GPUs, enabling broader applications beyond gaming [10][11] - The rise of general-purpose GPU (GPGPU) computing, initially used for scientific calculations and later for cryptography and AI, can be traced back to the programmability established by GeForce 3 [10][11] - GeForce 3's legacy is evident in the current dominance of NVIDIA in the AI data center market, demonstrating how early innovations in gaming graphics have had far-reaching implications across various industries [11]
花旗闭门会-CPO的演进路径和对中国光模块影响-技术架构解析和看好天孚通信
花旗· 2026-03-01 17:22
花旗闭门会-CPO 的演进路径和对中国光模块影响,技术架 构解析和看好天孚通信 20260227 摘要 2027 年 CPU 交换机基础假设为 20.9 万台,综合考虑横向扩展(4 万 台光模块相关交换机)和纵向扩展(6.9 万台)需求,以及上行链路容 量。光模块需求下调因升级需求驱动,需外部接口对接数据中心和 AI 集 群。 下一代机柜采用扩张式设计,单机柜可部署约 144 颗 GPU 芯片,高密 度计算需光模块替代铜缆。每个节点约需 648 个光引擎,FAU 数量相近; 外部光源(ELS)需求取决于通道带宽,用于交换机间通信和内部连接。 NVR 576 配备 24 个交换托盘组件,每个模块对应 6 个组件配置,整体 配置需匹配总传输带宽,保障 GPU 侧带宽需求。数据测算需在 GPU、 计算节点等维度验证,通过带宽与逻辑校验市场测算的合理性。 2027 年市场规模测算:FAU 连接器约 22 亿美元,ELS 约 77 亿美元, 光纤整形器约 33 亿美元,光纤托盘约 56 亿美元。下调可封装光传输展 望因 CSP 提前锁定产能,改变了对节奏与供需结构的判断。 Q&A 对 CPO 在供应链中的潜在落地时间点 ...
X @Avi Chawla
Avi Chawla· 2026-02-25 06:30
8x faster LLM inference than Cerebras is here!!And it generates 17,000 tokens per second.Today, a key bottleneck in how LLM inference works is that when you run a model on any GPU, the model weights live in memory, and the compute cores have to constantly fetch those weights to do math.That back-and-forth between memory and compute is the single biggest reason inference is slow. It's also the reason we need expensive HBM stacks, liquid cooling, and high-speed interconnects, making AI data centers costly.Taa ...
未知机构:华西计算机每日资讯0223169亿融资押注专用芯片Taalas要-20260224
未知机构· 2026-02-24 03:35
Summary of Key Points from Conference Call Records Industry and Company Involved - **Company**: Ant Group - **Company**: Zhifang Technology - **Company**: Taalas - **Industry**: AI and Technology Core Insights and Arguments - **Ant Group's AI Strategy**: Ant Group's CEO Han Xinyi introduced a dual AI strategy named "Two Flowers," focusing on wealth and health management through AI. The strategy aims to penetrate the vast health market and enhance professional service offerings while developing AI payment systems to create a new commercial ecosystem [1][2] - **Zhifang Technology's Financing**: Zhifang Technology completed a Series B financing round exceeding 1 billion RMB, achieving a valuation over 10 billion RMB. This marks the company as one of the fastest-growing embodied AI firms globally, having completed seven rounds of financing within six months [1][2] - **Taalas's Chip Development**: Taalas announced a new funding round of $169 million, bringing total funding to approximately $219 million. The company introduced its first functional demonstration chip, HC1, optimized for the open-source model Llama 3.1, claiming to generate 17,000 tokens per second, outperforming Nvidia's H200 by 73 times while consuming only one-tenth of its power [2] Other Important but Potentially Overlooked Content - **Technological Advancements**: The HC1 chip utilizes TSMC's 6nm process technology, indicating a significant advancement in specialized AI processing capabilities [2] - **Market Context**: The strong performance of Ant Group's AI initiatives during the 2026 Spring Festival reflects the rapid implementation of its strategic goals, particularly in the health sector [1] - **Global AI Cycle Impact**: The ongoing global AI cycle is supporting South Korea's export growth, which saw a 47.3% year-on-year increase in exports for the first 20 days of February, indicating a robust demand for technology-related exports [4]
ARM,失宠了
半导体行业观察· 2026-02-19 02:46
Core Viewpoint - NVIDIA has sold its remaining stake in ARM for approximately $140 million, marking a significant shift from its previous attempt to acquire the company, which is crucial for AI infrastructure development [2] Group 1: NVIDIA and ARM Relationship - NVIDIA's collaboration with ARM has been vital for launching key products like Grace Hopper and Blackwell, with ARM playing a critical role in the upcoming Vera CPU [2] - The sale of ARM shares coincides with growing skepticism about ARM's position in the AI competition [2] Group 2: CPU Market Dynamics - There is a notable shift in workload from GPU to CPU, particularly for agentic tasks, which emphasizes the increasing importance of CPUs [2] - Major cloud providers are experiencing a surge in demand for data center CPUs, contributing to the rapid expansion of the overall CPU market [3] Group 3: ARM vs. x86 Architecture - ARM-based CPUs are perceived to have weaker momentum in AI servers due to lower GPU scheduling efficiency compared to x86 [3] - x86 architecture is favored for agentic workloads due to its superior single-thread burst performance, which is critical in environments executing millions of micro-tasks per second [3] Group 4: Ecosystem and Market Trends - The x86 ecosystem is well-established in enterprise data centers, including firmware stacks and virtualization layers, driving demand for Intel and AMD server products [4] - NVIDIA's move to introduce x86 server racks aligns with the ongoing upgrade cycle among large cloud providers [4] Group 5: NVIDIA's Strategic Direction - NVIDIA is actively pursuing an x86 strategy in collaboration with Intel, integrating x86 solutions into NVLink-equipped server racks [5] - The sale of ARM shares is primarily a financial maneuver and does not significantly impact NVIDIA's overall product strategy, although future CPU generations may explore x86 diversification [5]