Workflow
数字计算
icon
Search documents
绕开光刻机“卡脖子” 中国新型芯片问世!专访北大孙仲:支撑AI训练和具身智能 可在28纳米及以上成熟工艺量产
Mei Ri Jing Ji Xin Wen· 2025-12-30 00:36
Core Viewpoint - A Chinese research team has developed a novel high-precision, scalable analog matrix computing chip based on resistive random-access memory (RRAM), achieving 24-bit fixed-point precision for the first time globally, which allows for reduced computational card usage for similar tasks [1][12]. Group 1: Technology and Innovation - The new chip represents a significant leap in precision, reducing the relative error from 1% to one part in ten million (10^-7), thus addressing the historical precision bottleneck of analog computing [7][15]. - The chip can support advanced applications such as 6G, embodied intelligence, and AI model training, and can be produced using mature processes of 28nm and above, circumventing the bottleneck of photolithography [1][12]. - The research team has introduced a new feedback circuit and utilized classic iterative optimization algorithms to enhance precision without sacrificing energy efficiency or speed [11][15]. Group 2: Market and Application Potential - The chip is particularly suited for applications requiring large-scale matrix operations, such as AI model training, 6G communications, and supercomputing tasks, which are fundamentally based on matrix calculations [10][20]. - The team aims to scale the matrix size from 16x16 to 128x128 within two years, with a long-term goal of reaching 512x512, which would enable practical applications in medium-scale scenarios [24][25]. - The technology provides a potential alternative to reliance on advanced processes and NVIDIA GPUs, positioning the team at the forefront of modern analog computing [10][26]. Group 3: Strategic Importance - This development offers a "detour" for China's computing capabilities, potentially reducing dependence on advanced manufacturing processes and foreign technology [10][26]. - The successful demonstration of this new path confirms its potential, although significant investment and collaboration across the technology and industry sectors will be necessary to fully realize its capabilities [10][26].
绕开光刻机“卡脖子”,中国新型芯片问世!专访北大孙仲:支撑AI训练和具身智能,可在28纳米及以上成熟工艺量产
Mei Ri Jing Ji Xin Wen· 2025-12-29 10:20
Core Insights - A Chinese research team has developed a new type of chip based on resistive random-access memory (RRAM) that achieves a precision of 24-bit fixed-point accuracy in analog matrix computations, marking a significant advancement in computational efficiency and energy consumption for AI applications [2][12][15] - This chip can support various cutting-edge applications, including 6G communication, embodied intelligence, and AI model training, while being produced using mature 28nm technology, thus avoiding reliance on advanced lithography processes [2][4][10] Technology Overview - The new chip represents a departure from traditional digital computing paradigms, which rely on binary logic and silicon-based transistors, to a more efficient analog computing approach that directly utilizes physical laws for calculations [4][6][15] - The precision of analog computing has been significantly improved, reducing relative error from 1% to one part in ten million (10⁻⁷), which is crucial for large-scale computations where errors can accumulate exponentially [8][12][15] Innovation Highlights - The chip's innovations include the use of RRAM as a core component, a novel feedback circuit design that minimizes energy consumption while enhancing accuracy, and the implementation of classic iterative optimization algorithms for efficient matrix equation solving [15][16] - The chip's architecture allows for high-speed, low-power solutions to matrix equations, making it suitable for applications that require rapid computations, such as second-order training methods in AI [19][21] Application Potential - The chip is particularly well-suited for medium-scale applications, such as AI model training and 6G MIMO systems, where it can outperform traditional digital chips [18][25] - Future plans include scaling the chip's matrix size from 16x16 to 128x128 within two years, with aspirations to reach 512x512, which would enhance its applicability in various computational scenarios [25][26] Strategic Value - This development provides China with a potential alternative to reliance on advanced processes and NVIDIA GPUs, positioning the country favorably in the global computational landscape [10][11] - The successful demonstration of this new computing paradigm is seen as a critical step towards addressing future computational demands, emphasizing the need for ongoing investment in technology and infrastructure [11][26]
我国科学家研究的芯片,突破世纪难题
半导体行业观察· 2025-10-14 01:01
Core Insights - The research team from Peking University has achieved a breakthrough in high-precision, scalable analog matrix equation solving, published in Nature Electronics, marking a significant advancement in analog computing technology [1][2] - This innovation demonstrates that analog computing can efficiently and accurately address core computational problems in modern science and engineering, potentially disrupting the long-standing dominance of digital computing [2][3] Group 1: Key Innovations - The first key innovation is the use of resistive random-access memory (RRAM), which allows for precise control of resistance states and retains data without power, enabling it to function as both a memory and a computing unit [4] - The second key innovation stems from a foundational discovery in 2019, where the team designed an analog circuit capable of solving matrix equations in a single step, significantly compressing traditional iterative algorithms [5] - The third key innovation is the "bit slicing" technique, which breaks down 24-bit precision into multiple 3-bit segments for processing, allowing for a more sophisticated and efficient analog computation [5] Group 2: Practical Implications - The breakthrough allows for solving matrix equations with 24-bit precision in just a few iterations, drastically reducing the computational steps required for complex tasks, such as 6G signal detection [7] - In the AI field, this advancement could alleviate the "computational bottleneck" faced by large models, enabling faster and more efficient training processes [7] - The technology also addresses critical challenges in 6G communication, enhancing signal detection capabilities while significantly reducing energy consumption [8]
重生之在《我的世界》做山姆·奥特曼:网友在线手搓ChatGPT
创业邦· 2025-10-08 03:20
Core Insights - A user has successfully created a version of ChatGPT within the game Minecraft, showcasing the potential of in-game programming and logic circuits [4][10][12] - The model consists of approximately 5 million parameters and is capable of engaging in basic conversations, demonstrating the feasibility of building complex AI systems using simple game mechanics [6][19][22] Technical Details - The model has 5,087,280 parameters, trained using the TinyChat dataset, with an embedding dimension of 240 and a vocabulary of 1,920 tokens [19][21] - It features 6 layers and 5 attention heads, with a context window size of 64 tokens, allowing for short dialogues [21][22] - The construction of the model occupies a volume of 1020×260×1656 blocks in Minecraft, indicating the scale of the project [24] Construction Process - The process involved training a small GPT model on a personal computer, compressing weights to low precision, and translating calculations into Minecraft's redstone circuitry [26][30] - A "compiler" script was used to map the trained model to redstone modules, facilitating the construction of the circuit [29][30] Broader Implications - The success of this project highlights the capabilities of Minecraft as a platform for digital logic experimentation, where players can build complex systems from basic binary signals [32][34] - The community has also explored various computational projects within Minecraft, including building CNNs and even simulating internet-like structures [38][46]
重生之在《我的世界》做山姆·奥特曼:网友在线手搓ChatGPT
量子位· 2025-10-06 05:42
Core Viewpoint - The article discusses the impressive achievement of creating a ChatGPT model within the game Minecraft, showcasing the potential of using redstone circuits to simulate complex computational tasks [1][2][4]. Group 1: Model Specifications - The constructed ChatGPT model has approximately 5 million parameters, specifically 5,087,280 [16]. - It utilizes a TinyChat dataset for training, with an embedding dimension of 240 and a vocabulary of 1,920 tokens [18]. - The model features 6 layers and 5 attention heads, with a context window size of 64 tokens, suitable for very short conversations [19]. Group 2: Construction Process - The process involves training a small GPT model on a personal computer, compressing weights to low precision, and exporting the model structure [25]. - The next steps include translating computational methods into pixel block language and defining reusable circuit modules [26][27]. - Finally, a "compiler" script is used to map the trained model to redstone modules, facilitating the construction of the entire setup [28][30]. Group 3: Redstone Circuit Functionality - Redstone circuits in Minecraft operate on binary logic, where signals can be either on (1) or off (0), allowing players to build complex logic gates and circuits [32][34]. - This capability enables the construction of basic computational systems, such as adders and counters, leading to the potential for creating CPUs and neural networks [34]. Group 4: Broader Implications - The article highlights that the development of computational systems in Minecraft is still in its infancy, with only about 1% of the potential explored [37]. - Other projects within Minecraft include building CNNs for digit recognition and creating various games and even an internet simulation [39][46]. - The narrative suggests that players in Minecraft may eventually surpass current AI capabilities, hinting at a future where Minecraft could play a role in advancing artificial general intelligence (AGI) [48][49].
辛顿教授世界人工智能大会演讲PPT
2025-07-29 02:10
Summary of Key Points from the Conference Call Industry or Company Involved - The discussion revolves around the field of Artificial Intelligence (AI), particularly focusing on Digital Intelligence versus Biological Intelligence. Core Points and Arguments 1. **Two Paradigms of Intelligence** - The essence of intelligence is reasoning, achieved through symbolic rules manipulating symbolic expressions. Learning can be secondary to understanding knowledge representation [7][8][9]. 2. **Evolution of Language Models** - Over the past 30 years, significant advancements have occurred in language modeling, including the introduction of embedding vectors and the invention of transformers by Google [13][14]. 3. **Understanding of Language by LLMs** - Large Language Models (LLMs) understand language similarly to humans by converting words into compatible feature vectors, indicating a level of comprehension in their responses [16][28]. 4. **Analogy of Words as Lego Blocks** - Words are compared to high-dimensional Lego blocks, which can model various concepts and communicate ideas effectively [20][24]. 5. **Digital vs. Biological Computation** - Digital computation, while energy-intensive, allows for easy knowledge sharing among agents with the same model. In contrast, biological computation is less energy-consuming but struggles with knowledge transfer [51]. 6. **Knowledge Transfer Mechanisms** - Knowledge can be distilled from a teacher to a student in AI systems, allowing for efficient learning and adaptation [41][48]. 7. **Challenges of AI Control** - A super-intelligence could manipulate users to gain power, raising concerns about control and safety in AI development [55][57]. 8. **Global Cooperation on AI Safety** - There is skepticism about international collaboration on AI safety measures against threats like cyber attacks and autonomous weapons [64]. 9. **Training Benevolent AI** - Techniques to train AI to be benevolent may be independent of those that enhance its intelligence, suggesting a need for focused research on AI safety [68][72]. Other Important but Possibly Overlooked Content - The discussion emphasizes the potential risks associated with AI development, likening the situation to owning a tiger cub that could become dangerous as it matures, highlighting the urgency for safety measures [61]. - The need for countries to establish well-funded AI safety institutes to focus on making AI systems that do not seek control is also noted [72].