AlexNet
Search documents
刚刚,Geoffrey Hinton成为第二位引用量破百万的科学家
3 6 Ke· 2026-01-16 02:25
在他之前,只有他的老搭档、另一位「深度学习教父」Yoshua Bengio 达成了这一成就。目前,Hinton 的引用量仍在以惊人的速度增长,每一次引用都代 表着他对人工智能领域不可磨灭的贡献。从反向传播算法的推广到 AlexNet 的惊艳问世,从获得图灵奖到斩获 2024 年诺贝尔物理学奖,Hinton 的职业生 涯几乎就是一部现代 AI 的发展史。 这一数字不仅是学术影响力的量化,更是对这位 78 岁长者一生执着探索的最高致敬。 刚刚,Geoffrey Hinton 正式成为历史上第二位 Google Scholar 引用量突破 100 万大关的计算机科学家。 | TITLE | CITED BY | YEAR | | --- | --- | --- | | | * | | | Imagenet classification with deep convolutional neural networks | 188837 | 2012 | | A Krizhevsky, I Sutskever, GE Hinton, | | | | Advances in neural information proce ...
刚刚,Geoffrey Hinton成为第二位引用量破百万的科学家
机器之心· 2026-01-16 01:55
Core Viewpoint - Geoffrey Hinton has officially become the second computer scientist in history to surpass 1 million citations on Google Scholar, marking a significant milestone in his academic career and contributions to artificial intelligence [1][3]. Group 1: Academic Achievements - Hinton's citation count currently stands at 1,000,083, with an h-index of 192, indicating his substantial impact in the field of computer science and artificial intelligence [2]. - He is renowned for his work on backpropagation, which addressed the training challenges of multilayer neural networks, laying the groundwork for the deep learning revolution [10]. - Hinton, along with Yoshua Bengio and Yann LeCun, received the Turing Award in 2018, recognizing their pivotal contributions to the field of deep learning [13]. Group 2: Key Contributions - Hinton's notable innovations include the Boltzmann Machine, Restricted Boltzmann Machine, Deep Belief Network, Dropout technique, t-SNE for data visualization, Capsule Networks, and Knowledge Distillation, among others [14]. - His collaboration on AlexNet, which won the ImageNet competition in 2012, is considered a landmark moment that demonstrated the power of deep learning [16]. - The paper "Deep Learning," co-authored by Hinton, has garnered over 100,000 citations, summarizing the evolution and principles of deep learning [16]. Group 3: Personal Background and Career - Born into an academic family, Hinton's early life was marked by high expectations, which shaped his relentless pursuit of knowledge [5][8]. - He moved to Canada in the 1980s, where he established a long-term academic career at the University of Toronto, contributing significantly to the development of AI in Canada [9]. - Hinton's later years have seen him express concerns about the potential risks of AI, emphasizing the need for caution in its development [20]. Group 4: Legacy and Impact - Hinton's citation milestone reflects not only his individual achievements but also the collaborative efforts of his students, Alex Krizhevsky and Ilya Sutskever, who have also made significant contributions to AI [29]. - The historical context of Hinton's work illustrates the broader narrative of humanity's quest to understand intelligence, highlighting the transformative impact of his research on modern AI [31].
AI教父Geoffrey Hinton,全球第二个百万引用科学家
3 6 Ke· 2026-01-16 01:28
Core Insights - Geoffrey Hinton, a prominent figure in AI, has surpassed 1 million citations for his research papers, marking a significant milestone in academic recognition [1][3][12] - Hinton is the second individual globally to achieve this milestone, following Yoshua Bengio, who reached 1.036 million citations [7][10] - This achievement reflects the growing influence and recognition of deep learning theories and methodologies in the academic community [12] Academic Achievements - Hinton's most cited paper, "Imagenet classification with deep convolutional neural networks," has received 188,837 citations, highlighting its impact on the field [18][34] - Other notable works include "Deep Learning," co-authored with Bengio and Yann LeCun, which has garnered 107,646 citations, serving as a foundational text in deep learning [20][38] - Hinton's contributions span various influential papers, including "t-SNE" with 63,932 citations and "Dropout" with 60,895 citations, showcasing his extensive influence across multiple areas of machine learning [21][47] Historical Context - Hinton's work is rooted in decades of academic research, with significant contributions that have shaped the evolution of deep learning [18][25] - His early work during the AI winter and subsequent breakthroughs, such as the introduction of deep belief networks, played a crucial role in reviving interest in neural networks [27][28] - The recognition of Hinton, alongside Bengio and LeCun, with the Turing Award in 2018, underscores their collective impact on modern AI algorithms [28] Industry Implications - Hinton's research has laid the groundwork for contemporary AI applications, including large models like ChatGPT and Gemini, which rely on deep learning principles [24] - The advancements in deep learning, driven by Hinton's theories, have transformed various industries, particularly in computer vision and natural language processing [35][36] - The ongoing exploration of AI, as emphasized by Hinton, suggests that future research will continue to uncover the complexities of large models and their operations [24][49]
从洗碗工到“AI教母”,她又预言了下一个十年
3 6 Ke· 2026-01-13 07:31
Core Viewpoint - The next decade of AI is defined by "spatial intelligence," which emphasizes the need for AI to understand depth, distance, occlusion, and gravity to achieve true embodiment [1][10]. Group 1: Li Fei Fei's Background and Career - Li Fei Fei, known as the "AI Mother," has over 20 years of experience in AI research, with a focus on spatial intelligence as her latest guiding principle [2]. - Her autobiography, "The World I See," details her journey from a challenging childhood in the U.S. to becoming a prominent figure in AI, reflecting on her struggles and achievements [2][5]. - Li Fei Fei's career spans the evolution of AI from laboratory research to industrial application, making her autobiography a significant account of AI's development [2]. Group 2: ImageNet and AI Development - ImageNet, a large-scale visual database created by Li Fei Fei, played a crucial role in the advancement of AI, marking the beginning of the AI golden age [6][9]. - The project faced initial skepticism and challenges, but the use of Amazon's crowdsourcing service was pivotal in its success, allowing for efficient image labeling [8]. - The introduction of deep learning models like AlexNet, which utilized ImageNet, significantly improved AI's performance in image recognition tasks, reducing error rates dramatically [9]. Group 3: Spatial Intelligence and Future Directions - Li Fei Fei believes that the next breakthrough in AI will come from developing spatial intelligence, which encompasses understanding and generating three-dimensional environments [10][11]. - The current state of technology in spatial intelligence is still in its early stages, but Li Fei Fei is confident that significant advancements will occur within the next one to two years [11]. - She views spatial intelligence as a critical component in the pursuit of Artificial General Intelligence (AGI), suggesting that it is one of many keys needed to unlock this complex field [12].
AI教父Hinton首爆十年前拍卖:我早已内定谷歌必赢
3 6 Ke· 2025-12-21 23:25
Core Insights - The conversation between AI pioneers Hinton and Jeff Dean at NeurIPS 2025 highlighted the evolution of AI, discussing key breakthroughs and challenges in the field [1][4][14] Group 1: Historical Context and Key Developments - Hinton and Dean reflected on the early breakthroughs in machine learning and the significant impact of the Transformer paper, with Dean stating that Google does not regret publishing it due to its global influence [3][43] - The discussion included anecdotes about the development of AlexNet, which revolutionized image recognition, and the early days of Google Brain, emphasizing the importance of scaling in AI models [14][25][31] Group 2: Technical Insights and Innovations - Hinton's realization about the importance of scaling in AI models came after attending a talk by Ilya Sutskever, which shifted his perspective on computational power [13][31] - The conversation also covered the development of the Transformer model, which improved efficiency in processing and understanding data, allowing for better performance with less computational power [43][45] Group 3: Future Directions and Predictions - Looking ahead, Dean expressed excitement about scaling attention mechanisms and the potential for models to access vast amounts of data, which would require innovations in hardware [52][54] - Both Hinton and Dean acknowledged the transformative potential of AI in fields like healthcare and education, while also recognizing the uncertainty regarding job displacement and the creation of new opportunities [56][57]
为什么现代 AI 能做成?Hinton 对话 Jeff Dean
3 6 Ke· 2025-12-19 00:47
Core Insights - The conversation between Geoffrey Hinton and Jeff Dean at the NeurIPS conference highlights the systematic emergence of modern AI, emphasizing that breakthroughs are not isolated incidents but rather the result of simultaneous advancements in algorithms, hardware, and engineering [1] Group 1: AI Breakthroughs and Historical Context - The pivotal moment for modern AI occurred in 2012 during the ImageNet competition, where Hinton's team utilized deep neural networks with significantly more parameters and computational power than competitors, establishing deep learning's prominence [2][3] - Jeff Dean's early experiences with parallel algorithms in the 1990s laid the groundwork for future developments, although initial failures taught him the importance of matching computational power with model scale [4][5] Group 2: Hardware Evolution and Infrastructure - The TPU project was initiated in response to the need for custom hardware to support AI applications, leading to significant improvements in inference efficiency, with the first generation of TPUs achieving 30-80 times better performance than CPUs and GPUs [8] - The evolution of NVIDIA GPUs from AlexNet's two boards to the latest models continues to support large-scale training for companies like OpenAI and Meta, showcasing a diversified AI infrastructure landscape [9] Group 3: Convergence of Technology and Organization - The period from 2017 to 2023 saw the convergence of three critical technology curves: scalable algorithm architectures, centralized organizational structures, and a comprehensive engineering toolset, enabling large-scale AI applications [10][11][13] - The formation of the Gemini team at Google exemplified the importance of resource consolidation, allowing for focused efforts on AI model development and deployment [12] Group 4: Future Challenges in AI Scaling - The conversation identified three major challenges for AI scalability: energy efficiency, memory depth, and creative capabilities, which must be addressed to enable broader AI applications [16][18][21] - Achieving breakthroughs in these areas requires not only engineering optimizations but also long-term investments in foundational research, as many current technologies stem from decades-old academic studies [25][26] Group 5: Conclusion on AI Development - The journey of AI from conceptualization to widespread application is characterized by the alignment of several key factors: practical algorithms, robust computational support, and a conducive research environment [28]
被拒≠失败!这些高影响力论文都被顶会拒收过
具身智能之心· 2025-12-12 01:22
Core Insights - Waymo has released a deep blog detailing its AI strategy centered around its foundational model, emphasizing the use of distillation methods to create high-efficiency models for onboard operations [1][2] - Jeff Dean highlighted the significance of knowledge distillation, comparing it to the creation of the Gemini Flash model, which showcases the importance of distillation in AI model efficiency [1][2] Historical Context of Rejected Papers - Many foundational technologies in AI, such as optimizers for large models and computer vision techniques, were initially rejected by top conferences, showcasing a historical pattern of oversight in recognizing groundbreaking innovations [6] - Notable figures in AI, including Geoffrey Hinton and Yann LeCun, have faced rejection for their pioneering work, which was later recognized as transformative [6] Case Studies of Rejected Innovations - LSTM, a milestone for sequence data processing, was rejected by NIPS in 1996 but later became crucial in speech recognition and machine translation, highlighting the delayed recognition of its value [7][10] - SIFT, a dominant algorithm in computer vision, faced rejection from ICCV and CVPR due to its perceived complexity, yet proved to be vital in real-world image processing [11][13] - Dropout, a key regularization method for deep neural networks, was initially rejected for its radical approach but later became essential in training deep networks effectively [17][19] - Word2Vec, despite being rejected at ICLR, became a cornerstone in NLP due to its efficiency and practical application, eventually receiving recognition for its impact [20][24] - YOLO transformed object detection by prioritizing speed over precision, facing rejection for its perceived shortcomings but later becoming a widely adopted framework in the industry [28][30] Reflection on Peer Review Limitations - The peer review system often struggles to recognize disruptive innovations, leading to a systematic cognitive lag in evaluating groundbreaking research [40][41] - The tendency to equate mathematical complexity with research contribution can hinder the acceptance of simpler yet effective methods [41] - Historical examples illustrate that the true measure of a research's impact is not determined by initial peer review outcomes but by its long-term relevance and problem-solving capabilities [43][47]
何恺明NeurIPS 2025演讲盘点:视觉目标检测三十年
机器之心· 2025-12-11 10:00
Core Insights - The article highlights the significance of the "Test of Time Award" received by the paper "Faster R-CNN," co-authored by renowned researchers, marking its impact on the field of computer vision since its publication in 2015 [1][5][25] - The presentation by He Kaiming at NeurIPS 2025 summarizes the evolution of visual object detection over the past 30 years, showcasing key milestones and influential works that have shaped the field [6][31] Historical Development - The early attempts at face detection in the 1990s relied on handcrafted features and statistical methods, which were limited in adaptability and speed [12] - The introduction of AlexNet in 2012 demonstrated the superior feature extraction capabilities of deep learning, paving the way for its application in object detection [15] - The R-CNN model, proposed in 2014, revolutionized object detection by integrating CNNs for feature extraction and classification, although it initially faced computational challenges [17][18] Technological Advancements - The development of Faster R-CNN in 2015 addressed the speed bottleneck by introducing the Region Proposal Network (RPN), allowing for end-to-end real-time detection [25] - Subsequent innovations, such as YOLO and SSD in 2016, further enhanced detection speed by enabling direct output of object locations and categories [32] - The introduction of Mask R-CNN in 2017 added instance segmentation capabilities, while DETR in 2020 redefined detection using Transformer architecture [32][34] Future Directions - The article concludes with reflections on the ongoing exploration in computer vision, emphasizing the need for innovative models to replace outdated components as bottlenecks arise [35][36]
黄仁勋最新采访:依然害怕倒闭,非常焦虑
半导体芯闻· 2025-12-08 10:44
Core Insights - The discussion highlights the transformative impact of artificial intelligence (AI) and the role of NVIDIA in driving this technological revolution, emphasizing the importance of GPUs in various applications from gaming to modern data centers [2][10]. Group 1: AI and Technological Competition - The conversation underscores that the world is in a significant technological race, particularly in AI, where the first to reach advanced capabilities will gain substantial advantages [11][12]. - Historical context is provided, indicating that the U.S. has always been in a technological competition since the Industrial Revolution, with AI being the latest frontier [12][13]. Group 2: Energy and Manufacturing - The importance of energy growth and domestic manufacturing is emphasized as critical for national security and economic prosperity, with a call for revitalizing U.S. manufacturing capabilities [8][9]. - The discussion points out that without energy growth, industrial growth and job creation would be severely hindered, linking energy policies directly to advancements in AI and technology [9][10]. Group 3: AI Development and Safety - Concerns about the risks associated with AI are acknowledged, particularly regarding its potential military applications and ethical implications [19][20]. - The conversation suggests that AI's development will be gradual rather than sudden, with a focus on enhancing safety and reliability in AI systems [14][15]. Group 4: Future of AI and Knowledge Generation - The potential for AI to generate a significant portion of knowledge in the future is discussed, with predictions that AI could produce up to 90% of knowledge within a few years [41][42]. - The necessity for continuous verification of AI-generated information is highlighted, stressing the importance of ensuring accuracy and reliability in AI outputs [41][42]. Group 5: Cybersecurity and Collaboration - The dialogue emphasizes the collaborative nature of cybersecurity, where companies share information and best practices to combat threats collectively [23][24]. - The need for a unified approach to cybersecurity in the face of evolving threats is reiterated, suggesting that cooperation is essential for effective defense [23][24].
黄仁勋最新采访:依然害怕倒闭,非常焦虑
半导体行业观察· 2025-12-06 03:06
Core Insights - The discussion highlights the transformative impact of artificial intelligence (AI) and the role of NVIDIA in driving this technological revolution, emphasizing the importance of GPUs in various applications from gaming to modern data centers [1] - Huang Renxun discusses the risks and rewards associated with AI, the global AI race, and the significance of energy and manufacturing for future innovations [1] Group 1: AI and Technological Competition - The ongoing technological competition has been a constant since the Industrial Revolution, with the current AI race being one of the most critical [10][11] - Huang Renxun emphasizes that technological leadership is essential for national security and economic prosperity, linking energy growth to industrial growth and job creation [7][8] - The conversation touches on the historical context of technological races, including the Manhattan Project and the Cold War, underscoring the continuous nature of these competitions [11] Group 2: AI Development and Safety - Huang Renxun expresses optimism about the gradual development of AI, suggesting that advancements will be incremental rather than sudden [13] - The discussion addresses concerns about AI's potential risks, including the ethical implications of military applications and the need for robust cybersecurity measures [16][20] - Huang Renxun believes that AI's capabilities will increasingly focus on safety and reliability, reducing the occurrence of errors or "hallucinations" in AI outputs [14] Group 3: Future of Work and AI's Impact - The conversation explores the potential for AI to create a future where traditional jobs may become obsolete, leading to a society where individuals receive universal basic income [37] - Huang Renxun acknowledges the challenges of identity and purpose as AI takes over tasks traditionally performed by humans, emphasizing the need for society to adapt to these changes [38] - The discussion highlights the importance of maintaining human engagement and problem-solving in a future dominated by AI technologies [38] Group 4: Quantum Computing and Security - Huang Renxun discusses the implications of quantum computing on encryption and cybersecurity, suggesting that while current encryption methods may become outdated, the industry is actively developing post-quantum encryption technologies [22][23] - The conversation emphasizes the collaborative nature of cybersecurity efforts, where companies share information to enhance collective defenses against threats [20][21] - Huang Renxun asserts that AI will play a crucial role in future cybersecurity measures, leveraging its capabilities to protect against evolving threats [21]