Workflow
Gemma
icon
Search documents
X @Decrypt
Decrypt· 2026-04-02 18:19
Google Jumps Back Into the Open Source AI Race With Gemma 4https://t.co/GIU89CEGjQ ...
X @Decrypt
Decrypt· 2026-04-02 18:04
Google Jumps Back Into the Open Source AI Race With Gemma 4https://t.co/lC2FjoBLY4 ...
X @Demis Hassabis
Demis Hassabis· 2026-04-02 16:08
Available now under Apache 2.0 license in @GoogleAIStudio or download the model weights from @HuggingFace, @Kaggle and @Ollama. 400M downloads and 100K variants to date, Gemma goes from strength to strength. More info: https://t.co/hCHFTpoQJ9 ...
美国开源AI最后的旗帜,也倒了
量子位· 2026-03-30 01:34
Core Viewpoint - The Allen Institute for Artificial Intelligence (AI2) is significantly reducing funding for open-source model development, including OLMo, and shifting focus towards AI applications, which has led to the departure of key personnel to Microsoft [1][27][39]. Group 1: Personnel Changes - Key members of AI2, including former CEO Ali Farhadi and COO Sophie Leibrecht, have left to join Mustafa Suleyman's superintelligence team at Microsoft [2][10]. - Farhadi's departure marks the end of over two and a half years of leadership at AI2, where he was instrumental in the development of various AI projects [11][13]. - Other notable departures include Hannah Hajishirzi and Ranjay Krishna, both of whom were involved in significant AI initiatives at AI2 [3][19]. Group 2: Funding and Strategic Shifts - AI2's board chairman, Bill Hilf, indicated that the organization struggles to compete with tech giants like OpenAI and Google, which invest billions in training advanced models [27][28]. - The current funding model, primarily supported by the Paul G. Allen Family Foundation, is shifting from annual funding to a project proposal-based model, which may limit AI2's ability to pursue long-term open-source projects [33][38]. - The estimated training cost for cutting-edge models like GPT-4 is between $100 million to $200 million, highlighting the financial challenges faced by non-profit organizations like AI2 [29][30]. Group 3: Impact on Open-Source AI - The reduction in AI2's commitment to open-source model development is seen as a significant setback for the open-source AI community, with many expressing concern over the future of open-source initiatives in the U.S. [39][41]. - AI2's OLMo series was recognized for its commitment to transparency and open-source principles, but the recent changes may undermine these efforts [42][46]. - The shift in focus towards AI applications rather than foundational model development could accelerate the gap between U.S. and Chinese open-source AI capabilities [58][65]. Group 4: Future Outlook - Despite the challenges, AI2's interim CEO, Peter Clark, has stated that the organization remains committed to its mission and ongoing collaborations, such as the OMAI project with NSF and Nvidia [52]. - The landscape of open-source AI is evolving, with U.S. companies increasingly adopting models from China, indicating a shift in the global open-source AI dynamics [64][66].
谷歌AI内存技术工程化失败?TurboQuant“横空出世”,科技圈呼“谷歌版DeepSeek”、“真实版Pied Piper”!华尔街“呵呵,抄底内存股”!
美股IPO· 2026-03-26 00:44
Core Viewpoint - Google's new AI memory compression technology, TurboQuant, claims to reduce cache memory usage by 6 times and improve performance by 8 times, causing significant market panic among storage giants like Micron Technology and SanDisk, which saw their stocks drop over 5% [1][2][6] Group 1: Market Reaction - Following the announcement of TurboQuant, the storage chip sector experienced a sharp decline, with the storage chip and hardware supply chain index dropping by 2.08% [2][6] - Major companies like SanDisk and Micron saw significant stock declines, with SanDisk dropping 6.5% at one point and Micron falling 4% [6] - The market's defensive reaction stemmed from concerns about long-term demand for storage hardware due to the potential reduction in physical memory procurement [6] Group 2: Analyst Perspectives - Analysts from Wall Street believe the market has overreacted to the news, suggesting that investors should consider buying memory stocks during the dip [4][8] - Despite impressive lab data showing significant compression efficiency, analysts argue that the actual impact of TurboQuant on storage demand may be overstated [4][9] - Lynx Equity Strategies questioned the "disruptive" nature of the technology, stating that advanced compression techniques are primarily aimed at alleviating computational bottlenecks and will not fundamentally alter memory and flash demand in the next three to five years [9] Group 3: Long-term Demand Outlook - Morgan Stanley highlighted that TurboQuant's impact is limited to the inference stage of key-value caching and does not affect model training tasks or high-bandwidth memory (HBM) [10] - The firm referenced the "Jevons Paradox," suggesting that increased efficiency often leads to lower usage costs, which can stimulate greater overall demand [11] - By significantly reducing the service cost per query, TurboQuant may enable models that were previously only feasible on expensive cloud clusters to be deployed locally, thus lowering the barriers for AI scalability [11]
DeepSeek、GPT、Qwen,所有大模型架构图都有,Karpathy:宝藏画廊!
机器之心· 2026-03-16 03:53
Core Insights - The large model landscape has become increasingly crowded with numerous models emerging rapidly, making it difficult to understand their architectures and innovations [2][3] - A significant gap exists in the availability of a clear visual representation of these models, despite the abundance of options [2] Summary by Sections - **Introduction to the Landscape**: The article highlights the rapid development of large models such as GPT, Llama, and others, noting the challenge in comprehending their diverse architectures [2] - **Creation of the LLM Architecture Gallery**: AI researcher Sebastian Raschka has created an online resource called the "LLM Architecture Gallery," which organizes and visualizes the architectures of mainstream large models [3][6] - **Content of the Gallery**: The gallery serves as a comprehensive directory of various models, ranging from those with millions to trillions of parameters, including notable names like Llama, DeepSeek, and Mistral [7] - **Model Cards**: Each model in the gallery is linked to a dedicated page that provides essential information such as architecture diagrams, key module designs, parameter sizes, and release dates, facilitating quick understanding for researchers [11][14] - **Utility for Researchers**: The gallery acts as a quick reference index for model architectures, allowing users to compare different designs and innovations efficiently, thus aiding in understanding the evolution of technology [14]
芯原股份20260311
2026-03-12 09:08
Summary of Chipone's Conference Call Company Overview - **Company**: Chipone Technology Co., Ltd. (芯原股份) - **Industry**: AI ASIC (Application-Specific Integrated Circuit) Key Points IPO and Fundraising Strategy - Chipone plans to issue up to 15% of its shares in Hong Kong, including a green shoe option, to establish an international financing platform for global marketing, key technology R&D, and potential strategic acquisitions [2][3] - The funds raised will be allocated to four main areas: R&D for key technologies, global marketing network development, strategic investments and acquisitions, and general operational funding [3][4] Market Trends and ASIC Demand - AI computing demand is shifting from training to inference and fine-tuning, with ASICs expected to see a fivefold increase in capital expenditure by 2027 [2][4] - The demand for ultra-low power small models is surging in edge AI applications, such as AR glasses and robotics, with Chipone focusing on 270M parameter-level ultra-small model applications [2][4] Design-Light Model - Chipone is promoting a "Design-Light" model to reduce operational costs for chip companies by providing IP and design services, addressing the challenges of high operational costs in traditional Fabless models [2][5] AI Development Trends - The AI industry is transitioning from a focus on large models to a combination of large and small models, emphasizing the importance of physical world interaction and understanding [6][10] - The emergence of new applications, such as AR glasses and autonomous driving, is driving the need for specialized AI chips [6][10] Collaboration with Google - Chipone is deeply integrated with Google's open-source ecosystem, supporting projects like Gemma to secure commercial opportunities and strengthen its IP position in video coding and AIGC [2][9] Domestic Market Dynamics - The domestic autonomous driving chip market is expected to see over 50% penetration of domestic solutions within three years, with Chipone addressing differentiated computing needs through a Chiplet model [2][12] ASIC Market Growth - The ASIC market is experiencing significant growth, with major players like Broadcom and Marvell reporting substantial revenue increases, indicating ASICs' critical role in the current AI wave [4][10] Future of AI and AIGC - The future of AI should not solely focus on large models; small models are equally important and can derive capabilities from larger models [10][17] - The growth of edge AI applications will drive demand for efficient, low-power AI chips, positioning Chipone favorably in the AIGC landscape [17] Conclusion - Chipone's strategic initiatives, including its IPO, focus on R&D, and collaboration with major tech players, position it well to capitalize on the growing demand for AI and ASIC technologies in various applications. The emphasis on both edge and cloud AI solutions reflects a comprehensive approach to market opportunities.
Google Research and Synaptics Launch Next-Generation Coral Dev Board for Developers to Bring Multimodal Edge AI Applications to Life
Globenewswire· 2026-03-10 07:00
Core Insights - Synaptics Incorporated has launched a limited-edition Coral Dev Board powered by the Astra™ SL2610 product line, featuring the industry's first implementation of the Coral NPU from Google Research, aimed at enhancing power-efficient AI experiences across various applications [1][4] Product Features - The Coral Dev Board is designed for ultra-low power, always-on applications, enabling efficient on-device inference for battery-constrained devices, and includes hardware interfaces such as camera and display support, microphone inputs, and Wi-Fi/Bluetooth connectivity [2] - The board is targeted at AI and ML engineers, system architects, and ODMs/OEMs, providing an open environment for experimentation and rapid prototyping [2] Software and Development Tools - Synaptics' MLIR-based Torq open-source toolchain supports popular machine learning frameworks, facilitating a unified development experience from model optimization to deployment [3] - The combination of the Coral NPU and Synaptics Torq toolchain offers a robust foundation for building private and efficient Edge AI applications [3] Partnerships and Ecosystem - Synaptics is collaborating with Grinn Global and RS to launch the Coral Dev Board, which will come pre-configured with an Edge AI experience featuring the Gemma 3 270M model, allowing immediate development of on-device AI workloads [4] - The partnership aims to provide a clear path from evaluation to scalable deployment, combining production-ready hardware with rapid prototyping capabilities [4] Market Position and Future Plans - Synaptics emphasizes its commitment to building an open, developer-first Edge AI ecosystem, with plans for future development boards in collaboration with Google Research [5] - The company aims to meet the demand for accessible, production-grade platforms that facilitate immediate development for Edge AI applications [5][6]
X @Avi Chawla
Avi Chawla· 2026-03-05 20:00
RT Avi Chawla (@_avichawla)You're in a Research Scientist interview at DeepMind.The interviewer asks:"Our investors want us to contribute to open-source.Gemini crushed benchmarks.But we'll lose competitive edge by open-sourcing it.What to do?"You: "Release a research paper."Here's what you missed:LLMs today don't just learn from raw text; they also learn from each other.For example:- Llama 4 Scout & Maverick were trained using Llama 4 Behemoth.- Gemma 2 and 3 were trained using Gemini.Distillation helps us ...
“谷歌天团”反击AI泡沫质疑:这是工业革命,但速度快10倍、规模大10倍
美股IPO· 2026-02-20 14:57
Core Insights - Google Cloud's backlog orders have doubled to $240 billion, demonstrating the justification for high capital expenditures [1][4][7] - The current AI wave is compared to a "10x faster industrial revolution," indicating significant potential for growth and value creation [6][30] - DeepMind's CEO predicts that achieving Artificial General Intelligence (AGI) will take at least 5 to 10 more years [8][33] Group 1: AI Investment and Market Position - Google executives addressed concerns about the return on investment in AI, emphasizing that the current moment is akin to a major infrastructure investment period [6][30] - The company views its investment in AI as a transformative opportunity that will enhance various sectors, including cloud services, search, YouTube, and emerging businesses [7][30] - India is being positioned not just as a market but as a "full-stack player" in the AI field, reflecting its potential to contribute across all layers of AI infrastructure and applications [10][11] Group 2: Employment and Economic Impact - The discussion highlighted the distinction between "tasks" and "jobs," suggesting that while some jobs may decline, many will evolve or be created [9][28] - AI is seen as a technology that can empower small and medium-sized enterprises (SMEs), enabling them to leverage advanced tools without needing to be tech experts [36][37] - The potential for AI to fundamentally change workflows in various industries, including healthcare and education, was emphasized as a significant opportunity for India [39] Group 3: AGI and Future Prospects - DeepMind's CEO set a high standard for AGI, stating that it must exhibit all human cognitive abilities, and acknowledged that current systems are not yet at that level [8][33] - The use of AI tools like AlphaFold is already impacting scientific research, with over 300 million researchers globally, including 200,000 in India, utilizing these advancements [8][24] - The future of AI is seen as a means to accelerate scientific discovery and address global challenges, with a focus on ensuring equitable benefits from these technologies [34][35]