Workflow
深度学习
icon
Search documents
Cell重磅:人类首次通过“虚拟细胞”捕捉到生命最基本过程——细胞分裂
生物世界· 2026-04-01 01:00
Core Insights - The article discusses the concept of AI Virtual Cells (AIVC), which are advanced digital systems that integrate multimodal data, AI algorithms, and biological mechanisms to create predictive and controllable models of cellular behavior [1][4] - A significant study published in the journal Cell established a 4D whole-cell model that accurately simulates the lifecycle of the JCVI-syn3A bacterium, demonstrating the potential of virtual cells in understanding life processes at a molecular level [1][4] Group 1: AI Virtual Cells - AI Virtual Cells are not merely computer simulations but are dynamic digital systems that can replicate real physiological states and predict cellular responses to various interventions [1][4] - The recent research successfully simulated the entire lifecycle of a minimal bacterium, with the cell division process taking 105 minutes, closely matching the actual division time of the bacterium [1][4] Group 2: Applications and Implications - The development of virtual cells provides new tools for understanding how life emerges from molecular interactions, potentially revolutionizing fields such as drug discovery and synthetic biology [4] - The article outlines various applications of AI in biological research, including protein design, antimicrobial peptide design, computer-aided drug design (CADD), and artificial intelligence-driven drug discovery (AIDD) [5][31][34][35] Group 3: Educational Offerings - The article lists several advanced courses related to AI applications in biology, including topics on virtual cell construction, protein design, and synthetic biology, aimed at equipping participants with practical skills and theoretical knowledge [5][31][34][35] - Each course is designed to provide a comprehensive understanding of the respective fields, combining theoretical foundations with hands-on practical sessions [31][34][35] Group 4: Course Benefits and Promotions - The article mentions promotional offers for course registrations, including discounts for multiple course sign-ups and additional benefits such as access to past course recordings and preparatory materials [51][52] - Participants who complete the training and pass the examination can receive a certification that may serve as a valuable credential in their professional development [51]
谷歌前研究员‌:仅靠规模化无法实现AGI
Core Insights - François Chollet, a prominent figure in AI and the creator of Keras, emphasizes the importance of understanding AI as a tool for empowerment and encourages individuals to leverage AI knowledge to enhance their capabilities and navigate the ongoing transformation in various fields [2]. Group 1: Definition and Goals of AGI - François defines AGI as a system that can understand and master new problems with human-like efficiency and minimal training data, contrasting it with the automation of economic tasks [2]. - He predicts that the realization of AGI will first involve automating most economic work before achieving the more efficient learning definition he proposes [2]. Group 2: Limitations of Current AI Paradigms - The current reliance on deep learning and large language models (LLMs) is effective but not optimal, as it depends heavily on vast amounts of training data for pattern matching [2]. - In fields requiring formal verification of reward signals, such as coding and mathematics, current AI shows strong performance, while in less verifiable areas like writing, progress is slow or stagnant [2]. - François's research lab, NIA, aims to explore a fundamentally different AI research paradigm through program synthesis, focusing on high data efficiency and model optimality [2]. Group 3: Predictions on AGI Technology and Timeline - François believes that the "fluid intelligence engine" for AGI will be a compact codebase, potentially under 10,000 lines, but will require a vast knowledge base to operate effectively [3]. - He forecasts that AGI could be achieved around 2030, coinciding with the release of Arc-AGI versions 6 or 7, based on current progress and investment levels [3]. Group 4: Recommendations for Researchers and Entrepreneurs - François encourages diversification in AI research, suggesting that the current focus on LLMs is counterproductive and advocating for exploration of alternative paths like genetic algorithms and state space models [4]. - He highlights that a successful AI system must be capable of self-improvement and expansion without continuous direct intervention from human engineers, which is a core advantage of deep learning [4].
机器学习因子选股月报(2026年4月)-20260331
Southwest Securities· 2026-03-31 08:05
Quantitative Models and Construction GAN_GRU Model - **Model Name**: GAN_GRU - **Construction Idea**: The GAN_GRU model combines Generative Adversarial Networks (GAN) for feature generation and Gated Recurrent Units (GRU) for time-series feature encoding to create a stock selection factor[4][13][22] - **Construction Process**: 1. **GAN Component**: - **Generator**: Generates realistic data samples from random noise using the loss function: $$L_{G}\,=\,-\mathbb{E}_{z\sim P_{z}(z)}[\log(D(G(z)))]$$ where \(z\) represents random noise, \(G(z)\) is the generated data, and \(D(G(z))\) is the discriminator's output probability that the generated data is real[24][25][26] - **Discriminator**: Distinguishes real data from generated data using the loss function: $$L_{D}=-\mathbb{E}_{x\sim P_{data}(x)}[\log\!D(x)]-\mathbb{E}_{z\sim P_{z}(z)}[\log(1-D(G(z)))]$$ where \(x\) is real data, \(D(x)\) is the discriminator's output probability for real data, and \(D(G(z))\) is the output probability for generated data[27][29][30] - **Training Process**: Alternating training of the generator and discriminator until convergence[30][34] 2. **GRU Component**: - Two GRU layers (GRU(128,128)) followed by an MLP (256,64,64) to encode time-series features and predict future returns[22] - Input features include 18 price-volume metrics (e.g., closing price, turnover rate) sampled over 40 days to predict cumulative returns for the next 20 trading days[14][18][19] - Data preprocessing involves outlier removal, normalization, and cross-sectional standardization[18] - Training uses semi-annual rolling windows with hyperparameters such as batch size equal to the number of stocks, Adam optimizer, learning rate of \(1e-4\), and IC-based loss function[18][22] 3. **Feature Generation**: - GAN's generator processes raw price-volume time-series features (Input_Shape=(40,18)) and outputs transformed features with preserved time-series properties[37] - **Evaluation**: The model effectively combines GAN's feature generation capabilities with GRU's time-series encoding, providing robust predictive power for stock selection[4][22][37] --- Model Backtesting Results GAN_GRU Model Performance Metrics - **IC Mean**: 0.1095*** - **ICIR (Non-Annualized)**: 0.88 - **Turnover Rate**: 0.82X - **Recent IC**: 0.1008*** - **One-Year IC Mean**: 0.0514*** - **Annualized Return**: 36.03% - **Annualized Volatility**: 21.87% - **IR**: 1.55 - **Max Drawdown**: 27.29% - **Annualized Excess Return**: 21.87%[41][42][45] Industry-Level Performance - **Top 5 Industries by Recent IC**: - Media: 0.4279*** - Coal: 0.2355*** - Retail: 0.2003*** - Food & Beverage: 0.1701*** - Chemicals: 0.1395***[41][42][45] - **Top 5 Industries by One-Year IC Mean**: - Media: 0.1304*** - Steel: 0.1212*** - Retail: 0.1191*** - IT: 0.1064*** - Food & Beverage: 0.0988***[41][42][45] - **Top 5 Industries by Recent Excess Return**: - Media: 4.57% - Agriculture: 3.26% - Construction Materials: 3.19% - Light Manufacturing: 2.53% - Coal: 2.22%[45][46][48] - **Top 5 Industries by One-Year Average Excess Return**: - Real Estate: 1.83% - Retail: 1.41% - Consumer Services: 1.39% - Automotive: 1.18% - Utilities: 1.07%[45][46][48] --- Quantitative Factors and Construction GAN_GRU Factor - **Factor Name**: GAN_GRU - **Construction Idea**: Derived from the GAN_GRU model, this factor leverages GAN for feature generation and GRU for time-series encoding to predict stock returns[4][13][22] - **Construction Process**: - Input features include 18 price-volume metrics sampled over 40 days[14][18][19] - GAN generates transformed features while preserving time-series properties[37] - GRU encodes these features and outputs predicted returns as the factor[22][37] - Factor values undergo industry and market-cap neutralization and standardization[22] - **Evaluation**: The factor demonstrates strong predictive power across multiple industries and time periods, with significant IC values and excess returns[4][22][37] --- Factor Backtesting Results GAN_GRU Factor Performance Metrics - **IC Mean**: 0.1095*** - **ICIR (Non-Annualized)**: 0.88 - **Turnover Rate**: 0.82X - **Recent IC**: 0.1008*** - **One-Year IC Mean**: 0.0514*** - **Annualized Return**: 36.03% - **Annualized Volatility**: 21.87% - **IR**: 1.55 - **Max Drawdown**: 27.29% - **Annualized Excess Return**: 21.87%[41][42][45] Industry-Level Performance - **Top 5 Industries by Recent IC**: - Media: 0.4279*** - Coal: 0.2355*** - Retail: 0.2003*** - Food & Beverage: 0.1701*** - Chemicals: 0.1395***[41][42][45] - **Top 5 Industries by One-Year IC Mean**: - Media: 0.1304*** - Steel: 0.1212*** - Retail: 0.1191*** - IT: 0.1064*** - Food & Beverage: 0.0988***[41][42][45] - **Top 5 Industries by Recent Excess Return**: - Media: 4.57% - Agriculture: 3.26% - Construction Materials: 3.19% - Light Manufacturing: 2.53% - Coal: 2.22%[45][46][48] - **Top 5 Industries by One-Year Average Excess Return**: - Real Estate: 1.83% - Retail: 1.41% - Consumer Services: 1.39% - Automotive: 1.18% - Utilities: 1.07%[45][46][48]
突发|华为诺亚方舟实验室主任王云鹤离职
机器之心· 2026-03-28 04:45
Core Viewpoint - The departure of Wang Yunhe, the director of Huawei's Noah's Ark Lab, marks a significant shift in the AI industry, indicating a profound structural transformation within the sector since 2026 [3][25]. Group 1: Wang Yunhe's Background - Wang Yunhe, born in 1991, graduated with a Bachelor's degree in Mathematics from Xi'an University of Electronic Science and Technology and obtained his PhD in Intelligent Science from Peking University in 2018, focusing on deep learning, model compression, machine learning, and computer vision [5][8]. - He has over 8 years of experience at Huawei, starting as an intern at Noah's Ark Lab and progressing to roles such as Senior Engineer, Chief Engineer, and eventually the director of the lab [8][25]. Group 2: Contributions and Achievements - Wang has a notable academic record with over 33,000 citations on Google Scholar, highlighting his influence in the field of AI [13]. - His research includes the development of GhostNet, a lightweight neural network architecture that achieved a Top-1 accuracy of 75.7% on the ImageNet classification task, surpassing MobileNetV3 [15][16]. - He has contributed significantly to the Vision Transformer research, with his survey article receiving 5,528 citations, establishing it as a key reference in the field [18]. Group 3: Insights on AI Models - Wang has provided unique insights into the mainstream technology routes in the era of large models, discussing the potential impact of diffusion models on autoregressive models and emphasizing the need for structural thinking in model design [21]. - His recent work on the DLLM Agent explores how different generative paradigms affect agent planning and decision-making, demonstrating the efficiency of the proposed model in global planning and interaction [22][24]. Group 4: Industry Impact - Wang's departure from Huawei is a focal point for the industry, as he has led several internationally influential algorithm innovations during his tenure [25]. - His future career path, particularly regarding his thoughts on unifying architectures for diffusion language models and general artificial intelligence, remains a topic of interest for the industry [26].
《Science Robotics》重磅!毫瓦级超声波,让手掌大飞行机器人“穿越”浓雾、黑暗及复杂障碍环境
机器人大讲堂· 2026-03-26 11:05
Core Viewpoint - The article discusses the innovative use of ultrasonic sensors in a new perception system named "Saranga," developed by Worcester Polytechnic Institute, which allows drones to navigate in challenging environments where traditional sensors like cameras and LiDAR fail [1][3]. Group 1: Technology and Innovation - The Saranga system utilizes a milliwatt-level ultrasonic sensor suite instead of relying on cameras and LiDAR, enabling a palm-sized flying robot to autonomously navigate through fog, darkness, snow, and transparent obstacles [3][5]. - The research team drew inspiration from nature, particularly from the ability of small bats to use echolocation effectively in dark and dusty environments, leading to the realization that ultrasonic technology could be optimized for drone applications [5][6]. - The system employs a combination of physical noise reduction and deep learning techniques to enhance the detection capabilities of ultrasonic signals, effectively increasing the detection range from 1 meter to 2 meters [6][7]. Group 2: Experimental Validation - The team conducted extensive testing in various environments, including indoor and outdoor scenarios, using a custom quadcopter named PeARBat160 equipped with ultrasonic sensors [11][12]. - In tests involving transparent obstacles, the flying robot achieved a success rate of 77.27% in 22 trials, while it reached 80.95% success with thin obstacles [12][14]. - The system demonstrated a 90% success rate in dense fog conditions and a perfect 100% success rate in low-light environments, showcasing its robustness in adverse conditions [18][20]. Group 3: Comparative Analysis - Saranga was compared with another ultrasonic obstacle avoidance system, BatDeck, where Saranga significantly outperformed BatDeck, successfully navigating 13 out of 17 trials in the same complex indoor environment [28]. - The success rate of Saranga decreased from 100% to 72.73% as the target speed increased from 1 m/s to 2 m/s, indicating areas for further optimization [28]. Group 4: Industry Implications - The success of Saranga signals a need for the robotics industry to reconsider the reliance on mainstream sensors like cameras and LiDAR, advocating for a focus on the physical reliability of sensors in specific environments [30]. - The article emphasizes that integrating older sensors with new computational technologies, such as deep learning, can revitalize their effectiveness, suggesting a shift in sensor selection criteria for autonomous systems [30].
离职特斯拉“隐身”14个月,杨硕创业终于亮牌:重新定义机器人训练范式
量子位· 2026-03-24 23:52
Core Viewpoint - Yang Shuo, co-founder and CTO of Mondo Robotics, has remained silent since leaving Tesla's Optimus team over a year ago, but recently unveiled the company's work on a new model called DiT4DiT, which focuses on training robots using video to enhance their action capabilities and adaptability in various scenarios [1][2]. Group 1: DiT4DiT Model Overview - DiT4DiT is an end-to-end model that integrates video diffusion and action diffusion into a cascading framework for robot learning [9]. - The model employs a unique design called "intermediate denoising," which extracts key features during the video generation process to guide robot action decisions without waiting for a complete video output [11][12]. - The model's performance has been validated, achieving a 98.6% average success rate on the LIBERO benchmark, demonstrating its state-of-the-art capabilities [30]. Group 2: Key Design Features - The model's two critical designs include intermediate denoising and a three-timestep scheme, which allows for efficient training of both video generation and action prediction tasks [10][25]. - The intermediate denoising process involves extracting features from a specific layer during the denoising stages, optimizing the robot's ability to understand physical interactions rather than relying on complete video clarity [19][22]. - The three-timestep scheme enables the video model and action model to operate independently yet cohesively, improving convergence speed by 7 times and data efficiency by over 10 times [29]. Group 3: Practical Applications and Performance - DiT4DiT has been deployed on the Yuzhu G1 humanoid robot, successfully completing tasks such as flower arrangement and drawer interactions, outperforming pre-trained models and demonstrating superior deployment potential on robot edge chips [41][42][43]. - The model's design allows it to adapt quickly to new objects and scenarios, addressing limitations of traditional visual-language-action models that struggle with dynamic physical understanding [36][40].
选股择时与多资产轮动的统一框架:深度学习系列之二:绝对收益视角下的技术形态专家模型
Soochow Securities· 2026-03-24 11:41
Core Insights - The report presents a deep learning model based on Gated Recurrent Units (GRU) for technical analysis, which demonstrates robust capabilities in stock selection and timing across multiple asset classes, achieving significant excess returns [1][10][11]. Group 1: Model Performance - The model shows a mean Information Coefficient (IC) of 9.14% in cross-sectional stock selection from 2018 to 2026, with an annualized excess return of 10.73% relative to an equal-weighted benchmark [1]. - In time-series timing, the model achieves annualized excess returns ranging from 15.94% to 19.92% when applied to the CSI All Share Index, with a drawdown ratio between 0.75 and 0.89 [1][3]. - The model's zero-sample inference capability is validated as it successfully predicts patterns not seen in the training data, indicating its generalizability [1][3]. Group 2: Asset Allocation Strategies - The model achieves significant excess returns in style rotation, industry rotation, and ETF rotation, with the ETF rotation strategy yielding an annualized excess return of 16.56% [2][3]. - The industry rotation strategy shows an annualized excess return of 12.60% with a drawdown ratio of 2.12, while the maximum drawdown is controlled within -5.95% [3]. - The model's adaptability across different investment scenarios highlights its robustness and provides a new technical pathway for quantitative investing [3]. Group 3: Technical Analysis Framework - The report emphasizes the limitations of traditional technical analysis, which relies heavily on manually defined patterns and is susceptible to market noise [10][11]. - The GRU model automates the extraction of K-line features and integrates both cross-sectional and time-series capabilities, overcoming the limitations of traditional methods [10][11][12]. - The model's architecture allows for multi-period information fusion, enhancing the robustness of trading decisions by leveraging features from different time frames [12][13].
英伟达首台DGX GB300,老黄亲自登门送给他
量子位· 2026-03-19 07:09
Core Viewpoint - The article discusses the significance of NVIDIA's CEO Jensen Huang personally delivering the first DGX Station (GB300) to Andrej Karpathy, highlighting the rise of individual developers in the AI era and the importance of computational power in the ongoing AI model competition [1][9][58]. Group 1: Delivery of DGX Station - Huang's delivery of the DGX Station to Karpathy symbolizes a milestone in the AI era, marking the emergence of personal developers as key players [1][9]. - This event is reminiscent of Huang's previous deliveries, such as the first DGX-1 to OpenAI, which played a crucial role in the deep learning revolution [8][39]. - The DGX Station (GB300) is designed for individual developers, providing data center-level AI computing power in a compact form [28][30]. Group 2: Significance of Individual Developers - Karpathy is recognized as a representative of individual developers, transforming AI from a corporate domain to a system manageable by individuals [17][19]. - His recent work focuses on creating systems that allow a single person to complete the entire process from idea to product [18][19]. - The choice of Karpathy for this delivery underscores the shift towards distributed computing and the importance of individual contributions in the AI landscape [58][61]. Group 3: Technical Specifications of DGX Station - The DGX Station (GB300) features 748GB of unified memory and 20 PFLOPS of computing power, enabling the execution of large-scale AI models [30]. - It allows seamless migration of local projects to cloud environments, addressing the need for continuous AI operation [31][32]. - The system is tailored for developing and running AI agents, reflecting the growing trend of personal AI applications [24][34]. Group 4: Broader Implications for the Industry - Huang's actions signal a strategic move by NVIDIA to position itself as a foundational supplier in the AI model competition, emphasizing the necessity of computational resources [50][56]. - The article suggests that the future of AI development will increasingly rely on individual developers rather than large organizations, as computational power becomes more accessible [58][61]. - NVIDIA is also enhancing its infrastructure for AI agents, indicating a comprehensive approach to support developers from hardware to software [34][36].
Cell:中国学者开发AI药物发现与设计平台GPS,一作已回国加入临港实验室
生物世界· 2026-03-18 04:37
Core Viewpoint - The article discusses the development of a deep learning-based drug discovery platform called GPS, which utilizes transcriptomic features to identify and optimize compounds for reversing disease-associated transcriptional phenotypes, marking a significant advancement in drug discovery methodologies [3][10][18]. Group 1: Research Background - Current virtual drug screening primarily relies on docking against specific protein targets or AI/ML models trained on screening data, with limited use of transcriptomics, particularly single-cell RNA sequencing, to characterize diseases and cellular states [2]. - The identification of drugs that reverse disease-related transcriptomic features has been explored as a strategy for discovering new uses for existing drugs, but this approach is limited to compounds already in databases and does not support the screening and optimization of novel compounds [2]. Group 2: Research Breakthrough - A research team from Michigan State University and other institutions developed the GPS platform, which screens large compound libraries based on transcriptomic features [3][9]. - The GPS platform predicts the impact of chemical structures on gene expression, allowing for the identification of compounds that can reverse disease-associated gene expression patterns [10][11]. Group 3: Methodology - The GPS platform operates in three key steps: predicting gene expression changes from chemical structures, calculating a "reversal score" to assess the potential of compounds to reverse disease signatures, and optimizing promising compounds using a Monte Carlo tree search algorithm [13][14]. - The research team trained a deep learning model using extensive drug-gene expression data from the LINCS database, enhancing prediction accuracy through a robust collaborative learning framework [11]. Group 4: Applications and Findings - In hepatocellular carcinoma research, the team identified a lead compound with an IC50 value of approximately 4μM against liver cancer cell lines, which was further optimized to achieve sub-micromolar activity [14]. - For idiopathic pulmonary fibrosis (IPF), the team discovered that the existing drug Pyrithyldion could effectively reverse gene expression features associated with IPF, and identified a novel compound, Drug 18, which significantly reduced key fibrosis markers in patient samples [15]. Group 5: Significance - The GPS platform represents a paradigm shift in drug discovery by focusing on disease gene expression features rather than relying solely on known protein targets or limited phenotypic screening data [18]. - This approach allows for the exploration of a vast chemical space and the discovery of novel mechanisms of action, potentially leading to more effective and personalized treatments for diseases like liver cancer and IPF [18].
量化选股策略周报:指增组合年内超额收益悉数转正
CAITONG SECURITIES· 2026-03-15 07:30
Market Performance - As of March 13, 2026, the Shanghai Composite Index decreased by 0.70%, while the Shenzhen Component Index increased by 0.76%[9] - The CSI 300 Index rose by 0.19% during the same period[9] - The ChiNext Index showed better performance amidst market fluctuations, with a weekly increase of 2.51%[10] Enhanced Index Fund Performance - For the CSI 300 enhanced index fund, the minimum excess return was -2.19%, the median was -0.01%, and the maximum was 0.88% for the week[13] - The CSI 500 enhanced index fund had a minimum excess return of -0.59%, a median of 0.80%, and a maximum of 2.92%[13] - The CSI 1000 enhanced index fund reported a minimum excess return of -0.53%, a median of 0.34%, and a maximum of 1.25%[13] Year-to-Date Performance - As of March 13, 2026, the CSI 300 Index increased by 0.8%, while the CSI 300 enhanced portfolio rose by 2.9%, resulting in an excess return of 2.1%[21] - The CSI 500 Index increased by 10.4%, with the enhanced portfolio rising by 11.2%, yielding an excess return of 0.8%[26] - The CSI A500 Index rose by 3.6%, while the enhanced portfolio increased by 5.4%, resulting in an excess return of 1.8%[33] - The CSI 1000 Index increased by 8.1%, with the enhanced portfolio rising by 8.2%, yielding an excess return of 0.1%[39] Risk Considerations - There are risks associated with factor failure, model failure, and changes in market style that could impact performance[5][44]