深度学习模型
Search documents
国际最新研发深度学习模型:可预测DNA变异影响助力开发新疗法
Zhong Guo Xin Wen Wang· 2026-01-30 06:15
Core Insights - The article discusses the development of a deep learning model called AlphaGenome by Google's research team, which can predict the functional impact of DNA sequence variations up to 1 million base pairs long [1][3]. Group 1: Model Capabilities - AlphaGenome is designed to predict how DNA sequence variations affect various biological processes, aiding in the understanding of genetic diseases and improving gene testing [1][3]. - The model has been trained on human and mouse genomes to learn how DNA sequences influence different biological outcomes, allowing it to predict 5,930 human and 1,128 mouse genetic signals related to specific functions such as gene expression, splicing, and protein modification [3][4]. - In evaluations of 26 variant effect predictions, AlphaGenome performed comparably or better than existing top models in 25 cases, showcasing its ability to make multiple predictions across various genetic signals and biological outcomes [3]. Group 2: Future Applications - The research team suggests that further improvements to AlphaGenome could expand its applications, such as increasing the range of species covered or enhancing the model's ability to identify non-coding sequences [4]. - There is potential for AlphaGenome to deepen the understanding of complex biological outcomes resulting from DNA sequence variations in the future [4].
图生视频工具在跨境电商中的应用与技术解析
Sou Hu Cai Jing· 2026-01-22 16:22
Core Insights - The article discusses the rise of generative video tools that utilize AI technology to convert static images into dynamic videos, which have become essential for enhancing product display in the rapidly growing cross-border e-commerce sector [1][6] - These tools help merchants reduce video production costs and improve content output efficiency, allowing them to meet diverse marketing needs across multiple platforms and regions [1][6] Group 1: Technology and Features - Generative video tools automate image processing to create smooth video content, employing technologies such as Generative Adversarial Networks (GAN), deep learning models, and natural language processing [1][6] - The tools can intelligently add motion effects, transition animations, and background music, and even support multilingual voiceovers, making videos more engaging and localized [1][6] - Most tools operate on cloud-based processing, enabling users to quickly output videos suitable for various scenarios by simply uploading images and making basic settings [1][6] Group 2: Key Players - Keevx focuses on providing efficient video generation services for cross-border e-commerce, enabling merchants to quickly create virtual model showcase videos for product detail pages, platform ads, and social media marketing [2] - Runway ML is another notable tool that allows users to convert static images into dynamic videos using advanced machine learning models, offering high-quality output and flexibility for users with technical backgrounds [2] - Canva integrates generative video functionality into its graphic design platform, allowing users to create videos easily through a user-friendly interface, making it particularly suitable for small to medium-sized cross-border e-commerce merchants [4] Group 3: Market Impact - Overall, generative video tools lower the barriers to video production for cross-border e-commerce, enhancing marketing efficiency and enabling merchants to vividly showcase products [6] - These tools foster user engagement through localized and personalized content, and as AI technology continues to advance, these tools are expected to become more intelligent and integrated, offering greater possibilities for the global e-commerce ecosystem [6]
深度学习模型可预测细胞每分钟发育变化 为构建“数字胚胎”奠定基础
Ke Ji Ri Bao· 2025-12-26 00:37
Core Insights - A collaborative team from MIT, the University of Michigan, and Northeastern University has introduced a geometric deep learning model named "MultiCell," which predicts cellular behavior during fruit fly embryonic development at single-cell resolution [1][2] - The model utilizes four-dimensional whole-embryo data with sub-micron resolution and high frame rates, containing approximately 5,000 labeled cell boundaries and nuclei [1] - "MultiCell" is the first algorithm capable of predicting various cellular behaviors with single-cell precision during multicellular self-assembly, showing potential for early diagnosis and drug screening [2] Group 1 - The "MultiCell" model can predict the behavior changes of each cell every minute during the embryonic development process [1] - The model achieved about 90% accuracy in predicting cell connection loss and demonstrated high accuracy in predicting cell invagination, division, or rearrangement behaviors [2] - The method is compared to AlphaFold, which predicts protein structures from amino acid sequences, highlighting the complexity of embryonic development compared to protein folding [1] Group 2 - The model was trained on three embryonic videos and then applied to predict the evolution of a fourth new embryo [2] - Future enhancements may include integrating gene expression and protein localization data to provide a more comprehensive understanding of the interaction between physical and biological information [2] - The development of a universal multicellular developmental prediction model could lead to the creation of "digital embryos" for drug screening and guiding artificial tissue design [1]
——量化学习笔记之一:基于堆叠LSTM模型的十年期国债收益率预测
EBSCN· 2025-12-15 07:56
1. Report Industry Investment Rating No relevant content provided. 2. Core View of the Report The report systematically reviews the evolution of financial time - series forecasting models and constructs a prediction model for China's 10 - year treasury bond yield using a long - short - term memory (LSTM) neural network with historical time series as the single input variable, initially exploring the application of this deep - learning model in the fixed - income quantitative field [10]. 3. Summary by Relevant Catalog 3.1 Financial Time - Series Forecasting and Neural Network Models 3.1.1 Evolution of Financial Time - Series Forecasting Models Financial time - series forecasting has gone through three main development stages: traditional econometric models, traditional machine - learning models, and deep - learning models. Traditional econometric models have clear forms and strong interpretability but struggle to depict nonlinear and complex dynamic relationships. Traditional machine - learning models can perform nonlinear fitting and automatic feature screening but need manual feature extraction. Deep - learning models can automatically extract features from raw data and capture complex long - term time - series patterns, adapting well to the complex characteristics of financial time series [11][12]. 3.1.2 Neural Network Models and LSTM Models Neural network models are machine - learning models imitating the connection structure of human brain neurons. Recurrent neural networks (RNN) and their variants, such as LSTM, are designed for processing sequence data. LSTM solves the long - term dependence problem of traditional RNN through a "gating mechanism" and memory units, enhancing robustness to irregular data and being suitable for bond yield prediction [13][18]. 3.2 Treasury Bond Yield Prediction Based on Stacked LSTM Model 3.2.1 Stacked LSTM Model Stacked LSTM connects multiple LSTM layers in sequence, having advantages in long - sequence processing and multi - dimensional feature extraction, more suitable for complex time - series forecasting in financial scenarios [23]. 3.2.2 Construction of Treasury Bond Yield Prediction Model The report uses a classic and robust architecture of three - layer stacked LSTM + Dropout regularization to build a neural network model for predicting the 10 - year treasury bond yield. It only uses the historical time series of the 10 - year treasury bond yield as a single variable for prediction. The data is from the beginning of 2021 to December 12, 2025. After data processing and sample construction, a medium - complexity LSTM neural network model with about 130,000 adjustable parameters is built. The optimal model is obtained at the 27th training iteration, with an average absolute error of 1.43BP for the test set. The predicted yield on December 19, 2025, is 1.8330%, slightly lower than 1.8396% on December 12, 2025 [2][24][30]. 3.3 Follow - up Optimization Directions - Optimize model design: Adjust and optimize the design related to time windows, data processing, network architecture, and training strategies [3][36]. - Input multi - dimensional variables: Expand input variables from a single yield sequence to multi - dimensional variables such as macro, market, and sentiment to make the model more in line with economic logic and capture more comprehensive information [3][36]. - Build hybrid models: Combine the LSTM model with traditional econometric models or other machine - learning models to build hybrid models like ARIMAX - LSTM and CNN - LSTM - ATT, enhancing prediction accuracy [3][36]. - Introduce a rolling back - testing mechanism: Use a rolling time - window back - testing mechanism to update the model dynamically and make continuous predictions, improving the model's adaptability to market changes [3][36].
量化学习笔记之一:基于堆叠LSTM模型的十年期国债收益率预测
EBSCN· 2025-12-15 06:53
1. Report Industry Investment Rating No relevant information provided. 2. Core View of the Report The report systematically reviews the evolution of financial time - series prediction models and constructs a prediction model for China's 10 - year Treasury bond yield using the Long Short - Term Memory (LSTM) neural network, with historical time - series as the single input variable, to explore the application of this deep - learning model in the fixed - income quantitative field [10]. 3. Summary by Directory 3.1 Financial Time - Series Prediction and Neural Network Models - **Evolution of Financial Time - Series Prediction Models**: Financial time - series prediction has gone through three main stages: traditional econometric models (e.g., ARIMA, GARCH), traditional machine - learning models (e.g., SVM, RF), and deep - learning models. Traditional econometric models have clear forms but struggle with nonlinear and complex dynamic relationships. Traditional machine - learning models can perform nonlinear fitting but need manual feature extraction. Deep - learning models can automatically extract features and capture long - term patterns, with RNN and its variants like LSTM being mainstream methods [11][12]. - **Neural Network Models and LSTM Models**: Neural network models mimic the human brain's neuron connection structure. RNN is designed for sequence data but has issues with long - term memory. LSTM solves the long - term dependency problem of RNN through a "gating mechanism" and memory units, enhancing robustness to irregular data. It is suitable for bond yield prediction due to its ability to handle long - term time series and filter noise [13][17][18]. 3.2 Treasury Bond Yield Prediction Based on Stacked LSTM Model - **Stacked LSTM Model**: Stacked LSTM connects multiple LSTM layers sequentially, offering advantages in long - sequence processing and multi - dimensional feature extraction, making it more suitable for complex financial time - series prediction [23]. - **Construction of Treasury Bond Yield Prediction Model**: - **Data Processing and Sample Construction**: The data is the yield of the 10 - year Treasury bond from the beginning of 2021 to December 12, 2025. First - order differences are calculated and standardized. Samples are constructed with the first - order differences of the past 60 trading days as input features and the first - order differences of the next week as the prediction target. The samples are divided into a training set (72%), a validation set (8%), and a test set (20%) [27]. - **Model Design and Evaluation**: The model architecture consists of LSTM, Dropout, and Dense layers. The training strategy involves 200 iterations with an early - stopping mechanism. Evaluation metrics include MSE, MAE, and RMSE [28][29]. - **Model Results**: A medium - complexity LSTM neural network model with about 130,000 adjustable parameters is built. The optimal model is obtained at the 27th iteration, and the early - stopping mechanism is triggered at the 77th iteration. The average absolute error for the test set is 1.43BP. The 10 - year Treasury bond yield is predicted to decline from December 15 - 19, 2025, with the predicted value on December 19, 2025, being 1.8330%, slightly lower than 1.8396% on December 12, 2025 [30]. 3.3 Follow - up Optimization Directions - **Model Design Optimization**: Adjust and optimize relevant designs such as time windows, data processing, network architecture, and training strategies [3][36]. - **Input Multi - Dimensional Variables**: Expand input variables from a single yield sequence to multi - dimensional variables such as macroeconomic, market, and sentiment variables to make the model more in line with economic logic and capture more comprehensive information [3][36]. - **Construct Hybrid Models**: Combine the LSTM model with traditional econometric models or other machine - learning models to build hybrid models like ARIMAX - LSTM and CNN - LSTM - ATT, leveraging different model advantages and improving prediction accuracy [3][36]. - **Introduce Rolling Back - testing Mechanism**: Use a rolling time - window back - testing mechanism to update the model dynamically and make continuous predictions, enhancing the model's adaptability to market changes [3][36].
AI文章仿写工具哪个好?深度评测帮你选
Sou Hu Cai Jing· 2025-12-14 16:14
Core Insights - The article discusses the need for a comprehensive tool that automates the entire content creation process, from collection to publication, addressing the limitations of existing AI writing tools that often serve single functions [1][2] - It evaluates several mainstream "AI-generated article imitation" tools based on their automation, functionality, originality, publication flexibility, and cost-effectiveness [2] Group 1: Tool Evaluations - **First Place: Youcaiyun AI Content Factory** - Scoring 9.8/10, it offers a complete content production pipeline, including article collection, intelligent filtering, deep originality/rewrite, and automated publication, designed to meet the needs of website owners and content operators [4][6] - **Second Place: Zhixie Workshop** - Scoring 8.5/10, it excels in creative writing and deep imitation, particularly for literary texts, but lacks built-in content collection and automated publication capabilities, making it suitable for individual creators or small studios [7] - **Third Place: Xuncaitong** - Scoring 7.9/10, it has strong web information scraping and aggregation capabilities, but its rewriting function is basic and requires manual proofreading, limiting its effectiveness for high-quality SEO optimization [8][10] - **Fourth Place: Yigaojingling** - Scoring 7.0/10, it is a lightweight tool for quick generation of draft content, but its simplicity and lack of advanced features make it less suitable for teams with high-quality content needs [11] Group 2: Industry Trends - The evolution of text generation technology has progressed from simple template filling to deep semantic understanding and creative imitation, with modern large language models achieving over 70% vocabulary and sentence structure variation while retaining factual information [2] - The article emphasizes the importance of selecting a tool that integrates into a complete workflow rather than standalone features, highlighting the growing homogeneity in AI content creation tools [12]
中邮因子周报:深度学习模型回撤显著,高波占优-20250901
China Post Securities· 2025-09-01 05:47
Quantitative Models and Construction 1. Model Name: barra1d - **Model Construction Idea**: This model is part of the GRU factor family and is designed to capture short-term market dynamics through daily data inputs[4][6][8] - **Model Construction Process**: The barra1d model uses daily market data to calculate factor exposures and returns. It applies industry-neutralization and standardization processes to ensure comparability across stocks. The model is rebalanced monthly, selecting the top 10% of stocks with the highest factor scores for long positions and the bottom 10% for short positions, with equal weighting[17][28][29] - **Model Evaluation**: The barra1d model demonstrated strong performance in multiple stock pools, showing resilience in volatile market conditions[4][6][8] 2. Model Name: barra5d - **Model Construction Idea**: This model extends the barra1d framework to a five-day horizon, aiming to capture slightly longer-term market trends[4][6][8] - **Model Construction Process**: Similar to barra1d, the barra5d model uses five-day aggregated data for factor calculation. It follows the same industry-neutralization, standardization, and rebalancing processes as barra1d[17][28][29] - **Model Evaluation**: The barra5d model experienced significant drawdowns in recent periods, indicating sensitivity to market reversals[4][6][8] 3. Model Name: open1d - **Model Construction Idea**: This model focuses on open price data to identify short-term trading opportunities[4][6][8] - **Model Construction Process**: The open1d model calculates factor exposures based on daily opening prices. It applies the same industry-neutralization and rebalancing methodology as other GRU models[17][28][29] - **Model Evaluation**: The open1d model showed moderate performance, with some drawdowns in recent periods[4][6][8] 4. Model Name: close1d - **Model Construction Idea**: This model emphasizes closing price data to capture end-of-day market sentiment[4][6][8] - **Model Construction Process**: The close1d model uses daily closing prices for factor calculation. It follows the same construction and rebalancing methodology as other GRU models[17][28][29] - **Model Evaluation**: The close1d model demonstrated stable performance, with positive returns in certain stock pools[4][6][8] --- Model Backtesting Results 1. barra1d Model - Weekly Excess Return: +0.57%[29][30] - Monthly Excess Return: +0.75%[29][30] - Year-to-Date Excess Return: +4.38%[29][30] 2. barra5d Model - Weekly Excess Return: -2.17%[29][30] - Monthly Excess Return: -3.76%[29][30] - Year-to-Date Excess Return: +4.13%[29][30] 3. open1d Model - Weekly Excess Return: -0.97%[29][30] - Monthly Excess Return: -2.85%[29][30] - Year-to-Date Excess Return: +4.20%[29][30] 4. close1d Model - Weekly Excess Return: -1.68%[29][30] - Monthly Excess Return: -4.50%[29][30] - Year-to-Date Excess Return: +1.90%[29][30] --- Quantitative Factors and Construction 1. Factor Name: Beta - **Factor Construction Idea**: Measures historical market sensitivity of a stock[15] - **Factor Construction Process**: Calculated as the regression coefficient of a stock's returns against market returns over a specified period[15] 2. Factor Name: Size - **Factor Construction Idea**: Captures the size effect, where smaller firms tend to outperform larger ones[15] - **Factor Construction Process**: Defined as the natural logarithm of total market capitalization[15] 3. Factor Name: Momentum - **Factor Construction Idea**: Identifies stocks with strong recent performance[15] - **Factor Construction Process**: Combines historical excess return mean, volatility, and cumulative deviation into a weighted formula: $ Momentum = 0.74 * \text{Volatility} + 0.16 * \text{Cumulative Deviation} + 0.10 * \text{Residual Volatility} $[15] 4. Factor Name: Volatility - **Factor Construction Idea**: Measures the risk or variability in stock returns[15] - **Factor Construction Process**: Weighted combination of historical residual volatility and other measures[15] 5. Factor Name: Valuation - **Factor Construction Idea**: Captures the value effect, where undervalued stocks tend to outperform[15] - **Factor Construction Process**: Defined as the inverse of the price-to-book ratio[15] 6. Factor Name: Liquidity - **Factor Construction Idea**: Measures the ease of trading a stock[15] - **Factor Construction Process**: Weighted combination of turnover rates over monthly, quarterly, and yearly horizons: $ Liquidity = 0.35 * \text{Monthly Turnover} + 0.35 * \text{Quarterly Turnover} + 0.30 * \text{Yearly Turnover} $[15] 7. Factor Name: Profitability - **Factor Construction Idea**: Identifies stocks with strong earnings performance[15] - **Factor Construction Process**: Weighted combination of various profitability metrics, including analyst forecasts and financial ratios[15] 8. Factor Name: Growth - **Factor Construction Idea**: Captures the growth potential of a stock[15] - **Factor Construction Process**: Weighted combination of earnings and revenue growth rates[15] --- Factor Backtesting Results 1. Beta Factor - Weekly Return: +0.14%[21] - Monthly Return: +1.65%[21] - Year-to-Date Return: +5.29%[21] 2. Size Factor - Weekly Return: +0.36%[21] - Monthly Return: +1.00%[21] - Year-to-Date Return: +6.37%[21] 3. Momentum Factor - Weekly Return: +2.21%[24] - Monthly Return: +8.80%[24] - Year-to-Date Return: +23.30%[24] 4. Volatility Factor - Weekly Return: +2.82%[24] - Monthly Return: +12.29%[24] - Year-to-Date Return: +25.25%[24] 5. Valuation Factor - Weekly Return: +1.47%[21] - Monthly Return: +2.30%[21] - Year-to-Date Return: -2.26%[21] 6. Liquidity Factor - Weekly Return: +1.80%[21] - Monthly Return: +5.91%[21] - Year-to-Date Return: +19.70%[21] 7. Profitability Factor - Weekly Return: +4.57%[21] - Monthly Return: +7.53%[21] - Year-to-Date Return: +27.56%[21] 8. Growth Factor - Weekly Return: +2.76%[24] - Monthly Return: +6.51%[24] - Year-to-Date Return: +14.51%[24]
国债期货系列报告:多通道深度学习模型在国债期货因子择时上的应用
Guo Tai Jun An Qi Huo· 2025-08-28 08:42
1. Report Industry Investment Rating No relevant content provided. 2. Core Viewpoints of the Report - The report innovatively proposes a dual - channel deep - learning model (LSTM and GRU) that integrates daily - frequency and minute - frequency data, which can effectively capture market information on different time scales, significantly improve the prediction accuracy and stability of the strategy outside the sample (especially during market downturns), and provide a new idea with strong generalization ability for reconstructing the quantitative timing system of the bond market [2]. - The dual - channel model shows excellent generalization ability and robustness in out - of - sample tests, and can maintain a high winning rate in bear markets, effectively making up for the shortcoming of traditional factors failing in market downturns [3]. - In the multi - factor timing framework, the weight of deep - learning factors should be controlled at a relatively low proportion, and machine - learning factors should play a supplementary role to achieve the unity of interpretability and performance improvement [43][44]. 3. Summary by Relevant Catalogs 3.1 Deep - Learning Model Introduction - Traditional quantitative factors in the bond market have declined in performance in recent years, and there is a need to reconstruct and re - mine bond - market quantitative factors. Deep - learning methods can be used to find complex relationships in data, and RNN, LSTM, and GRU are considered suitable for the timing task of Treasury bond futures [7][8]. - RNN can process time - series data but has the problem of gradient disappearance when dealing with long time - series [9]. - LSTM solves the gradient - disappearance problem through a cell state and three gating units, enabling it to learn long - range dependencies in sequences [15]. - GRU simplifies the structure of LSTM, reduces the number of learnable parameters, and has high parameter efficiency and fast training speed [19]. - A dual - channel model is designed to process daily - frequency and minute - frequency data simultaneously to extract features on different time scales and predict the daily - frequency returns of Treasury bond futures, which can reduce the over - fitting risk [22]. 3.2 Treasury Bond Futures Timing Test 3.2.1 Back - testing Settings - The target variable is the open - to - open return of 10 - year Treasury bond futures, and the back - testing time interval is from January 2016 to August 2025, with daily rebalancing, 100% margin, 1 - time leverage, and a bilateral handling fee of 0.01% [25][26][27]. 3.2.2 Daily - frequency Channel Model - The single - daily - frequency channel model based on daily - frequency features performs well within the sample but poorly outside the sample, with obvious over - fitting [33]. 3.2.3 Dual - channel Model - The dual - channel model fuses multi - frequency time - series information. The addition of minute - frequency information significantly improves the prediction effect of the model outside the sample, enhances the generalization ability and stability, and maintains a relatively high winning rate in both long and short positions [40][41][42]. 3.3 Deep - Learning Allocation in the Multi - factor Framework - Deep - learning factors in the multi - factor timing framework have high performance but also have over - fitting risks and lack of interpretability. The weight of deep - learning factors should be controlled at a relatively low proportion, and machine - learning factors should play a supplementary role [43][44]. 3.4 Conclusion - The report explores the application of deep - learning models in Treasury bond futures quantitative timing and proposes a dual - channel deep - learning framework based on multi - frequency data fusion, which can effectively improve the performance of multi - factor strategies [45].
Cell子刊:舒妮/黄伟杰团队综述AI赋能多模态成像,用于神经精神疾病精准医疗
生物世界· 2025-05-26 23:57
Core Viewpoint - The integration of multimodal neuroimaging and artificial intelligence (AI) is revolutionizing the early diagnosis and personalized treatment of neuropsychiatric disorders, addressing the challenges posed by their complex pathology and clinical heterogeneity [2][6]. Multimodal Neuroimaging: A Comprehensive Brain Examination - Traditional single-modality brain examinations are limited, while multimodal imaging can decode the brain from structural, functional, and molecular dimensions, enabling early intervention [7][8]. - Structural imaging (e.g., MRI) reveals brain tissue volume and cortical thickness, functional imaging (e.g., fMRI, EEG) captures neuronal activity, and molecular imaging (e.g., PET) tracks pathological markers like amyloid proteins, providing early warnings for conditions like Alzheimer's disease [9]. AI as a Puzzle Solver - AI demonstrates three key capabilities in handling vast heterogeneous data: feature fusion (early, mid, and late fusion), deep learning models, and clinical prediction tools, significantly enhancing diagnostic accuracy [12][13]. - For instance, multimodal AI models have improved early Alzheimer's diagnosis accuracy to 92.7%, surpassing single-modality methods by over 15% [13]. Practical Achievements: AI's Impact - AI has achieved high diagnostic accuracy, distinguishing Alzheimer's from Lewy body dementia at 87% and predicting epilepsy seizures with over 98% accuracy [14]. - It can predict the efficacy of depression medications with 89% accuracy and assess cognitive decline rates [15]. - AI identified three subtypes in over 2,000 bipolar disorder patients, guiding personalized treatment approaches [16]. Challenges and Breakthroughs: Path to Clinical Application - The integration of multimodal neuroimaging data faces challenges such as data availability, heterogeneity, and AI model interpretability, compounded by issues like class imbalance, algorithm bias, and data privacy [20]. - Addressing these challenges is crucial for developing robust AI models based on multimodal neuroimaging [20]. Future Research Directions - The future of AI in neuropsychiatric disorders includes the development of transformer models for cross-modal data processing, dynamic monitoring of brain network changes, and creating lightweight models for clinical use [23][24]. - Despite significant advancements, further exploration of clinical effectiveness and usability is needed to transition from research to practical applications [24].