终身行人重识别
Search documents
AAAI 2026|新突破:北大彭宇新团队提出可见光-红外终身行人重识别方法CKDA
机器之心· 2025-12-06 04:08
Core Insights - The article discusses the development of a novel method for lifelong pedestrian re-identification called CKDA, which aims to continuously learn new discriminative information from incoming data while retaining the ability to recognize known data across different modalities, specifically visible light and infrared images [2][6]. Group 1: Background and Motivation - Lifelong pedestrian re-identification focuses on recognizing the same individual across different scenarios by continuously learning from pedestrian data collected in various environments [6]. - Existing methods struggle to balance the acquisition of modality-specific knowledge and the retention of cross-modal common knowledge, leading to conflicts that hinder effective learning [9][11]. Group 2: Technical Solution - The CKDA method introduces a framework that includes three main modules: 1. Cross-modal common prompts, which extract shared discriminative knowledge by removing style information unique to each modality [12]. 2. Single-modal specific prompts, which enhance the retention of knowledge specific to each modality while avoiding interference between modalities [20]. 3. Cross-modal knowledge alignment, which aligns new and old knowledge in independent feature spaces to improve the model's ability to balance discriminative knowledge across modalities [24][25]. Group 3: Experimental Results - The CKDA method achieved the best performance on four commonly used visible-infrared pedestrian re-identification datasets, with an average mAP of 36.3% and R1 accuracy of 39.4% [28]. - The results indicate that CKDA effectively enhances the model's perception and retention capabilities for both visible and infrared modality information through complementary prompts [30][29].
IEEE TPAMI 2025 | 北京大学提出分布驱动的终身学习范式,用结构建模解决灾难性遗忘
机器之心· 2025-09-26 10:35
Core Viewpoint - The article discusses a recent research achievement in the field of artificial intelligence, specifically focusing on a new framework called DKP++ for lifelong person re-identification (LReID), which addresses the catastrophic forgetting problem in lifelong learning by enhancing memory retention of historical knowledge and improving cross-domain learning capabilities [2][3]. Research Background - Person re-identification (ReID) aims to match and associate images of the same individual across different camera views, locations, and times, with applications in surveillance, intelligent transportation, and urban safety management [3]. - The traditional ReID paradigm struggles with domain shift issues due to variations in data collection conditions, leading to inadequate adaptability in long-term dynamic environments [3]. Research Challenges - The core challenge in lifelong person re-identification is the catastrophic forgetting problem, where the model's retrieval performance for old domain data significantly decreases after learning new domain knowledge [5]. - Existing methods to mitigate forgetting, such as retaining historical samples or using knowledge distillation, face limitations related to data privacy risks, storage overhead, and model flexibility [5]. Research Motivation - The motivation behind DKP++ includes distribution-aware prototype learning to effectively retain historical knowledge without storing historical samples, and cross-domain distribution alignment to enhance the model's ability to learn new knowledge while utilizing historical information [8][10]. Method Design - DKP++ employs a distribution-aware knowledge aligning and prototyping framework, which includes: 1. Instance-level fine-grained modeling to capture local details of person instances [14]. 2. Distribution-aware prototype generation to create robust category-level prototypes that retain intra-class variation knowledge [14]. 3. Distribution alignment to bridge the feature distribution gap between new and old data [14]. 4. Prototype-based knowledge transfer to guide model learning using generated prototypes and labeled new data [14]. Experimental Analysis - The experiments utilized two typical training domain sequences and five widely used ReID datasets, evaluating the model's knowledge retention and generalization capabilities [16]. - DKP++ demonstrated an improvement of 5.2%-7% in average performance on known domains and 4.5%-7.7% in overall generalization performance on unseen domains compared to existing methods [17]. - The model showed higher historical knowledge retention and faster performance growth in unseen domains as the number of learned domains increased [20]. Technical Innovations - DKP++ introduces innovative designs focusing on distribution prototype modeling and representation, as well as sample alignment-guided prototype knowledge transfer to overcome distribution gaps between new and old domain data [23]. Future Outlook - The research suggests potential improvements in areas such as distribution alignment using larger models, active forgetting mechanisms to eliminate redundant knowledge, and multi-modal lifelong learning capabilities to enhance perception in complex environments [23].
IEEE TPAMI 2025 | 北京大学提出LSTKC++,长短期知识解耦与巩固驱动的终身行人重识别
机器之心· 2025-07-03 00:22
Core Viewpoint - The article discusses the introduction of LSTKC++, a framework developed by a research team at Peking University, aimed at addressing the challenges of lifelong person re-identification (LReID) by effectively balancing the learning of new knowledge and the retention of historical knowledge [1][2][9]. Summary by Sections 1. Introduction to LSTKC++ - LSTKC++ incorporates long-short term knowledge decoupling and dynamic correction mechanisms to enhance the model's ability to learn new knowledge while retaining historical knowledge during lifelong learning [2][9]. 2. Challenges in Person Re-Identification - The traditional ReID paradigm struggles with domain shifts due to variations in location, equipment, and time, leading to difficulties in adapting to long-term dynamic changes in test data [5][6]. - The core challenge of LReID is the catastrophic forgetting problem, where the model's performance on old domain data degrades after learning new domain knowledge [9][12]. 3. Framework Overview - The LSTKC framework, introduced in AAAI 2024, divides the lifelong learning model into short-term and long-term models, allowing for a collaborative fusion of knowledge [11]. - LSTKC++ improves upon LSTKC by decoupling the long-term model into two parts: one for earlier historical knowledge and another for more recent historical knowledge [19]. 4. Methodology Enhancements - LSTKC++ features complementary knowledge transfer between long-term and short-term models, correcting knowledge based on sample affinity matrices and optimizing knowledge weighting parameters using new domain training data [19][20]. - The framework introduces a sample relationship-guided long-term knowledge consolidation mechanism to facilitate reasoning with both long-term and short-term knowledge after learning new domains [20]. 5. Experimental Analysis - LSTKC++ shows an improvement in known domain average performance (Seen-Avg mAP and Seen-Avg R@1) by 1.5%-3.4% compared to existing methods, and an increase in overall generalization performance (Unseen-Avg mAP and Unseen-Avg R@1) by 1.3%-4% [25]. - The framework demonstrates significant performance advantages in intermediate domains while maintaining a balance between knowledge retention and learning new information [25]. 6. Technical Innovations - The work focuses on the challenges of new knowledge learning and historical knowledge forgetting, proposing innovative designs such as a decoupled knowledge memory system and a semantic-level knowledge correction mechanism [26]. 7. Practical Applications - LSTKC++ is suitable for dynamic environments where camera deployment conditions frequently change, making it applicable in smart cities, edge computing, and security scenarios [27]. - The framework is designed to meet privacy protection needs by not requiring access to historical samples during the continuous learning process [27]. 8. Future Directions - Future research may explore extending LSTKC++ to pre-trained visual large models and integrating multi-modal perception for enhanced stability and robustness in continuous learning [28].