Title
题目
CCSI: Continual Class-Specific Impression for data-free class incremental learning
CCSI: 用于无数据类别增量学习的持续类别特定印象
01
文献速递介绍
当前用于医学图像分类任务的深度学习模型表现出了令人鼓舞的性能。然而,这些模型大多要求在训练之前收集所有的训练数据并指定所有的类别。它们在部署时对深度学习模型进行一次性训练,并期望其能够在随后的所有数据上执行。然而,这种要求在实际的临床环境中具有局限性,因为医学图像数据是不断收集且随时间变化的,*例如*,当出现新疾病类型时。
一种有前途的解决这一机器学习挑战的方法是使系统具备持续或终身学习的能力,即部署的模型能够适应新数据,同时保持从先前数据中获得的信息。将这些学习技术整合到模型中,将使深度学习模型更能适应医疗数据集的不断扩展。在各种增量场景中,医学持续学习已经被广泛应用(van de Ven等,2022),这些场景考虑到了新数据的非平稳性。这些场景包括任务增量学习(González等,2023;Liao等,2022;Xu等,2022;Kaustaban等,2022;Chakraborti等,2021),即引入新的医学任务,*例如*,将分割网络扩展到另一身体区域;类别增量学习(Chee等,2023;Yang等,2021a;Li等,2020b),即在模型中添加新类别,*例如*,在分类任务中引入一种新疾病类型;以及领域增量学习(Yang等,2023;Srivastava等,2021;Bayasi等,2021),即模型面临其未经过训练的新医学领域。这些增量学习场景的大多数假设是能够访问先前模型的所有训练数据或其中的一部分,这些数据被存储在内存系统中。然后,重新训练过程在保存的数据和新数据上进行。
Abatract
摘要
In real-world clinical settings, traditional deep learning-based classification methods struggle with diagnosingnewly introduced disease types because they require samples from all disease classes for offline training. Classincremental learning offers a promising solution by adapting a deep network trained on specific disease classesto handle new diseases. However, catastrophic forgetting occurs, decreasing the performance of earlier classeswhen adapting the model to new data. Prior proposed methodologies to overcome this require perpetualstorage of previous samples, posing potential practical concerns regarding privacy and storage regulationsin healthcare. To this end, we propose a novel data-free class incremental learning framework that utilizesdata synthesis on learned classes instead of data storage from previous classes. Our key contributions includeacquiring synthetic data known as Continual Class-Specific Impression (CCSI) for previously inaccessible trainedclasses and presenting a methodology to effectively utilize this data for updating networks when introducingnew classes. We obtain CCSI by employing data inversion over gradients of the trained classification modelon previous classes starting from the mean image of each class inspired by common landmarks sharedamong medical images and utilizing continual normalization layers statistics as a regularizer in this pixelwise optimization process. Subsequently, we update the network by combining the synthesized data with newclass data and incorporate several losses, including an intra-domain contrastive loss to generalize the deepnetwork trained on the synthesized data to real data, a margin loss to increase separation among previousclasses and new ones, and a cosine-normalized cross-e