Abstract
Catastrophic forgetting is a key challenge for class-incremental learning with deep neural networks, where the performance decreases considerably while dealing with long sequences of new classes. To tackle this issue, in this paper, we propose a new exemplar-supported representation for incremental learning (ESRIL) approach that consists of three components. First, we use memory aware synapses (MAS) pre-trained on the ImageNet to retain the ability of robust representation learning and classification for old classes from the perspective of the model. Second, exemplar-based subspace clustering (ESC) is utilized to construct the exemplar set, which can keep the performance from various views of the data. Third, the nearest class multiple centroids (NCMC) is used as the classifier to save the training cost of the fully connected layer of MAS when the criterion is met. Intensive experiments and analyses are presented to show the influence of various backbone structures and the effectiveness of different components in our model. Experiments on several general-purpose and fine-grained image recognition datasets have fully demonstrated the efficacy of the proposed methodology.
| Original language | English |
|---|---|
| Article number | 9034001 |
| Pages (from-to) | 51276-51284 |
| Number of pages | 9 |
| Journal | IEEE Access |
| Volume | 8 |
| DOIs | |
| Publication status | Published - 12 Mar 2020 |
Funding
This work was supported in part by the Key Research and Development Plan of Shanxi Province under Grant 201703D111023, in part by the Shanxi International Cooperation Project under Grant 201803D421039, in part by the Hundred Talents Program of Shanxi, and in part by the Natural Science Foundation of Shanxi Province under Grant 201801D121144 and Grant 201801D221190.
Keywords
- exemplar-based subspace clustering
- image recognition
- incremental learning
- memory aware synapses