Official repository of
- Online Class-Incremental Continual Learning with Adversarial Shapley Value (AAAI 2021)
- Supervised Contrastive Replay: Revisiting the Nearest Class Mean Classifier in Online Class-Incremental Continual Learning (CVPR2021 Workshop)
- Online Continual Learning in Image Classification: An Empirical Survey (Neurocomputing), Official version
Create a virtual enviroment
virtualenv online-cl
Activating a virtual environment
source online-cl/bin/activate
Installing packages
pip install -r requirements.txt
- Split CIFAR10
- Split CIFAR100
- CORe50-NC
- Split Mini-ImageNet
- NonStationary-MiniImageNet (Noise, Occlusion, Blur)
- CORe50-NI
- CIFAR10 & CIFAR100 will be downloaded during the first run
- CORE50 download:
source fetch_data_setup.sh
- Mini-ImageNet: Download from https://www.kaggle.com/whitemoon/miniimagenet/download , and place it in datasets/mini_imagenet/
- NonStationary-MiniImageNet will be generated on the fly
- ASER: Adversarial Shapley Value Experience Replay(AAAI, 2021) [Paper]
- EWC++: Efficient and online version of Elastic Weight Consolidation(EWC) (ECCV, 2018) [Paper]
- iCaRL: Incremental Classifier and Representation Learning (CVPR, 2017) [Paper]
- LwF: Learning without forgetting (ECCV, 2016) [Paper]
- AGEM: Averaged Gradient Episodic Memory (ICLR, 2019) [Paper]
- ER: Experience Replay (ICML Workshop, 2019) [Paper]
- MIR: Maximally Interfered Retrieval (NeurIPS, 2019) [Paper]
- GSS: Gradient-Based Sample Selection (NeurIPS, 2019) [Paper]
- GDumb: Greedy Sampler and Dumb Learner (ECCV, 2020) [Paper]
- CN-DPM: Continual Neural Dirichlet Process Mixture (ICLR, 2020) [Paper]
- SCR: Supervised Contrastive Replay (CVPR Workshop, 2021) [Paper]
- Label trick [Paper]
- Cross entropy with knowledge distillation [Paper]
- Multiple iterations [Paper]
- Nearest Class Mean classifier [Paper]
- Separated Softmax [Paper]
- Review Trick [Paper]
Detailed descriptions of options can be found in general_main.py
#ER
python general_main.py --data cifar100 --cl_type nc --agent ER --retrieve random --update random --mem_size 5000
#MIR
python general_main.py --data cifar100 --cl_type nc --agent ER --retrieve MIR --update random --mem_size 5000
#GSS
python general_main.py --data cifar100 --cl_type nc --agent ER --retrieve random --update GSS --eps_mem_batch 10 --gss_mem_strength 20 --mem_size 5000
#LwF
python general_main.py --data cifar100 --cl_type nc --agent LWF
#iCaRL
python general_main.py --data cifar100 --cl_type nc --agent ICARL --retrieve random --update random --mem_size 5000
#EWC++
python general_main.py --data cifar100 --cl_type nc --agent EWC --fisher_update_after 50 --alpha 0.9 --lambda_ 100
#GDumb
python general_main.py --data cifar100 --cl_type nc --agent GDUMB --mem_size 1000 --mem_epoch 30 --minlr 0.0005 --clip 10
#AGEM
python general_main.py --data cifar100 --cl_type nc --agent AGEM --retrieve random --update random --mem_size 5000
#CN-DPM
python general_main.py --data cifar100 --cl_type nc --agent CNDPM --stm_capacity 1000 --classifier_chill 0.01 --log_alpha -300
#ASER
python general_main.py --data cifar100 --cl_type nc --agent ER --update ASER --retrieve ASER --mem_size 5000 --aser_type asvm --n_smp_cls 1.5 --k 3
#SCR
python general_main.py --data cifar100 --cl_type nc --agent SCR --retrieve random --update random --mem_size 5000 --head mlp --temp 0.07 --eps_mem_batch 100
python general_main.py --review_trick True --data cifar100 --cl_type nc --agent ER --retrieve MIR --update random --mem_size 5000
python main_tune.py --general config/general_1.yml --data config/data/cifar100/cifar100_nc.yml --default config/agent/mir/mir_1k.yml --tune config/agent/mir/mir_tune.yml
There are four config files controling the experiment.
- general config controls variables that are not changed during the experiment
- data config controls variables related to the dataset
- default method config controls variables for a specific method that are not changed during the experiment
- method tuning config controls variables that are used for tuning during the experiment
├──agents #Files for different algorithms
├──base.py #Abstract class for algorithms
├──agem.py #File for A-GEM
├──cndpm.py #File for CN-DPM
├──ewc_pp.py #File for EWC++
├──exp_replay.py #File for ER, MIR and GSS
├──gdumb.py #File for GDumb
├──iCaRL.py #File for iCaRL
├──lwf.py #File for LwF
├──scr.py #File for SCR
├──continuum #Files for create the data stream objects
├──dataset_scripts #Files for processing each specific dataset
├──dataset_base.py #Abstract class for dataset
├──cifar10.py #File for CIFAR10
├──cifar100,py #File for CIFAR100
├──core50.py #File for CORe50
├──mini_imagenet.py #File for Mini_ImageNet
├──openloris.py #File for OpenLORIS
├──continuum.py
├──data_utils.py
├──non_stationary.py
├──models #Files for backbone models
├──ndpm #Files for models of CN-DPM
├──...
├──pretrained.py #Files for pre-trained models
├──resnet.py #Files for ResNet
├──utils #Files for utilities
├──buffer #Files related to buffer
├──aser_retrieve.py #File for ASER retrieval
├──aser_update.py #File for ASER update
├──aser_utils.py #File for utilities for ASER
├──buffer.py #Abstract class for buffer
├──buffer_utils.py #General utilities for all the buffer files
├──gss_greedy_update.py #File for GSS update
├──mir_retrieve.py #File for MIR retrieval
├──random_retrieve.py #File for random retrieval
├──reservoir_update.py #File for random update
├──global_vars.py #Global variables for CN-DPM
├──io.py #Code related to load and store csv or yarml
├──kd_manager.py #File for knowledge distillation
├──name_match.py #Match name strings to objects
├──setup_elements.py #Set up and initialize basic elements
├──utils.py #File for general utilities
├──config #Config files for hyper-parameters tuning
├──agent #Config files related to agents
├──data #Config files related to dataset
├──general_*.yml #General yml (fixed variables, not tuned)
├──global.yml #paths to store results
The hyperparameters used in the ASER and SCR papers can be found in the folder config_CVPR
to duplicate the papers' results.
If you use this paper/code in your research, please consider citing us:
Supervised Contrastive Replay: Revisiting the Nearest Class Mean Classifier in Online Class-Incremental Continual Learning
Accepted at CVPR2021 Workshop.
@inproceedings{mai2021supervised,
title={Supervised Contrastive Replay: Revisiting the Nearest Class Mean Classifier in Online Class-Incremental Continual Learning},
author={Mai, Zheda and Li, Ruiwen and Kim, Hyunwoo and Sanner, Scott},
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
pages={3589--3599},
year={2021}
}
Online Continual Learning in Image Classification: An Empirical Survey
Published in Neurocomputing, official version
Preprint on arXiv here.
@article{MAI202228,
title = {Online continual learning in image classification: An empirical survey},
journal = {Neurocomputing},
volume = {469},
pages = {28-51},
year = {2022},
issn = {0925-2312},
doi = {https://doi.org/10.1016/j.neucom.2021.10.021},
url = {https://www.sciencedirect.com/science/article/pii/S0925231221014995},
author = {Zheda Mai and Ruiwen Li and Jihwan Jeong and David Quispe and Hyunwoo Kim and Scott Sanner}
}
Online Class-Incremental Continual Learning with Adversarial Shapley Value
@inproceedings{shim2021online,
title={Online Class-Incremental Continual Learning with Adversarial Shapley Value},
author={Shim, Dongsub and Mai, Zheda and Jeong, Jihwan and Sanner, Scott and Kim, Hyunwoo and Jang, Jongseong},
booktitle={Proceedings of the AAAI Conference on Artificial Intelligence},
volume={35},
number={11},
pages={9630--9638},
year={2021}
}
- Zheda Mai (Corresponding author)
zheda.mai@mail.utoronto.ca - Ruiwen Li
- Dongsub Shim
The PyTorch implementation of ASER in this repository is more efficient than the original TensorFlow implementation and has better performance. The results of the ASER paper can be reproduced in the original TensorFlow implementation repository.