TY - GEN
T1 - Plastic and Stable Gated Classifiers for Continual Learning
AU - Kuo, Nicholas I.Hsien
AU - Harandi, Mehrtash
AU - Fourrier, Nicolas
AU - Walder, Christian
AU - Ferraro, Gabriela
AU - Suominen, Hanna
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021/9/1
Y1 - 2021/9/1
N2 - Conventional neural networks are mostly high in plasticity but low in stability. Hence, catastrophic forgetting tends to occur over the sequential training of multiple tasks and a backbone learner loses its ability in solving a previously learnt task. Several studies have shown that catastrophic forgetting can be partially mitigated through freezing the feature extractor weights while only sequentially training the classifier network. Though these are effective methods in retaining knowledge, forgetting could still become severe if the classifier network is over-parameterised over many tasks. As a remedy, this paper presents a novel classifier design with high stability. Highway-Connection Classifier Networks (HCNs) leverage gated units to alleviate forgetting. When employed alone, they exhibit strong robustness against forgetting. In addition, they synergise well with many existing and popular continual learning archetypes. We release our codes at https://github.com/Nic5472K/CLVISION2021_CVPR_HCN
AB - Conventional neural networks are mostly high in plasticity but low in stability. Hence, catastrophic forgetting tends to occur over the sequential training of multiple tasks and a backbone learner loses its ability in solving a previously learnt task. Several studies have shown that catastrophic forgetting can be partially mitigated through freezing the feature extractor weights while only sequentially training the classifier network. Though these are effective methods in retaining knowledge, forgetting could still become severe if the classifier network is over-parameterised over many tasks. As a remedy, this paper presents a novel classifier design with high stability. Highway-Connection Classifier Networks (HCNs) leverage gated units to alleviate forgetting. When employed alone, they exhibit strong robustness against forgetting. In addition, they synergise well with many existing and popular continual learning archetypes. We release our codes at https://github.com/Nic5472K/CLVISION2021_CVPR_HCN
UR - http://www.scopus.com/inward/record.url?scp=85116042388&partnerID=8YFLogxK
U2 - 10.1109/CVPRW53098.2021.00394
DO - 10.1109/CVPRW53098.2021.00394
M3 - Conference contribution
T3 - IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops
SP - 3548
EP - 3553
BT - Proceedings - 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2021
PB - IEEE Computer Society
T2 - 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2021
Y2 - 19 June 2021 through 25 June 2021
ER -