Opportunistic dynamic architecture for class-incremental learning

Fahrurrozi Rahman*, Andrea Rosales Sanabria, Juan Ye

*Corresponding author for this work

Research output: Contribution to journalReview articlepeer-review

Abstract

Continual learning has attracted increasing attention over the last few years, as it enables to continually learn new tasks over time, which has significant implication to many real-world applications. A large number of continual learning techniques are proposed and achieve promising performance; however, many of them commit to a fixed, large architecture at the beginning, which can waste the memory space and incur high training cost. To directly tackle this challenge, we propose an Opportunistic Dynamic Architecture,ODA, based on mixture of experts. ODA can automatically grow with more experts for new incoming tasks and opportunistically shrink by merging experts with similar weights. We evaluated ODA on three commonly used datasets: CIFAR-100, CUB, and iNaturalist, and compared against eight existing continual learning techniques. ODA not only outperforms these techniques but does so with a parameter size that is slightly smaller on average, maintaining memory efficiency without compromising accuracy. Furthermore, ODA achieves this with only around 16% of the training time across all datasets when updating for each new task,making it a highly resource-efficient solution for continual learning applications.
Original languageEnglish
Article number10946184
Pages (from-to)1-11
Number of pages11
JournalIEEE Access
Volume13
DOIs
Publication statusPublished - 31 Mar 2025

Keywords

  • Class-incremental learning
  • Continual learning
  • Mixture of expert

Fingerprint

Dive into the research topics of 'Opportunistic dynamic architecture for class-incremental learning'. Together they form a unique fingerprint.

Cite this