When comparing continual learning methods, the choice of which model layers to train is as important as the method itself—different fine-tuning regimes can completely change which approach performs best.
This paper shows that how you choose which parts of a model to update during continual learning (learning new tasks sequentially) significantly changes which methods work best.