MagMax: Leveraging Model Merging for Seamless Continual Learning - Publication - Bridge of Knowledge

Search

MagMax: Leveraging Model Merging for Seamless Continual Learning

Abstract

This paper introduces a continual learning approach named MagMax, which utilizes model merging to enable large pre-trained models to continuously learn from new data without forgetting previously acquired knowledge. Distinct from traditional continual learning methods that aim to reduce forgetting during task training, MagMax combines sequential fine-tuning with a maximum magnitude weight selection for effective knowledge integration across tasks. Our initial contribution is an extensive examination of model merging techniques, revealing that simple approaches like weight averaging and random weight selection surprisingly hold up well in various continual learning contexts. More importantly, we present MagMax, a novel model-merging strategy that enables continual learning of large pre-trained models for successive tasks. Our thorough evaluation demonstrates the superiority of MagMax in various scenarios, including class- and domain-incremental learning settings. The code is available on github.

Citations

  • 0

    CrossRef

  • 0

    Web of Science

  • 0

    Scopus

Authors (4)

Cite as

Full text

full text is not available in portal

Keywords

Details

Category:
Conference activity
Type:
publikacja w wydawnictwie zbiorowym recenzowanym (także w materiałach konferencyjnych)
Language:
English
Publication year:
2024
Bibliographic description:
Marczak D., Twardowski B., Trzciński T., Cygert S.: MagMax: Leveraging Model Merging for Seamless Continual Learning// / : , 2024,
DOI:
Digital Object Identifier (open in new tab) 10.1007/978-3-031-73013-9_22
Sources of funding:
  • Poza PG
Verified by:
Gdańsk University of Technology

seen 28 times

Recommended for you

Meta Tags