AI Article Synopsis

  • Continual learning (CL) focuses on learning new tasks while retaining knowledge from previous tasks, but traditional methods often struggle due to the lack of available raw data because of copyright and privacy issues.
  • This paper introduces two innovative settings for CL: data-efficient CL (DECL-APIs) and data-free CL (DFCL-APIs), which utilize machine learning APIs to train models with little or no raw data, facing unique challenges like incomplete data and model variability.
  • A new cooperative continual distillation learning framework is proposed, using adversarial training with generators to create synthetic data from APIs, enabling effective knowledge transfer to a CL model while implementing a regularization term to prevent forgetting previous knowledge; results show that this approach can perform compar

Article Abstract

Continual learning (CL) aims to learn new tasks without forgetting previous tasks. However, existing CL methods require a large amount of raw data, which is often unavailable due to copyright considerations and privacy risks. Instead, stakeholders usually release pre-trained machine learning models as a service (MLaaS), which users can access via APIs. This paper considers two practical-yet-novel CL settings: data-efficient CL (DECL-APIs) and data-free CL (DFCL-APIs), which achieve CL from a stream of APIs with partial or no raw data. Performing CL under these two new settings faces several challenges: unavailable full raw data, unknown model parameters, heterogeneous models of arbitrary architecture and scale, and catastrophic forgetting of previous APIs. To overcome these issues, we propose a novel data-free cooperative continual distillation learning framework that distills knowledge from a stream of APIs into a CL model by generating pseudo data, just by querying APIs. Specifically, our framework includes two cooperative generators and one CL model, forming their training as an adversarial game. We first use the CL model and the current API as fixed discriminators to train generators via a derivative-free method. Generators adversarially generate hard and diverse synthetic data to maximize the response gap between the CL model and the API. Next, we train the CL model by minimizing the gap between the responses of the CL model and the black-box API on synthetic data, to transfer the API's knowledge to the CL model. Furthermore, we propose a new regularization term based on network similarity to prevent catastrophic forgetting of previous APIs. Our method performs comparably to classic CL with full raw data on the MNIST and SVHN datasets in the DFCL-APIs setting. In the DECL-APIs setting, our method achieves 0.97×, 0.75× and 0.69× performance of classic CL on the more challenging CIFAR10, CIFAR100, and MiniImageNet, respectively.

Download full-text PDF

Source
http://dx.doi.org/10.1109/TPAMI.2024.3460871DOI Listing

Publication Analysis

Top Keywords

raw data
16
stream apis
12
forgetting previous
12
continual learning
8
full raw
8
model
8
catastrophic forgetting
8
previous apis
8
synthetic data
8
apis
7

Similar Publications

Motivation: Recent advancements in parallel sequencing methods have precipitated a surge in publicly available short-read sequence data. This has encouraged the development of novel computational tools for the de novo assembly of transcriptomes from RNA-seq data. Despite the availability of these tools, performing an end-to-end transcriptome assembly remains a programmatically involved task necessitating familiarity with best practices.

View Article and Find Full Text PDF

Background: The Knight Alzheimer Research Imaging (KARI) dataset, a compilation of data from projects conducted at Washington University in St. Louis, represents a comprehensive effort to advance our understanding of Alzheimer disease (AD) through multimodal data collection. The overarching goal is to characterize normal aging and disease progression to contribute insights into the biological changes preceding AD symptom onset.

View Article and Find Full Text PDF

Background: Neuroinflammation is an integral part of Alzheimer's Disease (AD) pathology, whereby inflammatory processes contribute to the production of amyloid-β, the propagation of tau pathology, and neuronal loss. We recently investigated data-driven methods for determining distinct progression trajectory groups on the ADCOMS scale. This study evaluates whether biomarkers of inflammation in cerebrospinal fluid (CSF) can predict progression rate and membership of those progression rate groups.

View Article and Find Full Text PDF

Background: Structural and functional heterogeneity in the brains of patients with Alzheimer's disease (AD) leads to diagnostic and prognostic uncertainty and confounds clinical treatment planning. Normative modelling, where individual-level deviations in brain measures from a reference sample are computed to infer personalized effects of disease, allows parsing of disease heterogeneity. In this study, GAN based normative modelling technique quantifies individual level neuroanatomical abnormality thereby facilitating measurement of personalized disease related effects in AD patients.

View Article and Find Full Text PDF

Biomarkers.

Alzheimers Dement

December 2024

School of Psychological Sciences and Turner Institute for Brain and Mental Health, Monash University, Monash, VIC, Australia.

Background: Diagnostic and prognostic decisions about Alzheimer's disease (AD) are more accurate when based on large data sets. We developed and validated a machine learning (ML) data harmonization tool for aggregation of prospective data from neuropsychological tests applied to study AD. The online ML-combine application (OML-combine app) allows researchers to utilize the ML-harmonization method for harmonization of their own data with that from other large available data bases (e.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!