double PT: Enhancing Meta-Learning Performance with Pre-Train and Meta-Pre-Train Processes

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

With meta-learning, models are trained on multiple tasks, and resulting trained models are expected to be capable of ' learning' new tasks effectively. MAML (Model Agnostic Meta-Learning) was one such early design which allowed models to reuse learned features, but with limited performance. Pre-training is another known method to improve the performance of a final trained model. Pre-train starts by assisting models to seek better initialization points, thus offering better feature representations. In this paper, we propose doublePT (double-pretrain) which is a two-stage pre-training method with goals to 1) reduce the number of parameters, 2) acquire better feature representations, and 3) achieve competitive overall performances in different benchmark measurements. For the first stage operation, we use a universal pre-training model to capture general features from a large dataset. In the second stage, we propose to use MAML to fine-tune the pre-training model to enhance the feature adaptability. Since the first-stage pre-training model has already learned general feature representations, it reduces the training activities for the second-stage fine-tuning operations, and enables better feature extractions in new tasks. Validated through our experiments, we find that our proposed doublePT approach fine-tunes across different tasks, and performs better than that of one-stage pre-training approach. Upon combining doublePT and DINOv2, and comparing to the latest PMF meta-learning design, the number of parameters required by the PMF pipelining model needs 304.8% more parameters than in our proposed DINOv2+doublePT model design. Performance-wise, the DINOv2+doublePT also has the best accuracies across different benchmark measurements.

Original languageEnglish
Title of host publicationProceedings - 2023 IEEE 35th International Conference on Tools with Artificial Intelligence, ICTAI 2023
PublisherIEEE Computer Society
Pages688-692
Number of pages5
ISBN (Electronic)9798350342734
DOIs
Publication statusPublished - 2023
Event35th IEEE International Conference on Tools with Artificial Intelligence, ICTAI 2023 - Atlanta, United States
Duration: 6 Nov 20238 Nov 2023

Publication series

NameProceedings - International Conference on Tools with Artificial Intelligence, ICTAI
ISSN (Print)1082-3409

Conference

Conference35th IEEE International Conference on Tools with Artificial Intelligence, ICTAI 2023
Country/TerritoryUnited States
CityAtlanta
Period6/11/238/11/23

Keywords

  • double-pre-train
  • few-shot learning
  • meta-learning
  • meta-pre-train

Fingerprint

Dive into the research topics of 'double PT: Enhancing Meta-Learning Performance with Pre-Train and Meta-Pre-Train Processes'. Together they form a unique fingerprint.

Cite this