double PT: Enhancing Meta-Learning Performance with Pre-Train and Meta-Pre-Train Processes

研究成果: Conference contribution同行評審

摘要

With meta-learning, models are trained on multiple tasks, and resulting trained models are expected to be capable of ' learning' new tasks effectively. MAML (Model Agnostic Meta-Learning) was one such early design which allowed models to reuse learned features, but with limited performance. Pre-training is another known method to improve the performance of a final trained model. Pre-train starts by assisting models to seek better initialization points, thus offering better feature representations. In this paper, we propose doublePT (double-pretrain) which is a two-stage pre-training method with goals to 1) reduce the number of parameters, 2) acquire better feature representations, and 3) achieve competitive overall performances in different benchmark measurements. For the first stage operation, we use a universal pre-training model to capture general features from a large dataset. In the second stage, we propose to use MAML to fine-tune the pre-training model to enhance the feature adaptability. Since the first-stage pre-training model has already learned general feature representations, it reduces the training activities for the second-stage fine-tuning operations, and enables better feature extractions in new tasks. Validated through our experiments, we find that our proposed doublePT approach fine-tunes across different tasks, and performs better than that of one-stage pre-training approach. Upon combining doublePT and DINOv2, and comparing to the latest PMF meta-learning design, the number of parameters required by the PMF pipelining model needs 304.8% more parameters than in our proposed DINOv2+doublePT model design. Performance-wise, the DINOv2+doublePT also has the best accuracies across different benchmark measurements.

原文English
主出版物標題Proceedings - 2023 IEEE 35th International Conference on Tools with Artificial Intelligence, ICTAI 2023
發行者IEEE Computer Society
頁面688-692
頁數5
ISBN(電子)9798350342734
DOIs
出版狀態Published - 2023
事件35th IEEE International Conference on Tools with Artificial Intelligence, ICTAI 2023 - Atlanta, United States
持續時間: 6 11月 20238 11月 2023

出版系列

名字Proceedings - International Conference on Tools with Artificial Intelligence, ICTAI
ISSN(列印)1082-3409

Conference

Conference35th IEEE International Conference on Tools with Artificial Intelligence, ICTAI 2023
國家/地區United States
城市Atlanta
期間6/11/238/11/23

指紋

深入研究「double PT: Enhancing Meta-Learning Performance with Pre-Train and Meta-Pre-Train Processes」主題。共同形成了獨特的指紋。

引用此