JOURNAL ARTICLE

Enhancing Model Agnostic Meta-Learning via Gradient Similarity Loss

Jae-Ho TakByung‐Woo Hong

Year: 2024 Journal:   Electronics Vol: 13 (3)Pages: 535-535   Publisher: Multidisciplinary Digital Publishing Institute

Abstract

Artificial intelligence (AI) technology has advanced significantly, now capable of performing tasks previously believed to be exclusive to skilled humans. However, AI models, in contrast to humans who can develop skills with relatively less data, often require substantial amounts of data to emulate human cognitive abilities in specific areas. In situations where adequate pre-training data is not available, meta-learning becomes a crucial method for enhancing generalization. The Model Agnostic Meta-Learning (MAML) algorithm, which employs second-order derivative calculations to fine-tune initial parameters for better starting points, plays a pivotal role in this area. However, the computational demand of this method can be challenging for modern models with a large number of parameters. The concept of the Approximate Hessian Effect is introduced in this context, examining the effectiveness of second-order derivatives in identifying initial parameters conducive to high generalization performance. The study suggests the use of cosine similarity and squared error (L2 loss) as a loss function within the Approximate Hessian Effect framework to modify gradient weights, aiming for more generalizable model parameters. Additionally, an algorithm that relies on first-order calculations is presented, designed to achieve performance levels comparable to MAML. This approach was tested and compared with traditional MAML methods using both the MiniImagenet dataset and a modified MNIST dataset. The results were analyzed to evaluate its efficiency. Compared to previous studies that achieved good performance using only the first derivative, this approach is more efficient because it does not require iterative loops to converge on additional loss functions. Additionally, there is potential for further performance enhancement through hyperparameter tuning.

Keywords:
Hessian matrix MNIST database Generalization Computer science Similarity (geometry) Artificial intelligence Context (archaeology) Machine learning Function (biology) Algorithm Artificial neural network Mathematics

Metrics

3
Cited By
1.92
FWCI (Field Weighted Citation Impact)
31
Refs
0.81
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Domain Adaptation and Few-Shot Learning
Physical Sciences →  Computer Science →  Artificial Intelligence
Machine Learning and Data Classification
Physical Sciences →  Computer Science →  Artificial Intelligence
Hydrological Forecasting Using AI
Physical Sciences →  Environmental Science →  Environmental Engineering

Related Documents

JOURNAL ARTICLE

Learning Symbolic Model-Agnostic Loss Functions via Meta-Learning

Christian RaymondQi ChenBing XueMengjie Zhang

Journal:   IEEE Transactions on Pattern Analysis and Machine Intelligence Year: 2023 Vol: 45 (11)Pages: 1-15
JOURNAL ARTICLE

Model-agnostic multi-stage loss optimization meta learning

Xiao YaoJianlong ZhuGuanying HuoNing XuXiaofeng LiuCe Zhang

Journal:   International Journal of Machine Learning and Cybernetics Year: 2021 Vol: 12 (8)Pages: 2349-2363
JOURNAL ARTICLE

Modified Model-Agnostic Meta-Learning

Aashay Pawar

Year: 2020 Vol: 197 Pages: 1-4
JOURNAL ARTICLE

Meta weight learning via model-agnostic meta-learning

Zhixiong XuXiliang ChenWei TangJun LaiLei Cao

Journal:   Neurocomputing Year: 2020 Vol: 432 Pages: 124-132
© 2026 ScienceGate Book Chapters — All rights reserved.