Publications

Detailed Information

SHOT: Suppressing the Hessian along the Optimization Trajectory for Gradient-Based Meta-Learning

Cited 0 time in Web of Science Cited 1 time in Scopus
Authors

Lee, JunHoo; Yoo, Jayeon; Kwak, No Jun

Issue Date
2023-12
Publisher
Neural information processing systems foundation
Citation
Advances in Neural Information Processing Systems, Vol.36
Abstract
In this paper, we hypothesize that gradient-based meta-learning (GBML) implicitly suppresses the Hessian along the optimization trajectory in the inner loop.Based on this hypothesis, we introduce an algorithm called SHOT (Suppressing the Hessian along the Optimization Trajectory) that minimizes the distance between the parameters of the target and reference models to suppress the Hessian in the inner loop.Despite dealing with high-order terms, SHOT does not increase the computational complexity of the baseline model much.It is agnostic to both the algorithm and architecture used in GBML, making it highly versatile and applicable to any GBML baseline.To validate the effectiveness of SHOT, we conduct empirical tests on standard few-shot learning tasks and qualitatively analyze its dynamics.We confirm our hypothesis empirically and demonstrate that SHOT outperforms the corresponding baseline.Code is available at: https://github.com/JunHoo-Lee/SHOT.
ISSN
1049-5258
URI
https://hdl.handle.net/10371/204814
Files in This Item:
There are no files associated with this item.
Appears in Collections:

Related Researcher

  • Graduate School of Convergence Science & Technology
  • Department of Intelligence and Information
Research Area Feature Selection and Extraction, Object Detection, Object Recognition

Altmetrics

Item View & Download Count

  • mendeley

Items in S-Space are protected by copyright, with all rights reserved, unless otherwise indicated.

Share