Task-Specific Preconditioner for Cross-Domain Few-Shot Learning

Suhyun Kang, Jungwon Park, Wonseok Lee, Wonjong Rhee

Research output: Contribution to journalConference articlepeer-review

Abstract

Cross-Domain Few-Shot Learning (CDFSL) methods typically parameterize models with task-agnostic and task-specific parameters. To adapt task-specific parameters, recent approaches have utilized fixed optimization strategies, despite their potential sub-optimality across varying domains or target tasks. To address this issue, we propose a novel adaptation mechanism called Task-Specific Preconditioned gradient descent (TSP). Our method first meta-learns Domain-Specific Preconditioners (DSPs) that capture the characteristics of each meta-training domain, which are then linearly combined using task-coefficients to form the Task-Specific Preconditioner. The preconditioner is applied to gradient descent, making the optimization adaptive to the target task. We constrain our preconditioners to be positive definite, guiding the preconditioned gradient toward the direction of steepest descent. Empirical evaluations on the Meta-Dataset show that TSP achieves state-of-the-art performance across diverse experimental scenarios.

Original languageEnglish
Pages (from-to)17760-17769
Number of pages10
JournalProceedings of the AAAI Conference on Artificial Intelligence
Volume39
Issue number17
DOIs
StatePublished - 11 Apr 2025
Event39th Annual AAAI Conference on Artificial Intelligence, AAAI 2025 - Philadelphia, United States
Duration: 25 Feb 20254 Mar 2025

Fingerprint

Dive into the research topics of 'Task-Specific Preconditioner for Cross-Domain Few-Shot Learning'. Together they form a unique fingerprint.

Cite this