Skip to content
/ ALFA Public

Source code for NeurIPS 2020 paper "Meta-Learning with Adaptive Hyperparameters"

Notifications You must be signed in to change notification settings

baiksung/ALFA

Repository files navigation

ALFA - Meta-Learning with Adaptive Hyperparameters

Sungyong Baik, Myungsub Choi, Janghoon Choi, Heewon Kim, Kyoung Mu Lee

Source code for NeurIPS 2020 paper "Meta-Learning with Adaptive Hyperparameters" (previously titled "Adaptive Learning for Fast Adaptation")

This repository is the implementation of ALFA. The code is based off the public code of MAML++, where their reimplementation of MAML is used as the baseline.

Paper-arXiv

Requirements

  • Ubuntu 18.04
  • Anaconda3
  • Python==3.6
  • PyTorch==1.5
  • numpy==1.19.1

To install requirements, first download Anaconda3 and then run the following:

bash install.sh

Hardware Requirements

  • GPU with memory more than 27GB for a single-GPU ResNet12 backbone second-order training.
  • The current version does not support a multi-GPU setting. While running on a multi-GPU will not give errors, it will give incorrect results (due to uneven distribution of labels across GPUs).

Datasets

For miniIamgenet, the dataset can be downloaded from the link provided from MAML++ public code. make a directory named 'datasets' and place the downloaded miniImagnet under the 'datasets' directory.

Training

To train the model(s) in the paper, run this command in experiment_scripts folder:

For single GPU

bash alfa+maml.sh 0

where 0 represent GPU_ID.

Evaluation

After training is finished, the same command is run to evaluate:

For single GPU:

bash alfa+maml.sh 0

Releases

No releases published

Packages

No packages published