GithubHelp home page GithubHelp logo

l2t-ww's Introduction

Learning What and Where to Transfer (ICML 2019)

Learning What and Where to Transfer (ICML 2019) https://arxiv.org/abs/1905.05901

Requirements

  • python>=3.6
  • pytorch>=1.0
  • torchvision
  • cuda>=9.0

Note. The reported results in our paper were obtained in the old-version pytorch (pytorch=1.0, cuda=9.0). We recently executed again the experiment commands as described below using the recent version (pytorch=1.6.0, torchvision=0.7.0, cuda=10.1), and obtained similar results as reported in the paper.

Prepare Datasets

You can download CUB-200 and Stanford Dogs datasets

You need to run the below pre-processing script for DataLoader.

python cub200.py /data/CUB_200_2011
python dog.py /data/dog

Train L2T-ww

You can train L2T-ww models with the same settings in our paper.

python train_l2t_ww.py --dataset cub200 --datasplit cub200 --dataroot /data/CUB_200_2011
python train_l2t_ww.py --dataset dog --datasplit dog --dataroot /data/dog
python train_l2t_ww.py --dataset cifar100 --datasplit cifar100 --dataroot /data/ --experiment logs/cifar100_0/ --source-path logs --source-model resnet32 --source-domain tinyimagenet-200 --target-model vgg9_bn --pairs 4-0,4-1,4-2,4-3,4-4,9-0,9-1,9-2,9-3,9-4,14-0,14-1,14-2,14-3,14-4 --batchSize 128
python train_l2t_ww.py --dataset stl10 --datasplit stl10 --dataroot /data/ --experiment logs/stl10_0/ --source-path logs --source-model resnet32 --source-domain tinyimagenet-200 --target-model vgg9_bn --pairs 4-0,4-1,4-2,4-3,4-4,9-0,9-1,9-2,9-3,9-4,14-0,14-1,14-2,14-3,14-4 --batchSize 128

l2t-ww's People

Contributors

hankook avatar yunhunjang avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar

l2t-ww's Issues

Reproducibility issue on Stanford40-Actions

I am sincerely thankful for this shared code. I've really enjoyed the work.

While I try to reproduce your reported table, I faced an issue with the reproducibility for Stanford40-Actions (though I've confirmed CUB200, Dog and Indoor). It gives me around 53% compared to the reported score of 63.08%. Let me check with few details:

  1. how do you split the training data with what percent (I tried 20% and 25%)?
  2. Is there any change for hyper-parameters? (Adam optimizer for meta-networks has 1e-3 or 1e-4 learning rate? Or, the weight decay is 0 or 1e-4? (Appendix B)
  3. Any other comment?

Thanks in advance.

Jin-Hwa

about folder */split/

How did you get the file like dog-test,dota-train and dot-val?
What do they represent for?
How to get it or are there any files related to other data sets?
thanks a lot!

Questions about Bilevel optimization

Hi, I really enjoyed reading your paper, and I learned a lot!

I have a simple question. In your algorithm, when constructing the bilevel optimization framework, it seems that L_org does not distinguish between train and val. It means that the inner L_org and the outer L_org comes from the same (training) dataset. Can I learn the intuition of this setting (because most of bilevel optimization assumes exclusive train / val set for inner / outer)?

Thank you!

Could you specify details of TinyImageNet-ResNet32 to CIFAR100-VGG9

Hi, I would like to make sure some implemental details on TinyImageNet-ResNet32 to CIFAR100-VGG9 section including:

  1. The architecture of ResNet-32 (how to modify stem conv; others could be found in Figure 3)?
  2. The architecture of VGG-9 (how to modify the classifier, use BN)?
  3. Any change of the other hyperparameters, the learning rate is just fixed to 0.1 even though the batch size is 2x increased, an input image is smaller, and the capacity of networks is decreased?
  4. How to set aside validation splits for CIFAR-100 and STL-10?
  5. Please specify the augmentation strategy for 32x32 regimes.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.