Git Product home page Git Product logo

Comments (3)

wjmaddox avatar wjmaddox commented on July 21, 2024

Sure, first, you'll have to train models as described in the README here. Then, using a saved checkpoint, you can compute the uncertainties overall in this README. Running those scripts will save .npz files of all predictions.

For all of those models, we trained using the first half of the CIFAR classes, which is the --split_classes=0 flag throughout; that is, you'll have to add that flag into whatever training and evaluation scripts you use.

Hope this helps,

from swa_gaussian.

amitchandak avatar amitchandak commented on July 21, 2024

Hi,
Thanks for your reply and apologies for very late reply.
I tried to replicate transfer learning experiment. The models were trained as you mentioned here and then predictions were done on STL-10 dataset.

Example Command:
python3.6 ./experiments/uncertainty/uncertainty.py --data_path=<stl_dataPath> --dataset=STL10 --model=VGG16 --use_test --cov_mat --method=SWAG --scale=0.5 --file=<vgg16_swagModelPath> --save_path=<save_path>

Results: (Averaged over three runs with different random seed)

NLL Results:

Model VGG-16 PreResNet-164 WideResNet28x10
SWAG(Reported) 1.1402 ± 0.0342 0.9706 ± 0.00 0.8710 ± 0.00
SWAG(Reproduced) 0.8565 ± 0.0007 0.7783±0.0141 0.66± 0.0013

Accuracy Results

Model VGG-16 PreResNet-164 WideResNet28x10
SWAG(Reported) 72.19 ± 0.06 75.88 ± 0.00 77.09 ± 0.00
SWAG(Reproduced) 72.30 ±0.12 75.836±0.16 76.86± 0.07

I have used exact parameter values to train the models as mentioned so not sure why NLL numbers are different than once mentioned in paper. Can you please give some feedback on this ?

Thanks,
Amit

from swa_gaussian.

wjmaddox avatar wjmaddox commented on July 21, 2024

Hmm... so those are all numbers that are pretty similar to what we are reporting and in all but accuracy on WRN, slightly better than the numbers we reported (lower NLL is better, higher accuracy is better). I'm pretty confident that this is just random seeds (don't know what @izmailovpavel used 18 mos ago) and/or slight changes in pytorch since we last ran the experiments.

from swa_gaussian.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.