Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fashionMNIST results #3

Open
bkj opened this issue Jul 2, 2018 · 5 comments
Open

fashionMNIST results #3

bkj opened this issue Jul 2, 2018 · 5 comments

Comments

@bkj
Copy link

bkj commented Jul 2, 2018

This code is linked from the fashion-mnist repo, w/ very good results. Do you have a script somewhere I might be able to use to reproduce those numbers?

Thanks
Ben

@ajbrock
Copy link
Owner

ajbrock commented Jul 2, 2018

Just drop a FashionMNIST data-loader in utils.py and replace the mean / average statistics. Run it with the options that turn FreezeOut off, something like:
python train.py --t_0 1.0 --epochs 300 --which_dataset 10

and then whatever args you want for your model of choice. The scripts I used are on another machine I won't have access to for a few months, sorry.

Also, my results are not very good, it's just that no one has ever bothered to properly benchmark that dataset. With a tiny bit of hyperparameter tuning I wouldn't be surprised if you could exceed 98% accuracy, modulo how many mislabeled samples there are in the test set.

@bkj
Copy link
Author

bkj commented Jul 2, 2018

OK thanks. Do you have any pointers to a well-tuned model?

~ Ben

@ajbrock
Copy link
Owner

ajbrock commented Jul 2, 2018

Probably try one of those shake-shake or shake-drop variants, with a well-tuned SGDR cycle and distillation a-la born again neural nets or stochastic weighted averaging, and figure out the right width/depth. Googling any of those terms should point you in the right direction if they're not familiar =)

@bkj
Copy link
Author

bkj commented Jul 2, 2018

Yeah I bet that would do it... but have you ever seen a model that uses all of those techniques? I'm not sure any datasets are properly benchmarked by your definition! :)

Related -- have you ever seen any code implementing born again networks? It's fairly simple, but details can be finnicky.

~ Ben

@ajbrock
Copy link
Owner

ajbrock commented Jul 3, 2018

Hah, I'm not saying that datasets that haven't been hit with the latest-and-greatest aren't properly benchmarked, but no one's beaten a vanilla WRN40-4 with standard data aug. Kagglers would have a field day =p

As to BANN you can try asking Tommaso but i'm not familiar with any implementations.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants