-
Notifications
You must be signed in to change notification settings - Fork 7.2k
Add pretrained Wide ResNet #912
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Codecov Report
@@ Coverage Diff @@
## master #912 +/- ##
==========================================
+ Coverage 63.87% 63.97% +0.09%
==========================================
Files 66 66
Lines 5273 5279 +6
Branches 793 793
==========================================
+ Hits 3368 3377 +9
+ Misses 1673 1671 -2
+ Partials 232 231 -1
Continue to review full report at Codecov.
|
fmassa
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks a lot for the PR Sergey!
This is almost good to merge.
Can you add an entry for it in https://github.com/pytorch/vision/blob/master/docs/source/models.rst, with the accuracies as well in the table?
Then I'll copy the model weights to the pytorch website and let you know
|
Thanks for looking at it @fmassa , I added:
|
|
Sorry for the delay in replying. I've uploaded the pre-trained weights to Can you update the URLs and fix the conflicts? Then it's good to go! |
|
@fmassa done, updated links and rebased on master |
fmassa
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM, thanks!
Just waiting for CI to finish before merging the PR
|
Thanks Sergey! |
I trained WRN-50-2 and WRN-101-2 with master torchvision, which now allows making WRN models with simple width_per_group argument. I did not use the standard training procedure for ResNet though, here are the differences:
so the checkpoints are in torch.float16 to save space.
idk do we want these in torchvision? I could put them in wide-residual-networks instead.