-
Notifications
You must be signed in to change notification settings - Fork 7.2k
Description
🐛 Bug
WideResNets as described in the original paper ("Wide Residual Networks" https://arxiv.org/pdf/1605.07146.pdf) use Dropout. The Torchvision implementation does not.
If this for performance reasons, or similar, please add a comment. It would be very helpful to point out that difference clearly. When it comes to reproducing baselines, confusion about what is a real WideResNet vs not could complicate research efforts.
Thanks for your consideration!
To Reproduce
Steps to reproduce the behavior:
Examine model definition in https://github.com/pytorch/vision/blob/master/torchvision/models/resnet.py and compare to paper, or other implementations eg https://github.com/meliketoy/wide-resnet.pytorch/blob/master/networks/wide_resnet.py.
Expected behavior
Dropout layers to be added in the residual blocks.
Environment
torchvision 0.6/head