Skip to content

WideResNet implementation doesn't use Dropout  #2196

@BlackHC

Description

@BlackHC

🐛 Bug

WideResNets as described in the original paper ("Wide Residual Networks" https://arxiv.org/pdf/1605.07146.pdf) use Dropout. The Torchvision implementation does not.

If this for performance reasons, or similar, please add a comment. It would be very helpful to point out that difference clearly. When it comes to reproducing baselines, confusion about what is a real WideResNet vs not could complicate research efforts.

Thanks for your consideration!

To Reproduce

Steps to reproduce the behavior:

Examine model definition in https://github.com/pytorch/vision/blob/master/torchvision/models/resnet.py and compare to paper, or other implementations eg https://github.com/meliketoy/wide-resnet.pytorch/blob/master/networks/wide_resnet.py.

Expected behavior

Dropout layers to be added in the residual blocks.

Environment

torchvision 0.6/head

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions