Closed
Description
Looks like test_quantized_classification_model[resnet50]
is failing in some PRs like https://app.circleci.com/pipelines/github/pytorch/vision/11583/workflows/84517aa3-fa6b-4527-8ee6-8a09ae76199f/jobs/900594
and it's also failing internally: https://www.internalfb.com/intern/tests/search?search_id=757311188293548
It looks like the failure is related to the new expected values checks introduced in #4597
I think an easy fix is just to add resnet50 to the quantized_flaky_models
list, perhaps there's a better solution. It seems that the errors are consistently the same across executions, with the same atol
and rtol
differences:
Mismatched elements: 1 / 5 (20.0%)
Greatest absolute difference: 0.2876443862915039 at index (0, 1) (up to 0.1 allowed)
Greatest relative difference: 0.16666666666666666 at index (0, 1) (up to 0.1 allowed)
So there might be a source of variability that we're not controlling?