Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Reset Dynamo at Setup for all tests #9561

Merged
merged 2 commits into from
Mar 25, 2025
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
3 changes: 3 additions & 0 deletions backends/xnnpack/test/models/deeplab_v3.py
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,9 @@ def forward(self, *args):


class TestDeepLabV3(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

dl3 = DL3Wrapper()
dl3 = dl3.eval()
model_inputs = (torch.randn(1, 3, 224, 224),)
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/models/edsr.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,9 @@


class TestEDSR(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

edsr = edsr_r16f64(2, False).eval() # noqa
model_inputs = (torch.randn(1, 3, 224, 224),)

Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/models/emformer_rnnt.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,9 @@


class TestEmformerModel(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class EmformerRnnt(torch.nn.Module):
def __init__(self):
super().__init__()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/models/inception_v3.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,9 @@


class TestInceptionV3(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

ic3 = models.inception_v3(weights="IMAGENET1K_V1").eval() # noqa
model_inputs = (torch.randn(1, 3, 224, 224),)

Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/models/inception_v4.py
Original file line number Diff line number Diff line change
Expand Up @@ -12,6 +12,9 @@


class TestInceptionV4(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

ic4 = inception_v4(pretrained=False).eval()
model_inputs = (torch.randn(3, 299, 299).unsqueeze(0),)

Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/models/llama2_et_example.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,9 @@


class TestLlama2ETExample(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

def test_f32(self):
self._test()

Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/models/mobilebert.py
Original file line number Diff line number Diff line change
Expand Up @@ -12,6 +12,9 @@


class TestMobilebert(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

# pyre-ignore
mobilebert = MobileBertModel(MobileBertConfig()).eval()
example_inputs = (torch.tensor([[101, 7592, 1010, 2026, 3899, 2003, 10140, 102]]),)
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/models/mobilenet_v2.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,9 @@


class TestMobileNetV2(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

mv2 = models.mobilenetv2.mobilenet_v2(weights=MobileNet_V2_Weights)
mv2 = mv2.eval()
model_inputs = (torch.randn(1, 3, 224, 224),)
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/models/mobilenet_v3.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,9 @@


class TestMobileNetV3(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

mv3 = models.mobilenetv3.mobilenet_v3_small(pretrained=True)
mv3 = mv3.eval()
model_inputs = (torch.randn(1, 3, 224, 224),)
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/models/resnet.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,9 @@


class TestResNet18(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

inputs = (torch.randn(1, 3, 224, 224),)
dynamic_shapes = (
{
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/models/torchvision_vit.py
Original file line number Diff line number Diff line change
Expand Up @@ -12,6 +12,9 @@


class TestViT(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

vit = models.vision_transformer.vit_b_16(weights="IMAGENET1K_V1")
vit = vit.eval()
model_inputs = (torch.randn(1, 3, 224, 224),)
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/models/very_big_model.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestVeryBigModel(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class BigModel(torch.nn.Module):
def __init__(self):
super().__init__()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/models/w2l.py
Original file line number Diff line number Diff line change
Expand Up @@ -12,6 +12,9 @@


class TestW2L(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

batch_size = 10
input_frames = 700
vocab_size = 4096
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_abs.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestAbs(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class Abs(torch.nn.Module):
def __init__(self):
super().__init__()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_add.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestAdd(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class Add(torch.nn.Module):
def __init__(self):
super().__init__()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_avgpool2d.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestAvgPool2d(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class AvgPool2d(torch.nn.Module):
def __init__(
self, count_include_pad=False, ceil_mode=False, divisor_override=None
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_bilinear2d.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,9 @@


class TestUpsampleBilinear2d(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class StaticResizeBilinear2dModule(torch.nn.Module):
def forward(self, x):
a = torch.nn.functional.interpolate(
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_bmm.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestBMM(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class BMM(torch.nn.Module):
def __init__(self):
super().__init__()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_cat.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,9 @@


class TestCat(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class Cat(torch.nn.Module):
def __init__(self, dim=0):
super().__init__()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_ceil.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestCeil(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class Ceil(torch.nn.Module):
def __init__(self):
super().__init__()
Expand Down
4 changes: 4 additions & 0 deletions backends/xnnpack/test/ops/test_check_quant_params.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,9 @@


class TestCheckQuantParams(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

def create_invalid_value_injector(
self, invalid_value, is_per_channel=False, is_zp=False
):
Expand Down Expand Up @@ -46,6 +49,7 @@ def inject_invalid_scale_in_per_tensor(aten):
return inject_invalid_scale_in_per_tensor

def _test_check_quant_message(self, ep_modifier, expected_message):
torch._dynamo.reset()
mod = torch.nn.Linear(10, 10)
quantizer = XNNPACKQuantizer()
captured = export_for_training(mod, (torch.randn(1, 10),)).module()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_clamp.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestClamp(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class Clamp(torch.nn.Module):
def __init__(self, min_val=None, max_val=None):
super().__init__()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_conv1d.py
Original file line number Diff line number Diff line change
Expand Up @@ -19,6 +19,9 @@


class TestConv1d(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class Conv1d(torch.nn.Module):
def __init__(self, dtype: torch.dtype = torch.float):
groups = 1
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_conv2d.py
Original file line number Diff line number Diff line change
Expand Up @@ -170,6 +170,9 @@ def get_inputs(self):


class TestConv2d(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

def _test(
self,
m: torch.nn.Module,
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_div.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestDiv(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class Div(torch.nn.Module):
def __init__(self):
super().__init__()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_elu.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestElu(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class ELU(torch.nn.Module):
def __init__(self):
super().__init__()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_floor.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestFloor(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class Floor(torch.nn.Module):
def __init__(self):
super().__init__()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_hardswish.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestHardswish(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class Hardswish(torch.nn.Module):
def __init__(self):
super().__init__()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_hardtanh.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestHardTanh(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class HardTanh(torch.nn.Module):
def __init__(self, min_val=-1.0, max_val=1.0):
super().__init__()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_leaky_relu.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestLeakyRelu(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class LeakyReLU(torch.nn.Module):
def __init__(self, **kwargs):
super().__init__()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_linear.py
Original file line number Diff line number Diff line change
Expand Up @@ -219,6 +219,9 @@ class TestLinear(unittest.TestCase):
should produce strictly better results compared to Per-Tensor Quantization
"""

def setUp(self):
torch._dynamo.reset()

@staticmethod
def _get_4b_dqconfig() -> QuantizationConfig:
# Returns a QuantizationConfig for 4b dynamic quantization for XNNPACK.
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_lstm.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,9 @@


class TestLSTM(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class LSTMLinear(torch.nn.Module):
def __init__(self, input_size, hidden_size, out_size):
super().__init__()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_max_dim.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestMaxDim(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class Max(torch.nn.Module):
def forward(self, x):
max_values_1, max_indices_1 = torch.max(x, dim=2, keepdim=True)
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_maximum.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestMaximum(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class Maximum(torch.nn.Module):
def __init__(self):
super().__init__()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_maxpool2d.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,6 +15,9 @@


class TestMaxPool2d(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class MaxPool2d(torch.nn.Module):
def __init__(self, kernel_size=3, stride=1, padding=0, dilation=1):
super().__init__()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_mean_dim.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestMeanDim(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class MeanDim(torch.nn.Module):
def __init__(self, dims):
super().__init__()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_minimum.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestMinimum(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class Minimum(torch.nn.Module):
def __init__(self):
super().__init__()
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_multiply.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestMul(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class Mul(torch.nn.Module):
def forward(self, x, y):
z = x * y
Expand Down
3 changes: 3 additions & 0 deletions backends/xnnpack/test/ops/test_negate.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,9 @@


class TestNegate(unittest.TestCase):
def setUp(self):
torch._dynamo.reset()

class Negate(torch.nn.Module):
def __init__(self):
super().__init__()
Expand Down
Loading
Loading