diff --git a/examples/asr/emformer_rnnt/pipeline_demo.py b/examples/asr/emformer_rnnt/pipeline_demo.py index 4821c1b8ca..782bc1d539 100644 --- a/examples/asr/emformer_rnnt/pipeline_demo.py +++ b/examples/asr/emformer_rnnt/pipeline_demo.py @@ -16,10 +16,7 @@ from common import MODEL_TYPE_LIBRISPEECH, MODEL_TYPE_MUSTC, MODEL_TYPE_TEDLIUM3 from mustc.dataset import MUSTC from torchaudio.pipelines import EMFORMER_RNNT_BASE_LIBRISPEECH, RNNTBundle -from torchaudio.prototype.pipelines import ( - EMFORMER_RNNT_BASE_MUSTC, - EMFORMER_RNNT_BASE_TEDLIUM3, -) +from torchaudio.prototype.pipelines import EMFORMER_RNNT_BASE_MUSTC, EMFORMER_RNNT_BASE_TEDLIUM3 logger = logging.getLogger(__name__) diff --git a/examples/hubert/finetune.py b/examples/hubert/finetune.py index 601ef249cd..ecd0da60ab 100644 --- a/examples/hubert/finetune.py +++ b/examples/hubert/finetune.py @@ -9,11 +9,7 @@ import logging import pathlib -from argparse import ( - ArgumentDefaultsHelpFormatter, - ArgumentParser, - RawDescriptionHelpFormatter, -) +from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, RawDescriptionHelpFormatter from typing import Tuple from lightning import HuBERTFineTuneModule diff --git a/examples/hubert/train.py b/examples/hubert/train.py index 27108028ed..69d002133f 100644 --- a/examples/hubert/train.py +++ b/examples/hubert/train.py @@ -6,11 +6,7 @@ import logging import pathlib -from argparse import ( - ArgumentDefaultsHelpFormatter, - ArgumentParser, - RawDescriptionHelpFormatter, -) +from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, RawDescriptionHelpFormatter from typing import Optional, Tuple from lightning import HuBERTPreTrainModule diff --git a/examples/interactive_asr/asr.py b/examples/interactive_asr/asr.py index 8dd4184561..aab5ae3889 100644 --- a/examples/interactive_asr/asr.py +++ b/examples/interactive_asr/asr.py @@ -13,12 +13,7 @@ import logging from fairseq import options -from interactive_asr.utils import ( - add_asr_eval_argument, - get_microphone_transcription, - setup_asr, - transcribe_file, -) +from interactive_asr.utils import add_asr_eval_argument, get_microphone_transcription, setup_asr, transcribe_file def main(args): diff --git a/examples/pipeline_tacotron2/inference.py b/examples/pipeline_tacotron2/inference.py index 4467089a6d..581c65de78 100644 --- a/examples/pipeline_tacotron2/inference.py +++ b/examples/pipeline_tacotron2/inference.py @@ -12,12 +12,7 @@ import torch import torchaudio from datasets import InverseSpectralNormalization -from text.text_preprocessing import ( - available_phonemizers, - available_symbol_set, - get_symbol_list, - text_to_sequence, -) +from text.text_preprocessing import available_phonemizers, available_symbol_set, get_symbol_list, text_to_sequence from torchaudio.models import Tacotron2, tacotron2 as pretrained_tacotron2 from utils import prepare_input_sequence @@ -26,12 +21,8 @@ def parse_args(): r""" Parse commandline arguments. """ - from torchaudio.models.tacotron2 import ( - _MODEL_CONFIG_AND_URLS as tacotron2_config_and_urls, - ) - from torchaudio.models.wavernn import ( - _MODEL_CONFIG_AND_URLS as wavernn_config_and_urls, - ) + from torchaudio.models.tacotron2 import _MODEL_CONFIG_AND_URLS as tacotron2_config_and_urls + from torchaudio.models.wavernn import _MODEL_CONFIG_AND_URLS as wavernn_config_and_urls parser = argparse.ArgumentParser(description=__doc__) parser.add_argument( diff --git a/examples/pipeline_tacotron2/train.py b/examples/pipeline_tacotron2/train.py index a4de0cda99..afc618bea0 100644 --- a/examples/pipeline_tacotron2/train.py +++ b/examples/pipeline_tacotron2/train.py @@ -52,12 +52,7 @@ from datasets import SpectralNormalization, split_process_dataset, text_mel_collate_fn from loss import Tacotron2Loss -from text.text_preprocessing import ( - available_phonemizers, - available_symbol_set, - get_symbol_list, - text_to_sequence, -) +from text.text_preprocessing import available_phonemizers, available_symbol_set, get_symbol_list, text_to_sequence from utils import save_checkpoint diff --git a/test/torchaudio_unittest/backend/sox_io/roundtrip_test.py b/test/torchaudio_unittest/backend/sox_io/roundtrip_test.py index f4b6cdc7be..4185ab9d14 100644 --- a/test/torchaudio_unittest/backend/sox_io/roundtrip_test.py +++ b/test/torchaudio_unittest/backend/sox_io/roundtrip_test.py @@ -2,13 +2,7 @@ from parameterized import parameterized from torchaudio.backend import sox_io_backend -from torchaudio_unittest.common_utils import ( - get_wav_data, - PytorchTestCase, - skipIfNoExec, - skipIfNoSox, - TempDirMixin, -) +from torchaudio_unittest.common_utils import get_wav_data, PytorchTestCase, skipIfNoExec, skipIfNoSox, TempDirMixin from .common import get_enc_params, name_func diff --git a/test/torchaudio_unittest/backend/sox_io/smoke_test.py b/test/torchaudio_unittest/backend/sox_io/smoke_test.py index 4329209bc8..e394161044 100644 --- a/test/torchaudio_unittest/backend/sox_io/smoke_test.py +++ b/test/torchaudio_unittest/backend/sox_io/smoke_test.py @@ -3,12 +3,7 @@ from parameterized import parameterized from torchaudio.backend import sox_io_backend -from torchaudio_unittest.common_utils import ( - get_wav_data, - skipIfNoSox, - TempDirMixin, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_wav_data, skipIfNoSox, TempDirMixin, TorchaudioTestCase from .common import name_func diff --git a/test/torchaudio_unittest/common_utils/case_utils.py b/test/torchaudio_unittest/common_utils/case_utils.py index 29d2efd6a1..af499115ed 100644 --- a/test/torchaudio_unittest/common_utils/case_utils.py +++ b/test/torchaudio_unittest/common_utils/case_utils.py @@ -10,11 +10,7 @@ import torch import torchaudio from torch.testing._internal.common_utils import TestCase as PytorchTestCase -from torchaudio._internal.module_utils import ( - is_kaldi_available, - is_module_available, - is_sox_available, -) +from torchaudio._internal.module_utils import is_kaldi_available, is_module_available, is_sox_available from .backend_utils import set_audio_backend diff --git a/test/torchaudio_unittest/datasets/cmuarctic_test.py b/test/torchaudio_unittest/datasets/cmuarctic_test.py index b7fe94dd85..f8d782190d 100644 --- a/test/torchaudio_unittest/datasets/cmuarctic_test.py +++ b/test/torchaudio_unittest/datasets/cmuarctic_test.py @@ -2,13 +2,7 @@ from pathlib import Path from torchaudio.datasets import cmuarctic -from torchaudio_unittest.common_utils import ( - get_whitenoise, - normalize_wav, - save_wav, - TempDirMixin, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_whitenoise, normalize_wav, save_wav, TempDirMixin, TorchaudioTestCase def get_mock_dataset(root_dir): diff --git a/test/torchaudio_unittest/datasets/commonvoice_test.py b/test/torchaudio_unittest/datasets/commonvoice_test.py index 1607c47c1c..ba4155f485 100644 --- a/test/torchaudio_unittest/datasets/commonvoice_test.py +++ b/test/torchaudio_unittest/datasets/commonvoice_test.py @@ -5,13 +5,7 @@ from torch import Tensor from torchaudio.datasets import COMMONVOICE -from torchaudio_unittest.common_utils import ( - get_whitenoise, - normalize_wav, - save_wav, - TempDirMixin, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_whitenoise, normalize_wav, save_wav, TempDirMixin, TorchaudioTestCase _ORIGINAL_EXT_AUDIO = COMMONVOICE._ext_audio _SAMPLE_RATE = 48000 diff --git a/test/torchaudio_unittest/datasets/dr_vctk_test.py b/test/torchaudio_unittest/datasets/dr_vctk_test.py index 26f733486d..6caa58d8f5 100644 --- a/test/torchaudio_unittest/datasets/dr_vctk_test.py +++ b/test/torchaudio_unittest/datasets/dr_vctk_test.py @@ -2,12 +2,7 @@ import pytest from torchaudio.datasets import dr_vctk -from torchaudio_unittest.common_utils import ( - get_whitenoise, - save_wav, - TempDirMixin, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_whitenoise, save_wav, TempDirMixin, TorchaudioTestCase _SUBSETS = ["train", "test"] diff --git a/test/torchaudio_unittest/datasets/fluentcommands_test.py b/test/torchaudio_unittest/datasets/fluentcommands_test.py index 944cb95e42..00b9d34f36 100644 --- a/test/torchaudio_unittest/datasets/fluentcommands_test.py +++ b/test/torchaudio_unittest/datasets/fluentcommands_test.py @@ -5,12 +5,7 @@ from pathlib import Path from torchaudio.datasets import fluentcommands -from torchaudio_unittest.common_utils import ( - get_whitenoise, - save_wav, - TempDirMixin, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_whitenoise, save_wav, TempDirMixin, TorchaudioTestCase HEADER = ["", "path", "speakerId", "transcription", "action", "object", "location"] SLOTS = ["action", "object", "location"] diff --git a/test/torchaudio_unittest/datasets/gtzan_test.py b/test/torchaudio_unittest/datasets/gtzan_test.py index 24bdb1e9e9..1e7e2b279c 100644 --- a/test/torchaudio_unittest/datasets/gtzan_test.py +++ b/test/torchaudio_unittest/datasets/gtzan_test.py @@ -2,13 +2,7 @@ from pathlib import Path from torchaudio.datasets import gtzan -from torchaudio_unittest.common_utils import ( - get_whitenoise, - normalize_wav, - save_wav, - TempDirMixin, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_whitenoise, normalize_wav, save_wav, TempDirMixin, TorchaudioTestCase def get_mock_dataset(root_dir): diff --git a/test/torchaudio_unittest/datasets/librilightlimited_test.py b/test/torchaudio_unittest/datasets/librilightlimited_test.py index 9c30955652..0577ee2003 100644 --- a/test/torchaudio_unittest/datasets/librilightlimited_test.py +++ b/test/torchaudio_unittest/datasets/librilightlimited_test.py @@ -1,12 +1,7 @@ import os from torchaudio.datasets import librilight_limited -from torchaudio_unittest.common_utils import ( - get_whitenoise, - save_wav, - TempDirMixin, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_whitenoise, save_wav, TempDirMixin, TorchaudioTestCase # Used to generate a unique transcript for each dummy audio file diff --git a/test/torchaudio_unittest/datasets/librispeech_test_impl.py b/test/torchaudio_unittest/datasets/librispeech_test_impl.py index 0eeebdbe58..4116fe2d6f 100644 --- a/test/torchaudio_unittest/datasets/librispeech_test_impl.py +++ b/test/torchaudio_unittest/datasets/librispeech_test_impl.py @@ -2,12 +2,7 @@ from pathlib import Path from torchaudio.datasets import librispeech -from torchaudio_unittest.common_utils import ( - get_whitenoise, - normalize_wav, - save_wav, - TempDirMixin, -) +from torchaudio_unittest.common_utils import get_whitenoise, normalize_wav, save_wav, TempDirMixin # Used to generate a unique transcript for each dummy audio file _NUMBERS = ["ZERO", "ONE", "TWO", "THREE", "FOUR", "FIVE", "SIX", "SEVEN", "EIGHT", "NINE"] diff --git a/test/torchaudio_unittest/datasets/libritts_test.py b/test/torchaudio_unittest/datasets/libritts_test.py index a0499686d5..d04247ea79 100644 --- a/test/torchaudio_unittest/datasets/libritts_test.py +++ b/test/torchaudio_unittest/datasets/libritts_test.py @@ -2,13 +2,7 @@ from pathlib import Path from torchaudio.datasets.libritts import LIBRITTS -from torchaudio_unittest.common_utils import ( - get_whitenoise, - normalize_wav, - save_wav, - TempDirMixin, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_whitenoise, normalize_wav, save_wav, TempDirMixin, TorchaudioTestCase _UTTERANCE_IDS = [ [19, 198, "000000", "000000"], diff --git a/test/torchaudio_unittest/datasets/ljspeech_test.py b/test/torchaudio_unittest/datasets/ljspeech_test.py index 565f12f283..485e0c382d 100644 --- a/test/torchaudio_unittest/datasets/ljspeech_test.py +++ b/test/torchaudio_unittest/datasets/ljspeech_test.py @@ -3,13 +3,7 @@ from pathlib import Path from torchaudio.datasets import ljspeech -from torchaudio_unittest.common_utils import ( - get_whitenoise, - normalize_wav, - save_wav, - TempDirMixin, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_whitenoise, normalize_wav, save_wav, TempDirMixin, TorchaudioTestCase _TRANSCRIPTS = [ "Test transcript 1", diff --git a/test/torchaudio_unittest/datasets/musdb_hq_test.py b/test/torchaudio_unittest/datasets/musdb_hq_test.py index b3684132a9..18ae40d218 100644 --- a/test/torchaudio_unittest/datasets/musdb_hq_test.py +++ b/test/torchaudio_unittest/datasets/musdb_hq_test.py @@ -4,12 +4,7 @@ from parameterized import parameterized from torchaudio.datasets import musdb_hq from torchaudio.datasets.musdb_hq import _VALIDATION_SET -from torchaudio_unittest.common_utils import ( - get_whitenoise, - save_wav, - TempDirMixin, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_whitenoise, save_wav, TempDirMixin, TorchaudioTestCase _SOURCE_SETS = [ (None,), diff --git a/test/torchaudio_unittest/datasets/quesst14_test.py b/test/torchaudio_unittest/datasets/quesst14_test.py index 9f1472a05d..9ee2128006 100644 --- a/test/torchaudio_unittest/datasets/quesst14_test.py +++ b/test/torchaudio_unittest/datasets/quesst14_test.py @@ -4,12 +4,7 @@ from parameterized import parameterized from torchaudio.datasets import quesst14 -from torchaudio_unittest.common_utils import ( - get_whitenoise, - save_wav, - TempDirMixin, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_whitenoise, save_wav, TempDirMixin, TorchaudioTestCase def _get_filename(folder, index): diff --git a/test/torchaudio_unittest/datasets/speechcommands_test.py b/test/torchaudio_unittest/datasets/speechcommands_test.py index 94ebb4ac5e..6c63d8085c 100644 --- a/test/torchaudio_unittest/datasets/speechcommands_test.py +++ b/test/torchaudio_unittest/datasets/speechcommands_test.py @@ -2,13 +2,7 @@ from pathlib import Path from torchaudio.datasets import speechcommands -from torchaudio_unittest.common_utils import ( - get_whitenoise, - normalize_wav, - save_wav, - TempDirMixin, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_whitenoise, normalize_wav, save_wav, TempDirMixin, TorchaudioTestCase _LABELS = [ "bed", diff --git a/test/torchaudio_unittest/datasets/tedlium_test.py b/test/torchaudio_unittest/datasets/tedlium_test.py index f7c2e299e4..61846de1be 100644 --- a/test/torchaudio_unittest/datasets/tedlium_test.py +++ b/test/torchaudio_unittest/datasets/tedlium_test.py @@ -3,13 +3,7 @@ from pathlib import Path from torchaudio.datasets import tedlium -from torchaudio_unittest.common_utils import ( - get_whitenoise, - save_wav, - skipIfNoSox, - TempDirMixin, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_whitenoise, save_wav, skipIfNoSox, TempDirMixin, TorchaudioTestCase # Used to generate a unique utterance for each dummy audio file _UTTERANCES = [ diff --git a/test/torchaudio_unittest/datasets/vctk_test.py b/test/torchaudio_unittest/datasets/vctk_test.py index 88f66074e5..e49e1dbcec 100644 --- a/test/torchaudio_unittest/datasets/vctk_test.py +++ b/test/torchaudio_unittest/datasets/vctk_test.py @@ -2,13 +2,7 @@ from pathlib import Path from torchaudio.datasets import vctk -from torchaudio_unittest.common_utils import ( - get_whitenoise, - normalize_wav, - save_wav, - TempDirMixin, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_whitenoise, normalize_wav, save_wav, TempDirMixin, TorchaudioTestCase # Used to generate a unique transcript for each dummy audio file _TRANSCRIPT = [ diff --git a/test/torchaudio_unittest/datasets/yesno_test.py b/test/torchaudio_unittest/datasets/yesno_test.py index b6c0b48f11..da4429b2cd 100644 --- a/test/torchaudio_unittest/datasets/yesno_test.py +++ b/test/torchaudio_unittest/datasets/yesno_test.py @@ -2,13 +2,7 @@ from pathlib import Path from torchaudio.datasets import yesno -from torchaudio_unittest.common_utils import ( - get_whitenoise, - normalize_wav, - save_wav, - TempDirMixin, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_whitenoise, normalize_wav, save_wav, TempDirMixin, TorchaudioTestCase def get_mock_data(root_dir, labels): diff --git a/test/torchaudio_unittest/example/souce_sepration/wsj0mix_test.py b/test/torchaudio_unittest/example/souce_sepration/wsj0mix_test.py index 5fa51b442b..dd67a8e37c 100644 --- a/test/torchaudio_unittest/example/souce_sepration/wsj0mix_test.py +++ b/test/torchaudio_unittest/example/souce_sepration/wsj0mix_test.py @@ -1,13 +1,7 @@ import os from source_separation.utils.dataset import wsj0mix -from torchaudio_unittest.common_utils import ( - get_whitenoise, - normalize_wav, - save_wav, - TempDirMixin, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_whitenoise, normalize_wav, save_wav, TempDirMixin, TorchaudioTestCase _FILENAMES = [ diff --git a/test/torchaudio_unittest/example/tacotron2/tacotron2_loss_cpu_test.py b/test/torchaudio_unittest/example/tacotron2/tacotron2_loss_cpu_test.py index 9b79be2680..bfe019ba41 100644 --- a/test/torchaudio_unittest/example/tacotron2/tacotron2_loss_cpu_test.py +++ b/test/torchaudio_unittest/example/tacotron2/tacotron2_loss_cpu_test.py @@ -1,11 +1,7 @@ import torch from torchaudio_unittest.common_utils import PytorchTestCase -from .tacotron2_loss_impl import ( - Tacotron2LossGradcheckTests, - Tacotron2LossShapeTests, - Tacotron2LossTorchscriptTests, -) +from .tacotron2_loss_impl import Tacotron2LossGradcheckTests, Tacotron2LossShapeTests, Tacotron2LossTorchscriptTests class TestTacotron2LossShapeFloat32CPU(Tacotron2LossShapeTests, PytorchTestCase): diff --git a/test/torchaudio_unittest/example/tacotron2/tacotron2_loss_gpu_test.py b/test/torchaudio_unittest/example/tacotron2/tacotron2_loss_gpu_test.py index 2427a9c604..030e83d740 100644 --- a/test/torchaudio_unittest/example/tacotron2/tacotron2_loss_gpu_test.py +++ b/test/torchaudio_unittest/example/tacotron2/tacotron2_loss_gpu_test.py @@ -1,11 +1,7 @@ import torch from torchaudio_unittest.common_utils import PytorchTestCase, skipIfNoCuda -from .tacotron2_loss_impl import ( - Tacotron2LossGradcheckTests, - Tacotron2LossShapeTests, - Tacotron2LossTorchscriptTests, -) +from .tacotron2_loss_impl import Tacotron2LossGradcheckTests, Tacotron2LossShapeTests, Tacotron2LossTorchscriptTests @skipIfNoCuda diff --git a/test/torchaudio_unittest/functional/autograd_impl.py b/test/torchaudio_unittest/functional/autograd_impl.py index 8e51686283..b26cade6af 100644 --- a/test/torchaudio_unittest/functional/autograd_impl.py +++ b/test/torchaudio_unittest/functional/autograd_impl.py @@ -6,12 +6,7 @@ from parameterized import parameterized from torch import Tensor from torch.autograd import gradcheck, gradgradcheck -from torchaudio_unittest.common_utils import ( - get_spectrogram, - get_whitenoise, - rnnt_utils, - TestBaseMixin, -) +from torchaudio_unittest.common_utils import get_spectrogram, get_whitenoise, rnnt_utils, TestBaseMixin class Autograd(TestBaseMixin): diff --git a/test/torchaudio_unittest/functional/functional_cpu_test.py b/test/torchaudio_unittest/functional/functional_cpu_test.py index b4e3cc0938..1bf3b99f6f 100644 --- a/test/torchaudio_unittest/functional/functional_cpu_test.py +++ b/test/torchaudio_unittest/functional/functional_cpu_test.py @@ -3,11 +3,7 @@ import torch import torchaudio.functional as F from parameterized import parameterized -from torchaudio_unittest.common_utils import ( - PytorchTestCase, - skipIfNoSox, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import PytorchTestCase, skipIfNoSox, TorchaudioTestCase from .functional_impl import Functional, FunctionalCPUOnly diff --git a/test/torchaudio_unittest/functional/librosa_compatibility_test_impl.py b/test/torchaudio_unittest/functional/librosa_compatibility_test_impl.py index 4de769f1be..4e8d4d3d5f 100644 --- a/test/torchaudio_unittest/functional/librosa_compatibility_test_impl.py +++ b/test/torchaudio_unittest/functional/librosa_compatibility_test_impl.py @@ -13,12 +13,7 @@ import numpy as np -from torchaudio_unittest.common_utils import ( - get_spectrogram, - get_whitenoise, - nested_params, - TestBaseMixin, -) +from torchaudio_unittest.common_utils import get_spectrogram, get_whitenoise, nested_params, TestBaseMixin @unittest.skipIf(not LIBROSA_AVAILABLE, "Librosa not available") diff --git a/test/torchaudio_unittest/functional/torchscript_consistency_impl.py b/test/torchaudio_unittest/functional/torchscript_consistency_impl.py index 961f905a25..dfe30cf834 100644 --- a/test/torchaudio_unittest/functional/torchscript_consistency_impl.py +++ b/test/torchaudio_unittest/functional/torchscript_consistency_impl.py @@ -5,12 +5,7 @@ import torchaudio.functional as F from parameterized import parameterized from torchaudio_unittest import common_utils -from torchaudio_unittest.common_utils import ( - skipIfRocm, - TempDirMixin, - TestBaseMixin, - torch_script, -) +from torchaudio_unittest.common_utils import skipIfRocm, TempDirMixin, TestBaseMixin, torch_script class Functional(TempDirMixin, TestBaseMixin): diff --git a/test/torchaudio_unittest/models/ctc_decoder_test.py b/test/torchaudio_unittest/models/ctc_decoder_test.py index 55533bb107..3b7c0fe779 100644 --- a/test/torchaudio_unittest/models/ctc_decoder_test.py +++ b/test/torchaudio_unittest/models/ctc_decoder_test.py @@ -2,12 +2,7 @@ import torch from parameterized import parameterized -from torchaudio_unittest.common_utils import ( - get_asset_path, - skipIfNoCtcDecoder, - TempDirMixin, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_asset_path, skipIfNoCtcDecoder, TempDirMixin, TorchaudioTestCase NUM_TOKENS = 8 diff --git a/test/torchaudio_unittest/models/rnnt_decoder/rnnt_decoder_cpu_test.py b/test/torchaudio_unittest/models/rnnt_decoder/rnnt_decoder_cpu_test.py index 508bcc925d..47b6687c3e 100644 --- a/test/torchaudio_unittest/models/rnnt_decoder/rnnt_decoder_cpu_test.py +++ b/test/torchaudio_unittest/models/rnnt_decoder/rnnt_decoder_cpu_test.py @@ -1,8 +1,6 @@ import torch from torchaudio_unittest.common_utils import PytorchTestCase -from torchaudio_unittest.models.rnnt_decoder.rnnt_decoder_test_impl import ( - RNNTBeamSearchTestImpl, -) +from torchaudio_unittest.models.rnnt_decoder.rnnt_decoder_test_impl import RNNTBeamSearchTestImpl class RNNTBeamSearchFloat32CPUTest(RNNTBeamSearchTestImpl, PytorchTestCase): diff --git a/test/torchaudio_unittest/models/rnnt_decoder/rnnt_decoder_gpu_test.py b/test/torchaudio_unittest/models/rnnt_decoder/rnnt_decoder_gpu_test.py index 8917ee3837..e5b0ad33c7 100644 --- a/test/torchaudio_unittest/models/rnnt_decoder/rnnt_decoder_gpu_test.py +++ b/test/torchaudio_unittest/models/rnnt_decoder/rnnt_decoder_gpu_test.py @@ -1,8 +1,6 @@ import torch from torchaudio_unittest.common_utils import PytorchTestCase, skipIfNoCuda -from torchaudio_unittest.models.rnnt_decoder.rnnt_decoder_test_impl import ( - RNNTBeamSearchTestImpl, -) +from torchaudio_unittest.models.rnnt_decoder.rnnt_decoder_test_impl import RNNTBeamSearchTestImpl @skipIfNoCuda diff --git a/test/torchaudio_unittest/models/tacotron2/model_test_cpu_test.py b/test/torchaudio_unittest/models/tacotron2/model_test_cpu_test.py index 4efb9f960e..cfb7b247fa 100644 --- a/test/torchaudio_unittest/models/tacotron2/model_test_cpu_test.py +++ b/test/torchaudio_unittest/models/tacotron2/model_test_cpu_test.py @@ -1,11 +1,7 @@ import torch from torchaudio_unittest.common_utils import PytorchTestCase -from .model_test_impl import ( - Tacotron2DecoderTests, - Tacotron2EncoderTests, - Tacotron2Tests, -) +from .model_test_impl import Tacotron2DecoderTests, Tacotron2EncoderTests, Tacotron2Tests class TestTacotron2EncoderFloat32CPU(Tacotron2EncoderTests, PytorchTestCase): diff --git a/test/torchaudio_unittest/models/tacotron2/model_test_gpu_test.py b/test/torchaudio_unittest/models/tacotron2/model_test_gpu_test.py index 0bde83e3b0..ac7912e5ce 100644 --- a/test/torchaudio_unittest/models/tacotron2/model_test_gpu_test.py +++ b/test/torchaudio_unittest/models/tacotron2/model_test_gpu_test.py @@ -1,11 +1,7 @@ import torch from torchaudio_unittest.common_utils import PytorchTestCase, skipIfNoCuda -from .model_test_impl import ( - Tacotron2DecoderTests, - Tacotron2EncoderTests, - Tacotron2Tests, -) +from .model_test_impl import Tacotron2DecoderTests, Tacotron2EncoderTests, Tacotron2Tests @skipIfNoCuda diff --git a/test/torchaudio_unittest/models/wav2vec2/fairseq_integration_test.py b/test/torchaudio_unittest/models/wav2vec2/fairseq_integration_test.py index 9bc05cf2c1..770df45af8 100644 --- a/test/torchaudio_unittest/models/wav2vec2/fairseq_integration_test.py +++ b/test/torchaudio_unittest/models/wav2vec2/fairseq_integration_test.py @@ -11,11 +11,7 @@ wav2vec2_large_lv60k, ) from torchaudio.models.wav2vec2.utils import import_fairseq_model -from torchaudio_unittest.common_utils import ( - get_asset_path, - skipIfNoModule, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_asset_path, skipIfNoModule, TorchaudioTestCase def _load_config(*paths): @@ -102,10 +98,7 @@ def _get_model(self, config, num_out=None): from fairseq.models.hubert.hubert import HubertConfig, HubertModel from fairseq.models.hubert.hubert_asr import HubertCtcConfig, HubertEncoder from fairseq.models.wav2vec.wav2vec2 import Wav2Vec2Config, Wav2Vec2Model - from fairseq.models.wav2vec.wav2vec2_asr import ( - Wav2Vec2CtcConfig, - Wav2VecEncoder, - ) + from fairseq.models.wav2vec.wav2vec2_asr import Wav2Vec2CtcConfig, Wav2VecEncoder from fairseq.tasks.hubert_pretraining import HubertPretrainingConfig from omegaconf import OmegaConf diff --git a/test/torchaudio_unittest/models/wav2vec2/huggingface_intergration_test.py b/test/torchaudio_unittest/models/wav2vec2/huggingface_intergration_test.py index 3fd37639b0..16e0c150a6 100644 --- a/test/torchaudio_unittest/models/wav2vec2/huggingface_intergration_test.py +++ b/test/torchaudio_unittest/models/wav2vec2/huggingface_intergration_test.py @@ -2,17 +2,9 @@ import torch from parameterized import parameterized -from torchaudio.models.wav2vec2 import ( - wav2vec2_base, - wav2vec2_large, - wav2vec2_large_lv60k, -) +from torchaudio.models.wav2vec2 import wav2vec2_base, wav2vec2_large, wav2vec2_large_lv60k from torchaudio.models.wav2vec2.utils import import_huggingface_model -from torchaudio_unittest.common_utils import ( - get_asset_path, - skipIfNoModule, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import get_asset_path, skipIfNoModule, TorchaudioTestCase def _load_config(*paths): @@ -76,11 +68,7 @@ def _get_model(self, config): # However, somehow, once "transformers" is imported, `is_module_available` # starts to fail. Therefore, we defer importing "transformers" until # the actual tests are started. - from transformers.models.wav2vec2 import ( - Wav2Vec2Config, - Wav2Vec2ForCTC, - Wav2Vec2Model, - ) + from transformers.models.wav2vec2 import Wav2Vec2Config, Wav2Vec2ForCTC, Wav2Vec2Model if config["architectures"] == ["Wav2Vec2Model"]: return Wav2Vec2Model(Wav2Vec2Config(**config)) diff --git a/test/torchaudio_unittest/models/wav2vec2/model_test.py b/test/torchaudio_unittest/models/wav2vec2/model_test.py index a4f07ded74..4828f4feea 100644 --- a/test/torchaudio_unittest/models/wav2vec2/model_test.py +++ b/test/torchaudio_unittest/models/wav2vec2/model_test.py @@ -12,12 +12,7 @@ wav2vec2_large, wav2vec2_large_lv60k, ) -from torchaudio_unittest.common_utils import ( - skipIfNoCuda, - skipIfNoQengine, - torch_script, - TorchaudioTestCase, -) +from torchaudio_unittest.common_utils import skipIfNoCuda, skipIfNoQengine, torch_script, TorchaudioTestCase TORCH_VERSION: Tuple[int, ...] = tuple(int(x) for x in torch.__version__.split(".")[:2]) if TORCH_VERSION >= (1, 10): diff --git a/test/torchaudio_unittest/sox_effect/dataset_test.py b/test/torchaudio_unittest/sox_effect/dataset_test.py index cbc0d3b0f1..ba3b6edf8e 100644 --- a/test/torchaudio_unittest/sox_effect/dataset_test.py +++ b/test/torchaudio_unittest/sox_effect/dataset_test.py @@ -8,13 +8,7 @@ import numpy as np import torch import torchaudio -from torchaudio_unittest.common_utils import ( - get_whitenoise, - PytorchTestCase, - save_wav, - skipIfNoSox, - TempDirMixin, -) +from torchaudio_unittest.common_utils import get_whitenoise, PytorchTestCase, save_wav, skipIfNoSox, TempDirMixin class RandomPerturbationFile(torch.utils.data.Dataset): diff --git a/test/torchaudio_unittest/transforms/autograd_test_impl.py b/test/torchaudio_unittest/transforms/autograd_test_impl.py index 2a781f49dd..8bd81f0af8 100644 --- a/test/torchaudio_unittest/transforms/autograd_test_impl.py +++ b/test/torchaudio_unittest/transforms/autograd_test_impl.py @@ -5,13 +5,7 @@ import torchaudio.transforms as T from parameterized import parameterized from torch.autograd import gradcheck, gradgradcheck -from torchaudio_unittest.common_utils import ( - get_spectrogram, - get_whitenoise, - nested_params, - rnnt_utils, - TestBaseMixin, -) +from torchaudio_unittest.common_utils import get_spectrogram, get_whitenoise, nested_params, rnnt_utils, TestBaseMixin class _DeterministicWrapper(torch.nn.Module): diff --git a/test/torchaudio_unittest/transforms/librosa_compatibility_test_impl.py b/test/torchaudio_unittest/transforms/librosa_compatibility_test_impl.py index 16f9efa455..118fff0373 100644 --- a/test/torchaudio_unittest/transforms/librosa_compatibility_test_impl.py +++ b/test/torchaudio_unittest/transforms/librosa_compatibility_test_impl.py @@ -4,13 +4,7 @@ import torchaudio.transforms as T from parameterized import param, parameterized from torchaudio._internal.module_utils import is_module_available -from torchaudio_unittest.common_utils import ( - get_sinusoid, - get_spectrogram, - get_whitenoise, - nested_params, - TestBaseMixin, -) +from torchaudio_unittest.common_utils import get_sinusoid, get_spectrogram, get_whitenoise, nested_params, TestBaseMixin LIBROSA_AVAILABLE = is_module_available("librosa") diff --git a/test/torchaudio_unittest/transforms/transforms_test_impl.py b/test/torchaudio_unittest/transforms/transforms_test_impl.py index 4f69c6ca1f..b036ae7665 100644 --- a/test/torchaudio_unittest/transforms/transforms_test_impl.py +++ b/test/torchaudio_unittest/transforms/transforms_test_impl.py @@ -2,12 +2,7 @@ import torchaudio.transforms as T from parameterized import param, parameterized from torchaudio.functional.functional import _get_sinc_resample_kernel -from torchaudio_unittest.common_utils import ( - get_spectrogram, - get_whitenoise, - nested_params, - TestBaseMixin, -) +from torchaudio_unittest.common_utils import get_spectrogram, get_whitenoise, nested_params, TestBaseMixin from torchaudio_unittest.common_utils.psd_utils import psd_numpy diff --git a/tools/convert_voxpopuli_models.py b/tools/convert_voxpopuli_models.py index a744b6b2dd..0be7d5e036 100755 --- a/tools/convert_voxpopuli_models.py +++ b/tools/convert_voxpopuli_models.py @@ -93,9 +93,7 @@ def _main(args): import torch import torchaudio - from torchaudio.models.wav2vec2.utils.import_fairseq import ( - _convert_state_dict as _convert, - ) + from torchaudio.models.wav2vec2.utils.import_fairseq import _convert_state_dict as _convert cfg, state_dict = _load(args.input_file) params = _parse_model_param(cfg, state_dict) diff --git a/torchaudio/prototype/models/conv_emformer.py b/torchaudio/prototype/models/conv_emformer.py index 2a459b2760..e6b4d4f4f1 100644 --- a/torchaudio/prototype/models/conv_emformer.py +++ b/torchaudio/prototype/models/conv_emformer.py @@ -2,11 +2,7 @@ from typing import List, Optional, Tuple import torch -from torchaudio.models.emformer import ( - _EmformerAttention, - _EmformerImpl, - _get_weight_init_gains, -) +from torchaudio.models.emformer import _EmformerAttention, _EmformerImpl, _get_weight_init_gains def _get_activation_module(activation: str) -> torch.nn.Module: diff --git a/torchaudio/sox_effects/__init__.py b/torchaudio/sox_effects/__init__.py index 11f8d0da62..9de818dda6 100644 --- a/torchaudio/sox_effects/__init__.py +++ b/torchaudio/sox_effects/__init__.py @@ -1,12 +1,6 @@ from torchaudio._internal import module_utils as _mod_utils -from .sox_effects import ( - apply_effects_file, - apply_effects_tensor, - effect_names, - init_sox_effects, - shutdown_sox_effects, -) +from .sox_effects import apply_effects_file, apply_effects_tensor, effect_names, init_sox_effects, shutdown_sox_effects if _mod_utils.is_sox_available():