-
Notifications
You must be signed in to change notification settings - Fork 28
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
the question about dataset feature #29
Comments
hi there, Yes, in the paper we compare the performance of multiple models on ESC-50 (figure 1), the purpose was to show the advantage of Whisper on that. So we may have code of this in this repo. Since this project is mostly on Whisper, and Whisper has the strongest performance, we only release the feature of Whisper. If you want to get features of other datasets, we do have the code at https://github.com/YuanGongND/whisper-at/tree/main/src/noise_robust_asr/intermediate_feat_extract/esc-50, you can run it by yourself. Hope this helps. -Yuan |
Thank you for your quick answer, |
Can you reproduce the result of TL-TR methods? The only hyper parameter we might tuned for different models is the learning rate. You can try 5X 10X times larger or 5X 10X time smaller. -Yuan |
And are you using the following code? whisper-at/src/whisper_at_train/models.py Lines 112 to 132 in 17d94d6
|
thank you for your response, Yes, I can reproduce the results of Tl-TR. I'm also using the same code as you say, and I will continue to try different learning rates to see how it performs on the baseline. |
Hi Yuan,
The features of the ESC dataset you provided seem to only have whisper-large-v1,But it seems that the provided code includes features from more than one model.
Thanks
The text was updated successfully, but these errors were encountered: