-
Notifications
You must be signed in to change notification settings - Fork 298
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Sequential Training: How to fine-tune on MLM? #1287
Comments
Hi, To clarify, you want to do MLM on some data (potentially from an existing task, but really any text corpus), and then fine-tune on RTE after, correct? The MLM (specifically MLM-simple) task will read from a file and treat each line as an input: jiant/jiant/tasks/lib/mlm_simple.py Lines 58 to 67 in 1ad8628
You would likely want to:
Also note that we only support MLM for BERT, RoBERTa, ALBERT and XLM-R models. If it helps, I can write a quick colab to show this. |
Thank you very much! Yes, that's exactly what I'm trying to do. At the moment I'm experimenting on I made a config file for mlm_simple:
Then I took premises and hypotheses from RTE I'm not sure I understand how to use the load mode correctly, though. Can you please elaborate a bit further? I initialize an MLM run like this.
Here Then I run RTE task like this.
Do I still need to specify |
Sorry for the delay in my reply. Yes, Do let me know if you run into any other issues! |
I'm trying to do Sequential Training based on this example. I want to use MLM as the intermediate task and RTE as the target task. Expectedly, runscript.py in download_data doesn't support MLM.
As far as I understand, I can repurpose config.json from some other task by editing corresponding paths and provide my own data. However, there a few questions:
Perhaps, I'm missing some easier way to do this. If so, please point me to it. Ideally, I would like to simply repurpose text examples from some previously downloaded task; possibly, from the same target task. Is there a way to do it? Using existing functionality, that is.
The text was updated successfully, but these errors were encountered: