You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardexpand all lines: README.md
+27-2
Original file line number
Diff line number
Diff line change
@@ -48,6 +48,25 @@ be saved under `logs/multilingual`. Note that this process is highly memory inte
48
48
12+ GB of GPU memory (requirements are half if fp16 is enabled in AllenNLP, but this [requires custom changes to the library](https://github.com/allenai/allennlp/issues/2149)).
49
49
The training may take 20 or more days to complete all 80 epochs depending on the type of your GPU.
50
50
51
+
### Training on Other Datasets
52
+
53
+
An example config is given for fine-tuning on just English EWT. Just run:
1. When fine-tuning, my scores/metrics show poor performance.
112
131
113
-
It should take about 10 epochs to start seeing good scores coming from all the metrics, and 80 epochs to be competitive with UDPipe Future.
132
+
It should take about 10 epochs to start seeing good scores coming from all the metrics, and 80 epochs to be competitive
133
+
with UDPipe Future.
114
134
115
-
One caveat is that if you use a subset of treebanks for fine-tuning instead of all 124 UD v2.3 treebanks, *you must modify the configuration file*. Make sure to tune the learning rate scheduler to the number of training steps. Copy the [`udify_bert_finetune_multilingual.json`](https://github.com/Hyperparticle/udify/blob/master/config/ud/multilingual/udify_bert_finetune_multilingual.json) config and modify the `"warmup_steps"` and `"start_step"` values. A good initial choice would be to set both to be equal to the number of training batches of one epoch ( run the training script first to see the batches remaining).
135
+
One caveat is that if you use a subset of treebanks for fine-tuning instead of all 124 UD v2.3 treebanks,
136
+
*you must modify the configuration file*. Make sure to tune the learning rate scheduler to the number of
137
+
training steps. Copy the [`udify_bert_finetune_multilingual.json`](https://github.com/Hyperparticle/udify/blob/master/config/ud/multilingual/udify_bert_finetune_multilingual.json)
138
+
config and modify the `"warmup_steps"` and `"start_step"` values. A good initial choice would be to set both to be
139
+
equal to the number of training batches of one epoch (run the training script first to see the batches remaining, to
0 commit comments