Training a multilingual `T5-large` or `T5-XL` model on the HPLT 3.1 datasets, aiming at providing a modern alternative to `mT5` and `mT0`. [mT5 was trained on about 1 trillion tokens](https://aclanthology.org/2021.naacl-main.41/). Let's may be aim at minimally 50% of this. More details and context [here](https://github.com/hplt-project/HPLT-WP4/issues/3)