Skip to content

Multilingual T5 model #10

@akutuzov

Description

@akutuzov

Training a multilingual T5-large or T5-XL model on the HPLT 3.1 datasets, aiming at providing a modern alternative to mT5 and mT0.

mT5 was trained on about 1 trillion tokens. Let's may be aim at minimally 50% of this.

More details and context here

Metadata

Metadata

Labels

No labels
No labels

Type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions