We plan to have a bunch of T5-like language models trained on the HPLT v3 datasets:
- Monolingual models, following more or less the same list of languages as HPLT v2 BERT models
- A large multilingual model, aiming at providing a modern alternative to mT5