Skip to content

Features

zhezhaoa edited this page Oct 1, 2022 · 2 revisions

TencentPretrain has the following features:

  • Reproducibility TencentPretrain has been tested on many datasets and should match the performances of the original pre-training model implementations such as BERT, GPT-2, ELMo, T5, CLIP.
  • Model modularity TencentPretrain is divided into the following parts: embedding, encoder, target embedding (optional), decoder (optional), and target. Ample modules are implemented in each part. Clear and robust interface allows users to combine modules to construct pre-training models with as few restrictions as possible.
  • Multimodal TencentPretrain supports different modalities such as text, vision, and audio.
  • Model training TencentPretrain supports CPU mode, single GPU mode, distributed training mode, and gigantic model training with DeepSpeed.
  • Model zoo With the help of TencentPretrain, we pre-train and release models of different properties. Proper selection of pre-trained models is important to the performances of downstream tasks.
  • SOTA results TencentPretrain supports comprehensive downstream tasks (e.g. classification and machine reading comprehension) and provides winning solutions of many competitions.
  • Abundant functions TencentPretrain provides abundant functions related with pre-training, such as feature extractor and text generation.
Clone this wiki locally