Skip to content

BLOOM-zh is a modification from BLOOM. BLOOM-zh is trained extendedly on larger amounts of Traditional Chinese text data while it still maintains its pretrained English ability

CKIP-Joint/BLOOM-zh-1B1

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

13 Commits
 
 

Repository files navigation

BLOOM-zh: Traditional Chinese-Improved BLOOM

The multilingual language model BLOOM-zh features enhanced support for Traditional Chinese over BLOOM. BLOOM-zh has its origins in the open-source BLOOM models presented by BigScience in 2022. The technical specification can be found in the corresponding paper.

The first 1B1 parameter model was released on March 9, 2023. The 3B parameter model was released on April 10, 2023. We will continue to release larger models trained on more tokens with performance/behavior fixes.

Features

  • Performance: BLOOM-zh performs significantly better on Traditional Chinese than the stock BLOOM

  • Free-to-download: model weights are free to download

News

  • [2023.04.10] Bloom-zh 3B is out!

  • [2023.04.10] First iteration of Bloom-zh 1B1, this one pre-trained with 11B Traditional Chinese tokens.

  • [2023.03.09] Our paper for Bloom-zh is out!

  • [2023.03.07] Bloom-zh 1B1 is out!

Getting Started

Model weights

Model Additional training tokens Download
Bloom-zh 1B1 11.5 B https://huggingface.co/ckip-joint/bloom-1b1-zh
Bloom-zh 3B 13 B https://huggingface.co/ckip-joint/bloom-3b-zh

Evaluation

For perplexity evaluation, we used the evaluation pipeline presented in the original megatron-deepspeed library. For evaluating the downstream task performance, please follow the instructions of the Traditional Chinese Evaluation Suite. We evaluated Bloom-zh on following downstream tasks:

  • TTQA: Taiwan Trivia Question Answering. Please refer to our paper
  • DRCD: Please refer to DRCD
  • FGC: Please refer to 科技大擂台

License

The use of BLOOM-zh model weights is subject to the Model License.

Citation

If you find our work useful, please consider citing BLOOM-zh:

@article{ennen2023extending,
  title={Extending the Pre-Training of BLOOM for Improved Support of Traditional Chinese: Models, Methods and Results},
  author={Ennen, Philipp and Hsu, Po-Chun and Hsu, Chan-Jan and Liu, Chang-Le and Wu, Yen-Chen and Liao, Yin-Hsiang and Lin, Chin-Tung and Shiu, Da-Shan and Ma, Wei-Yun},
  journal={arXiv preprint arXiv:2303.04715},
  year={2023}
}

About

BLOOM-zh is a modification from BLOOM. BLOOM-zh is trained extendedly on larger amounts of Traditional Chinese text data while it still maintains its pretrained English ability

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •