Skip to content

BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding 中文 汉语

License

Notifications You must be signed in to change notification settings

tifoit/BERT-chinese

 
 

Repository files navigation

BERT-chinese

BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding 中文 汉语

requirements

python3

tensorflow >= 1.10

jieba

使用方法

1, 准备数据,参考data文件夹和vocab文件夹,data里空行代表document的分隔

2, 数据处理成tfrecord create_pretraining_data.py

3, 预训练 run_pretraining.py

TODO

实验结果

TODO

TODO LIST

GPU并行训练

License

MIT.

About

BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding 中文 汉语

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%