Code is heavily based on OpenAI's gpt-2 code samples and their original transformer language model
The initial codebase lacks code for training and the latter is oriented towards finetuning the model for supervised tasks. This repository aims to provide an implementation allowing straightforward text generation from the model.
- numpy
- tensorflow>=1.12.0
python3 train.py path/to/text/corpus
python3 sample.py path/to/model/checkpoint path/to/vocab path/to/hyperparams