Refactored and Abstracted the GPT-2 model code while putting it beside a flask API and dockerized for easy deployment.
Code and samples from the paper "Language Models are Unsupervised Multitask Learners".
See more details in blog post.
Download the model data (needs gsutil):
- Download the model (Linux):
sh download_model.sh 117M
- Download the model (Windows - Bash):
sh download_model_windows.sh 117M
In case it fails to download the
models/117M/checkpoint file, just create
checkpoint file at
and add these lines:
model_checkpoint_path: "model.ckpt" all_model_checkpoint_paths: "model.ckpt"
Build the docker image:
Start the services
visit the external IP/localhost to see the live demo.