CogLTX is a framework to apply current BERT-like pretrained language models to long texts. CogLTX does not need new Transformer structures or pretraining, but want to put forward a solution in finetuning and inference. See the paper (http://keg.cs.tsinghua.edu.cn/jietang/publications/NIPS20-Ding-et-al-CogLTX.pdf) for details.
This repo is a preview version, we will work out a more easy-to-use version in the future. You need to wait for some time due to another project for the first author, and be assured that we are not forgetting!
setup_env.sh describe how to set up an environment by conda. The most important part is
pip install torch==1.3.1 torchvision==0.4.2 transformers==2.4.1 pytorch-lightning==0.6 gensim ujson fuzzywuzzy
Note that the backward compatiblity of pytorch-lightning is not good, and some bugs in the 0.6 version are fixed manually in the code. We will change the framework in the later version of CogLTX.
The preprocess is different for different types of data, but in general can be split into 3 phases:
- Read and tokenize the data.
- Split the long text into blocks by
Buffer.split_document_into_blocks(document, tokenizer, cnt=0, hard=True, properties=None). The
hardparameter switches the mode, i.e. whether to use dynamic programming to decide the separation.
propertiesare associated information with the document, usually labels.
- Save them as a list of samples by pickle.
See the folders (
newsqa hotpotqa 20news) for details.
To train CogLTX on a specific task, we need write an entry like
run_newsqa.py, run_hotpotqa.py, run_20news.py, which includes the configuration, postprocess and evaluation.
The data of NewsQA, hotpotQA and 20news can be found in the original dataset paper, but we do not release the codes and data about Alibaba due to commercial reasons.