My Implementation of LiLT: A Simple yet Effective Language-Independent Layout Transformer,for Structured Document Understanding
LiLT can be pre-trained on the structured documents of a single language and then directly fine-tuned on other languages with the corresponding off-the-shelf monolingual/multilingual pre-trained textual models. Experimental results on eight languages have shown that LiLT can achieve competitive or even superior performance on diverse widely-used downstream benchmarks, which enables language-independent benefit from the pre-training of document layout structure
An interactive demo for the same can be found out here
pip install -q pytesseract
sudo apt install -q tesseract-ocr
pip install -q transformers
pip install -q pytorch-lightning
pip install -q einops
pip install -q tqdm
pip install -q 'Pillow==7.1.2'
pip install -q wandb
pip install -q torchmetrics
- The training of LiLT from scratch with PyTorch Lightning can be referred here
Currently, I used the following configurations:
config = {
"hidden_dropout_prob": 0.1,
"hidden_size_t": 768,
"hidden_size" : 768,
"hidden_size_l": 768 // 6,
"intermediate_ff_size_factor": 4,
"max_2d_position_embeddings": 1001,
"max_seq_len_l": 512,
"max_seq_len_t" : 512,
"num_attention_heads": 12,
"num_hidden_layers": 12,
'dim_head' : 64,
"shape_size": 96,
"vocab_size": 50265,
"eps": 1e-12,
"fine_tune" : True
}
- The results of all the experiments can be found out here
The same weights can be downloaded by the command as follows:
pip install gdown
gdown 1eRV4fS_LFwI5MHqcRwLUNQZgewxI6Se_
- The script of the same can be found out here
MIT
@inproceedings{wang-etal-2022-lilt,
title = "{L}i{LT}: A Simple yet Effective Language-Independent Layout Transformer for Structured Document Understanding",
author={Wang, Jiapeng and Jin, Lianwen and Ding, Kai},
booktitle = "Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
month = may,
year = "2022",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2022.acl-long.534",
doi = "10.18653/v1/2022.acl-long.534",
pages = "7747--7757",
}