Skip to content

deepvk/runner

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

27 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

UniversalNER 🚀

UniversalNER: Targeted Distillation from Large Language Models for Open Named Entity Recognition
Wenxuan Zhou*, Sheng Zhang*, Yu Gu, Muhao Chen, Hoifung Poon (*Equal Contribution)

[Project Page] [Demo] [Paper] [Data] [Model]

Release

  • [12/4] We add the evaluation code for NER datasets.
  • [9/14] We add our training code for finetuning the LLama base model with UniversalNER data.
  • [8/11] We release two more UniNER models, UniNER-7B-type-sup and UniNER-7B-all, which were finetuned on ChatGPT-generated data and 40 supervised datasets of various domains and offers better NER performance.
  • [8/10] We have released the inference code for running the model checkpoints. The code for pretraining and evaluation will be released soon.

Model License Data License Usage and License Notices: The data and model checkpoints are intended and licensed for research use only. They are also restricted to uses that follow the license agreement of LLaMA, Vicuna, and ChatGPT.

Install

This project relies on vllm. Ensure you have gcc version 5 or later, and CUDA versions between 11.0 and 11.8, as specified in the installation requirements for vllm.

  1. Clone this repository and navigate to the folder
git clone https://github.com/universal-ner/universal-ner.git
cd universal-ner
  1. Install the required packages
pip install -r requirements.txt

Demo

We use vllm for inference. The inference can be run with a single V100 16G GPU.

Gradio Web UI

To launch a Gradio demo locally, run the following command:

python -m src.serve.gradio_server \
    --model_path Universal-NER/UniNER-7B-type \
    --tensor_parallel_size 1 \
    --max_input_length 512

CLI Inference

Run the following command to use vllm for inference:

python -m src.serve.cli \
    --model_path Universal-NER/UniNER-7B-type \
    --tensor_parallel_size 1 \
    --max_input_length 512

Run the following command to use Huggingface Transformers for inference:

python -m src.serve.hf \
    --model_path Universal-NER/UniNER-7B-type \
    --tensor_parallel_size 1 \
    --max_input_length 512

Finetuning

Our training code is adapted from FastChat. See here for how to finetune the LLama base model with UniversalNER data.

Evaluation

To execute the evaluation process:

python -m src.eval.evaluate \
    --model_path Universal-NER/UniNER-7B-type \
    --data_path ./src/eval/test_data/CrossNER_AI.json \
    --tensor_parallel_size 1

Due to licensing restrictions associated with many NER datasets used in our evaluation, only the CrossNER and MIT datasets are included in this repository.

Citation

If you find UniversalNER helpful for your research and applications, please cite using this BibTeX:

@article{zhou2023universalner,
      title={UniversalNER: Targeted Distillation from Large Language Models for Open Named Entity Recognition}, 
      author={Wenxuan Zhou and Sheng Zhang and Yu Gu and Muhao Chen and Hoifung Poon},
      year={2023},
      eprint={2308.03279},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 94.1%
  • JavaScript 2.5%
  • HTML 1.9%
  • Other 1.5%