Skip to content

hiaoxui/long-range-transformers

Repository files navigation

Long Range Transformers

Some variants of transformers are claimed to be able to process long contexts, while most of them were ony tested on pseudo tasks, like LRA, or language modeling, leaving their capability of comprehending long texts to be explored. The goal of this project is to verify the effectiveness of long-range transformers on more practical NLP tasks: Do they really work on NLP tasks concerning with long texts? If not, why, and how can we make it work?

Tasks

  1. Coreference
  2. NLI
  3. Abstractive QA
  4. Extractive QA
  5. Summarization

Datasets

  • Ontonotes for coref
  • DocNLI for NLI
  • Qasper for abstractive QA
  • Triviaqa for extractive QA
  • SummFD and CNN

Model

  • Coarse2fine model for coref (located in this folder)
  • A baseline model for DocNLI (located in this folder)
  • A baseline model for abstractive QA (located in this folder)
  • A baseline model for extractive QA (located in this folder)
  • A baseline model for summarization (located in this folder)

Experiments

Coref

See this doc

2.2 NLI

See this doc

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published