Skip to content

Topic 04: Entity Linking and Entity Disambiguation

Sherry Lin edited this page Oct 19, 2020 · 17 revisions

Surveys and Analysis

  1. Entity Linking with a Knowledge Base: Issues, Techniques, and Solutions (TKDE 2014) [Paper] 🌟
  2. Neural Entity Linking: A Survey of Models based on Deep Learning (2020) [Paper]
  • a survey of state-of-the-art neural entity linking models
  • a survey of entity embedding techniques;
  • a discussion of recent domain-independent (zero-shot) and cross-lingual EL approaches;
  • a survey of EL applications to modeling word representations.
  1. Error analysis of the well known DeepED model [Link]
  2. Towards holistic Entity Linking: Survey and directions (Information Systems 2020) [Link]
  3. Neural Collective Entity Linking (COLING 2018) [Paper]

Notes for Entity Linking

  1. Candiate Entity Ranking [Notes]
  2. NLP-progress for Entity Linking [Notes] [GitHub]
  3. Recent Trend for Entity Linking [Notes]
  4. Summary of Entity Linking [Notes 1] [Notes 2, check the appendices here]

General Papers

  1. Zero-Shot Entity Linking by Reading Entity Descriptions (ACL 2019) [Paper][Code and Datasets]
  2. Keyphrase Overlap Relatedness for Entity Disambiguation (CIKM 2012), LSH 🌟
  3. Old is Gold: Linguistic Driven Approach for Entity and Relation Linking of Short Text (NAACL 2019), with Relation Linking
  4. Improving Entity Linking by Modeling Latent Relations between Mentions (ACL 2018)
  5. Entity Linking for Tweets (ACL 2013)
  6. Pangloss: Fast Entity Linking in Noisy Text Environments (KDD 2018) [Presentation] 🌟
  7. THINKER - Entity Linking System for Turkish Language (TKDE 2018) 🌟
  8. SHINE+: A General Framework for Domain-Specific Entity Linking with Heterogeneous Information Networks (TKDE 2018) 🌟
  9. SVM ensembles for named entity disambiguation (Computing, 102(4), 2020) [Paper]
  10. Attention-Based Joint Entity Linking with Entity Embedding (Information 10.2, 2019)
  11. Entity Disambiguation Leveraging Multi-Perspective Attention (IEEE Access 2019)
  12. CLEEK: A Chinese Long-text Corpus for Entity Linking (LREC 2020)
  13. A Novel Approach for Analyzing Entity Linking Between Words and Entities for a Knowledge Base Using an Attention-Based Bilinear Joint Learning and Weighted Summation Model (IEEE Access 2019)

Global Coherence

  1. Relational Inference for Wikification (ACL 2013)
  2. Robust Disambiguation of Named Entities in Text (EMNLP 2011)
  3. Liege: Link Entities in Web Lists with Knowledge Base (KDD 2012)🌟
  4. Collective entity linking in web text: A graph-based method (SIGIR 2011)
  5. Collective Annotation of Wikipedia Entities in Web Text (KDD 2009)🌟
  6. Local and Global Algorithms for Disambiguation to Wikipedia (ACL 2011)
  7. Learning entity representation for entity disambiguation (ACL 2013)
  8. Robust named entity disambiguation with random walks (Semantic Web 2018)
  9. Graph ranking for collective named entity disambiguation (ACL 2014)
  10. Personalized page rank for named entity disambiguation (ACL 2015)
  11. Collective entity resolution with multi-focal attention (2016)
  12. To link or not to link? a study on end-to-end tweet entity linking (NAACL 2013)
  13. An entity-topic model for entity linking (EMNLP 2012)
  14. Deep joint entity disambiguation with local neural attention (EMNLP 2017)
  15. Improving entity linking by modeling latent relations between mentions (ACL 2018)
  16. Neural Collective Entity Linking Based on Recurrent Random Walk Network Learning (IJCAI 2019), introduces external knowledge to model the semantic interdependence between different EL decisions
  17. ELDEN: Improved Entity Linking using Densified Knowledge Graphs (NAACL-HLT 2018) [Paper][Code], supervised EL system
  18. KBPearl: A Knowledge Base Population System Supported by Joint Entity and Relation Linking (VLDB 2020) [Paper], with relation linking 🌟
  19. Joint Embedding in Named Entity Linking on Sentence Level [Paper]
  20. Improving Entity Linking by Modeling Latent Relations between Mentions (ACL 2018) [Paper]
  21. High Quality Candidate Generation and Sequential Graph Attention Network for Entity Linking (WWW 2020) [Paper], BERT+SeqGAT
  22. A collective entity linking algorithm with parallel computing on large-scale knowledge base (The Journal of Supercomputing, 2020)
  23. A Novel Path-based Entity Relatedness Measure for Efficient Collective Entity Linking (ISWC 2020)

Relax the Global Coherence Assumption

  1. Joint entity linking with deep reinforcement learning (WWW 2019) [Paper]

Reinforcement learning, apply LSTM to be able to maintain long term memory for previous decisions.

  1. Learning Dynamic Context Augmentation for Global Entity Linking (DCA, ACL 2019) [Paper]

Reinforcement learning, previous decisions are collected as dynamic context to improve the following predictions.

  1. Global Entity Disambiguation with Pretrained Contextualized Embeddings of Words and Entities (2020) [Paper]
  • BERT+MLM.
  • Note: Paper 1,2,3 address ED as a sequential decision task that disambiguates mentions one by one, and uses words and already disambiguated entities to disambiguate new mentions.
  1. Joint Learning of Local and Global Features for Entity Linking via Neural Networks (COLING 2016), [Paper], CNN+RNN
  • Global-RNN utilizes convolutional neural networks to induce the representations for local contexts and takes advantage of recurrent neural networks to adaptively compress variable length sequences of predictions for global constraints.
  1. Dynamic Graph Convolutional Networks for Entity Linking (WWW 2020) [Paper]
  • Resorts to GNN to automatically decide the relevant linked nodes and then generate the global feature vector for every node.
  • Then a score function is proposed to directly utilize the feature to compute the ranking score and do not need other additional inference steps
  1. Pair-Linking for Collective Entity Disambiguation: Two Could Be Better Than All (TKDE 2018) 🌟
  2. CoNEREL: Collective Information Extraction in News Articles (SIGIR 2018, demo of Paper 6) [Paper]
  3. KBPearl: A Knowledge Base Population System Supported by Joint Entity and Relation Linking (VLDB 2020) 🌟
  4. Joint Entity Linking for Web Tables with Hybrid Semantic Matching (ICCS 2020)
  • Converts table entity linking into a sequence decision problem and uses hybrid semantic features to disambiguate the mentions in web tables
  1. Using Knowledge Base Semantics in Context-Aware Entity Linking (DocEng 2019) [Paper]
  • Supervised CEL. Retained the sum, max@1,max@2 and max@3 as global contextual features, which can be seen as a kind of flexibility in selecting and aggregating the relatedness score.
  1. High Quality Candidate Generation and Sequential Graph Attention Network for Entity Linking (WWW 2020)
  • Graph-based models treat all candidate entities equally which may introduce much noise information
  • Sequence models can only observe previous referred entities, ignoring the relevance between the current mention and its subsequent entities
  • Contribution: (1) propose a multi-strategy based candidate generation method to generate high recall candidate sets; (2) design a Sequential Graph Attention Network (SeqGAT) which combines the advantages of graph and sequence methods

Some related works:

  1. Evaluating the Impact of Knowledge Graph Context on Entity Disambiguation Models (CIKM 2020, short) [Paper]

Employ SPASQL to fetch triples of target entities and incorporate the triples as kg context in pre-trained ED models. DCA is used as baselines on the AIDA-CONLL dataset.

  1. Attention-based Deep Reinforcement Learning Model for Pair-Wise Interaction Recommendation (ICISCE 2019)

Claim that the one-pass sequential decision should consider both positive feedback and negative feedback.

Entity Linking with Type Info

  1. Improving Entity Linking through Semantic Reinforced Entity Embeddings (ACL 2020) [Paper] [Data and Code] [Details]
  • Fine-grained semantic types of entities can let the linking models learn contextual commonality about semantic relatedness.
  • fine-grained semantic words appear frequently as apposition (e.g., Defense contractor Raytheon), coreference (e.g., the company) or anonymous mentions (e.g., American defense firms). These fine-grained types of entities can help capture local contexts and relations of entities.
  1. Improving Entity Linking by Modeling Latent Entity Type Information (AAAI 2020) [Paper]
  • Conduct error analysis of the well known DeepED1 model (Ganea and Hofmann 2017) on the development set of AIDA-CoNLL, and found that more than half of their error cases fall into the category of type errors where the predicted entity’s type is different from the golden entity’s type.
  • Inject latent entity type information into the entity embeddings by modeling the immediate context surrounding the mention.
  • Apply pre-trained BERT to represent the entity context.
  1. A joint model for entity analysis: Coreference, typing, and linking (TACL 2014)
  2. Joint entity recognition and disambiguation (EMNLP 2015)
  3. J-nerd: joint named entity recognition and disambiguation with rich linguistic features (TACL 2016)
  • Paper 3,4,5 integrate type information into the entity linking task by jointly NER+EL, which captures the mutual dependency between them using structured CRF. These methods mainly differ in the design of hand-engineered features.
  1. Joint learning of named entity recognition and entity linking. (ACL: Student Research Workshop, 2019)
  • Multi-task learning using learned features by extending Stack-LSTM.
  • Paper 3,4,5,6 rely on extensive annotation of the type of mentions

Meta EL Note: how to combine the outputs of multiple EL tools for providing a unified set of entity annotations?

  1. Better Together - An Ensemble Learner for Combining the Results of Ready-made Entity Linking Systems (SAC 2020) [Paper]
  2. A Novel Ensemble Method for Named Entity Recognition and Disambiguation Based on Neural Network (ISWC 2018)
  3. MicroNeel: Combining NLP Tools to Perform Named Entity Detection and Linking on Microposts (Final Workshop 7 December 2016, Naples. 2016)
  4. Combining open source annotators for entity linking through weighted voting (SEM 2015)

Joint NER and EL

  1. Joint Learning of Named Entity Recognition and Entity Linking (ACL 2019) [Paper]
  • Multi-task learning of NER and EL based on Stack-LSTM approach.
  • Supervised EL system with learned features.
  • Future extension: training entity contextual embeddings and extend it to be cross-lingual.
  1. Re-ranking for joint named-entity recognition and linking (CIKM 2013) [Paper]
  • The reranking model then chooses among the set of all possible mention and entity link labelings for the whole phrase to determine the best choice. It can use features for known relationships between the television channel ABC and the television program The_View to encourage these as outputs. For efficiency, we use the pipeline models to prune the set of all possible candidate mentions and entity links to a manageable size while maintaining high recall. The reranking model can then use more sophisticated features for collective classification over this pruned set.
  • Rely on existing NER tools. Only NER is beneficial to EL, not vice versa.
  • Hand-engineered features.
  • Uses a large number of heuristically obtained Noun phrase (NP) chunks and word n-grams as additional input to the EL stage.
  1. To link or not to link? a study on end-to-end tweet entity linking (NAACL 2013)
  • Only suitable for short-text such as tweets.
  1. Joint entity recognition and disambiguation (EMNLP 2015) [Paper]
  • NER is beneficial to EL. EL is also beneficial to NER.
  • Supervised EL system.
  • Hand-engineered features.
  1. J-nerd: joint named entity recognition and disambiguation with rich linguistic features (TACL 2016) [Paper] [Code]
  • Supervised, non-linear probabilistic graphical model that captures mention spans, mention types, and the mapping of mentions to entities in a knowledge base.
  • Hand-engineered features.
  • Relies on fully labeled training data where each tagged entity needs to have an NER and EL label.
  1. A Joint Model for Entity Analysis: Coreference, Typing, and Linking (TACL 2014) [Paper] [System] [GitHub]
  • Joint learning of entity typing, EL, and coreference.
  • Hand-engineered features.
  1. Contextualized End-to-End Neural Entity Linking [Paper]
  • An end-to-end differentiable neural EL model that jointly performs MD and ED, based on BERT, while eliminating external knowledge so that we can study the impact of external knowledge to the EL model.
  1. Noise-robust Named Entity Understanding for Virtual Assistants [Paper]
  • Combining NER and EL information in a joint reranking module for noisy spoken language queries in the context of a digital voice assistant, our proposed framework improves accuracy in both tasks.
  1. End-to-End Neural Entity Linking [Paper] [Code]
  • The main idea is to consider all possible spans as potential mentions and learn contextual similarity scores over their entity candidates that are useful for both MD and ED decisions.

Unlinkable Mention Predication General

  1. No Noun Phrase Left Behind: Detecting and Typing Unlinkable Entities (EMNLP-ACL 2011)
  2. Entity Linking at Web Scale (AKBC-WEKEX 2012)

Explores handling these unlinkable entities over 3 steps: Detect Entities, Predict Types, Disambiguation

Unsupervised

  1. Using encyclopedic knowledge for named entity disambiguation (2006) [PDF]
  2. Nlpr_kbp in tac 2009 kbp track: A two-stage method to entity linking (TAC workshop 2009)
  3. LCC approaches to knowledge base population at TAC 2010 (TAC workshop 2010)
  4. Linking entities to a knowledge base with query expansion (EMNLP 2011)
  5. LINDEN: Linking named entities with knowledge base via semantic knowledge (WWW 2012)
  6. Linking named entities in tweets with knowledge base via user interest modeling (KDD 2013)
  7. Tagme: On-the-fly annotation of short text fragments (by wikipedia entities) (CIKM 2010)
  8. From names to entities using thematic context distance (CIKM 2011)
  9. Mining evidences for named entity disambiguation (KDD 2013)

Paper 1-9: If the score of a mention is smaller than a NIL threshold, then this mention is predicted as unlinkable. The threshold is learned from the training data.

Supervised

  1. NUS-I2R: Learning a combined system for entity linking (TAC workshop 2010)
  2. LCC approaches to knowledge base population at TAC 2010 (TAC workshop 2010)
  3. I2R-NUS-MSRA at TAC 2011: Entity linking (TAC workshop 2011)
  4. Entity linking with effective acronym expansion, instance selection and topic modeling (IJCAI 2011)
  5. Cross-lingual cross-document coreference with entity linking (TAC workshop 2011)

Utilize the binary classification technique (such as SVM). Positive --> entity_top, Negative --> NIL.

  1. Learning to link entities with knowledge base (NAACL 2010)
  2. Local and global algorithms for disambiguation to wikipedia (ACL 2011)

Design some features for for unlinkable mention prediction, such as the score of the top-ranked candidate and whether the entity mention is detected by some NER as a named entity.

  1. Entity disambiguation for knowledge base population (COLING 2010) [Paper]
  2. HLTCOE efforts in entity linking at TAC KBP 2010 (TAC workshop 2011)

Incorporate the unlinkable mention prediction process into the entity ranking process. They added a NIL entity into the candidate entity set, and considered NIL as a distinct candidate.

  1. A generative entity-mention model for linking entities with knowledge base (ACL 2011)

The model assumes that for the entity mention which refers to some specific entity, the probability of this entity mention generated by this specific entity’s model should be significantly higher than the probability of this mention generated by a general language model.

Entity Embeddings (Pre-trained)

  1. Wembedder: Wikidata entity embedding web service [Intro] [Github] [Web service]
  • Web service: only the "most similar" service.
  1. Pre-trained embeddings for Wikidata [Link]
  2. Pre-trained embeddings [Link]
  • 100-dimention and 50-dimention, parsed by numpy.memmap. However, I can only read one float (instead of a vector) for each entity. No sure whether there is mistake.