Skip to content

Recent trends of Entity Linking, Disambiguation, and Representation.

Notifications You must be signed in to change notification settings

izuna385/Entity-Linking-Recent-Trends

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

65 Commits
 
 
 
 
 
 

Repository files navigation

Recent Trends of Entity Linking

This repository aims to track the progress in Entity Linking. Studies on how to prepare Entity Representations are also listed, as Entity Representations are mandatory with Entity Linking.

Contents

Sub Contents


Trends (NAACL'21 and ICLR'21)

  • AUTOREGRESSIVE ENTITY RETRIEVAL

    • Traditional entity linking systems suffer from memory consumption due to prior encoding of entities in the knowledge base, computational resource cost due to comparison of all entities in the knowledge base, and cold start problems.

    • Instead of previous architecture, they exploits a sequence-to-sequence one to generate entity names in an autoregressive fashion conditioned on the context. They used constrained Beam Search, forcing to only decode valid entity identifier.

  • Linking Entities to Unseen Knowledge Bases with Arbitrary Schemas

    • Traditional entity linking systems assume that the schema of the knowledge base that ties the predicted entities together is known. They proposed a new method to convert the schema of unknown entities to BERT embedding using attributes and auxiliary tokens.

    • At the same time, they also proposed a training method to deal with unknown attributes.

Trends (~EMNLP'20 and CoNLL'20)

  • In Media Res: A Corpus for Evaluating Named Entity Linking with Creative Works [paper] [codes]

    • They pointed out that there was a lack of annotation guidances and styles in the Named entity linking, specifically for Creative Works. They designed new corpus with the aim of understanding different annotation styles for this domain, including franchise, books, TV shows, and so on.
  • LUKE: Deep Contextualized Entity Representations with Entity-aware Self-attention [Link] [Codes]

    • They proposed new pretraining task based on BERT, in which randomly masked words and entities are predicted in entity-annotated corpus from Wikipedia.

    • Also in pretraining task, they proposed an extended version of the transformer, which considers an entity-aware self-attention and the types of tokens (words or entities) when computing attention scores.

  • Scalable Zero-shot Entity Linking with Dense Entity Retrieval

  • Entity Linking in 100 Languages [paper]

    • They proposed a new formulation for multilingual entity linking, in which language-specific mentions resolve to a language-agnostic Knowledge Base. Also they provide Mewsli-9, a large new multilingual dataset matched to this setting.
  • COMETA: A Corpus for Medical Entity Linking in the Social Media [paper]

    • They presented COMETA, a unique corpus of large scale and coverage which is consisting of 20k English biomedical entity mentions with expert-annotated annotations, linked to SNOMED CT knowledge graph.
  • Zero-shot Entity Linking with Efficient Long Range Sequence Modeling [paper]

    • They propose a new position embeddings initialization method called Embeddings-repeat, which initializes larger position embeddings by repeating the small one from BERT-Base. For the zero-shot entity linking, their method improves the SOTA from 76.06% to 79.08% on its dataset.

Trends(~ACL'20)

  • From Zero to Hero: Human-In-The-Loop Entity Linking in Low Resource Domains [Link]

  • Improving Entity Linking through Semantic Reinforced Entity Embeddings

Trends (~ICLR'20)

  • Pretrained Encyclopedia: Weakly Supervised Knowledge-Pretrained Language Model (ICLR'20) [paper]

    • Adding a binary prediction task classifying whether entity is replaced or not, they trained the language model. For their purpose of checking whether LM itself contains knowledge, they evaluated model with Zero-Shot Fact Completion using Wikidata triplets.
  • K-Adapter: Infusing Knowledge into Pre-Trained Models with Adapters [paper]

    • Previous LM models with knowledge injected learn knowledge during training and fail to continual learning. With adapter module which is proposed, they show entire model including adapter captures richer factual and commonsense knowledge than previous models, like RoBERTa.

Trends (~EMNLP'19, CoNLL'19, ICLR'19)

  • Although there seems to be research related to EL model itself, some are now studying about (how to get) "Entity Representation" itself.

Models for Entity Linking

  • Fine-Grained Evaluation for Entity Linking (EMNLP'19)

  • Learning Dynamic Context Augmentation for Global Entity Linking (EMNLP'19)

  • Fine-Grained Entity Typing for Domain Independent Entity Linking

  • Investigating Entity Knowledge in BERT With Simple Neural End-To-End Entity Linking (CoNLL '19) [paper]

    • They proposed end-to-end EL including candidate generation step for the first time, by casting EL as token-classification problem.

  • Global Entity Disambiguation with Pretrained Contextualized Embeddings of Words and Entities [paper]

Entity Representation

  • Learning Dense Representations for Entity Retrieval (CoNLL '19)

    • paper, repo

    • They proposed no use of alias table(which was based on wikipedia statistics or prepared one) and searching all entities by brute-force/approximate nearest search for linking entity per mention.

  • EntEval: A Holistic Evaluation Benchmark for Entity Representations (EMNLP '19)

  • Learning Entity Representations for Few-Shot Reconstruction of Wikipedia Categories (ICLR '19)

  • Knowledge Enhanced Contextual Word Representations (EMNLP '19) [paper]

    • By injecting KB entity information into the mention-span representations through BERT, they improves the quality of the masked LM and its ability to recall facts.


Trends (~ACL'19)

  • Trends of leveraging all information(e.g. mention's type and definition and documents in which mention exists, etc...) seems to be disappering.

  • Although Wikipedia domain can use its hyperlink(=mention-entity pairs, about 7,500,000) for training linking model, under some domain-specific situations there are not so much mention-entity pairs.

  • Therefore, some papers are now challenging distant-learning and zero-shot learning of Entity linking.

  • BERT-based entity representation learning also emerged.

    • ERNIE: Enhanced Language Representation with Informative Entities [paper]

Baselines (~ACL'18)

  • (Commented @ Nov, 19') In those days, researches for improving Entity linking model itself were flourishing.

  • Bold style indicates its SoTA score of a specific dataset.

Baseline models Year Dataset code Run? Code address
Entity Linking via Joint Encoding of Types,Descriptions,and Context EMNLP2017 CoNLL-YAGO(82.9,acc),ACE2004,ACE2005,WIKI(89.0,f1) Tensorflow Only Traind model is uploaded here
┗ (Very Similar to the above) Joint Multilingual Supervision for Cross-lingual Entity Linking EMNLP2018 TH-Test,McN-Test,TAC2015 Pytorch Checking here
Neural Collective Entity Linking(NCEL) CL2018 CoNLL-YAGO, ACE2004, AQUAINT,TAC2010(91.0,mic-p),WW pytorch Bug here
Improving Entity Linking by Modeling Latent Relations between Mentions ACL2018 CoNLL-YAGO(93.07,mic-acc),AQUAINT,ACE2004,CWEB,WIKI(84.05,f1) pytorch Evaluation Done here
ELDEN NAACL2018 CoNLL-PPD(93.0,p-mic),TAC2010(89.6,mic-p) lua,torch(lua) Bug here
Deep Joint Entity Disambiguation with Local Neural Attention EMNLP2017 CoNLL-YAGO(92.22,mic-acc),CWEB,WW,ACE2004,AQUAINT,MSNBC lua,torch(lua) Train Running(2019/01/15) here
Hierarchical Losses and New Resources for Fine-grainid Entity Typing and Linking ACL2018 Medmentions,Typenet pytorch Bug here
Joint Learning of the Embedding of Words and Entities for Named Entity Disambiguation(Yamada,Shindo) CoNLL2016 CoNLL-YAGO(91.5,mic-acc),CoNLL-PPD(93.1,p-mic),TAC2010(85.5,mic-acc) pytorch/Tensorflow(original), checking Baseline Original
Learning Distributed Representations of Texts and Entities from Knowledge Base(Yamada,Shindo) ACL2017 CoNLL-PPD(94.7,p-mic),TAC2010(87.7,mic-acc) pytorch/Keras(original) checking Torch, Torch, Original

Datasets

General

Note: major datasets for benchmarking this task are listed at BLINK repository.

Multilingual

  • Mewsli-9 Dataset

    • This dataset was created from WikiNews, and covers 9 diverse languages, 5 language families and 6 writing systems.

Domain-Specific

  • Biomedical

    • Medmentions ([Mohan and Li, AKBC '19])

      • MedMentions was created as a benchmark dataset for named entity recognition and entity linking in the biomedical domain.

      • As it contains many concepts that are too broad to be of practical use, St21pv was constructed by filtering out those broad concepts from MedMentions.

    • BC5CDR ([Li et al., '15'])

      • BC5CDR is a dataset created for the BioCreative V Chemical and Disease Mention Recognition task.

      • It comprises 1,500 articles, containing 15,935 chemical and 12,852 disease mentions.

      • The reference knowledge base is MeSH, and almost all mentions have a gold entity in the reference knowledge base.

    • WikiMed and PubMedDS ([Shikhar et al., '20])

      • WIKIMED includes over 650,000 mentions normalized to concepts in UMLS. (Cited)

      • Also, they created annotated corpus PUBMEDDS with more than 5 million normalized mentions. Note that this dataset was created by distant supervision, which leads to cause some noisy annotations.

  • Zero-shot

    • Wikia datasets ([Logeswaran et al., '19])

      • From Wikia hyperlinks and its related themes, they created dataset for evaluating domain generalization of entity linking task.

      • They created 16 worlds dataset, which were split to 8 / 4 / 4 for train / dev / test and completely independent for each other.


Bi-Encoder vs Cross-Encoder


How to Get/Prepare Entity Representations?

  • Based on papers@December,'19


Another Trend: BERT x KB


Entity Linking Introductions

Local Model and Global Model

  • Details are wrintten in Neural Collective Entity Linking. paper

Trend in the Point of local vs global

What is local/global Model?


Misc

  • Bi-encoder based entity linking tutorial and its implementation. [link]

  • Comprehensive survey paper of Entity Linking has appeared.

  • (Commented @ March, 2020) Currently this repository includes papers for both of Entity linking and Entity LM. The former requires encoding entity representation for disambiguation, while the latter means injecting entity knowledge into LM during the training. So they are completely diffrent, although some work of entity LM evaluated its model with entity disambiguation. We will separate them at this repository, in the near future.

  • Cross-lingual EL papers.

    Cited from Gupta et al. (EMNLP '18)

    Cross-lingual Entity Linking (XEL) aims to ground entity mentions written in any language to an English Knowledge Base (KB), such as Wikipedia.

  • Entity Linking for Noisy/Short Texts

    • Aggregated Semantic Matching for Short Text Entity Linking (ACL'18) [paper]

    • Effective Use of Context in Noisy Entity Linking (EMNLP'18) [paper]

  • Multimodal Entity Linking

    • Multimodal Named Entity Disambiguation for Noisy Social Media Posts (ACL '18) [paper]
  • Some other papers

    • List-only Entity Linking paper

    • Joint Learning of Named Entity Recognition and Entity Linking paper