Entity Linking (EL) is the task of recognizing (cf. Named Entity Recognition) and disambiguating (Named Entity Disambiguation) named entities to a knowledge base (e.g. Wikidata, DBpedia, or YAGO). It is sometimes also simply known as Named Entity Recognition and Disambiguation.
EL can be split into two classes of approaches:
- End-to-End: processing a piece of text to extract the entities (i.e. Named Entity Recognition) and then disambiguate these extracted entities to the correct entry in a given knowledge base (e.g. Wikidata, DBpedia, YAGO).
- Disambiguation-Only: contrary to the first approach, this one directly takes gold standard named entities as input and only disambiguates them to the correct entry in a given knowledge base.
Example:
Barack | Obama | was | born | in | Hawaï |
---|---|---|---|---|---|
https://en.wikipedia.org/wiki/Barack_Obama | https://en.wikipedia.org/wiki/Barack_Obama | O | O | O | https://en.wikipedia.org/wiki/Hawaii |
More in details can be found in this survey.
Raiman is the current SOTA in Cross-lingual Entity Linking for WikiDisamb30 and TAC KBP 2010 datasets (note: Mulang’ et al. 2020 is the current Sota for ConLL-AIDA dataset). They construct a type system, and use it to constrain the outputs of a neural network to respect the symbolic structure. They achieve this by reformulating the design problem into a mixed integer problem: create a type system and subsequently train a neural network with it. They propose a 2-step algorithm: 1) heuristic search or stochastic optimization over discrete variables that define a type system informed by an Oracle and a Learnability heuristic, 2) gradient descent to fit classifier parameters. They apply DeepType to the problem of Entity Linking on three standard datasets (i.e. WikiDisamb30, CoNLL (YAGO), TAC KBP 2010) and find that it outperforms all existing solutions by a wide margin, including approaches that rely on a human-designed type system or recent deep learning-based entity embeddings, while explicitly using symbolic information lets it integrate new entities without retraining.
- Micro-Precision: Fraction of correctly disambiguated named entities in the full corpus.
- Macro-Precision: Fraction of correctly disambiguated named entities, averaged by document.
- Gerbil Micro-F1 - strong matching: micro InKB F1 score for correctly linked and disambiguated mentions in the full corpus as computed using the Gerbil platform. InKB means only mentions with valid KB entities are used for evaluation.
- Gerbil Macro-F1 - strong matching: macro InKB F1 score for correctly linked and disambiguated mentions in the full corpus as computed using the Gerbil platform. InKB means only mentions with valid KB entities are used for evaluation.
The AIDA CoNLL-YAGO Dataset by [Hoffart] contains assignments of entities to the mentions of named entities annotated for the original [CoNLL] 2003 NER task. The entities are identified by YAGO2 entity identifier, by Wikipedia URL, or by Freebase mid.
Paper / Source | Micro-Precision | Macro-Precision | Paper / Source | Code |
---|---|---|---|---|
Mulang’ et al. (2020) | 94.94 | - | Evaluating the Impact of Knowledge Graph Context on Entity Disambiguation Models | - |
Raiman et al. (2018) | 94.88 | - | DeepType: Multilingual Entity Linking by Neural Type System Evolution | Official |
Sil et al. (2018) | 94.0 | - | Neural Cross-Lingual Entity Linking | |
Radhakrishnan et al. (2018) | 93.0 | 93.7 | ELDEN: Improved Entity Linking using Densified Knowledge Graphs | |
Le et al. (2018) | 93.07 | - | Improving Entity Linking by Modeling Latent Relations between Mentions | Official |
Ganea and Hofmann (2017) | 92.22 | - | Deep Joint Entity Disambiguation with Local Neural Attention | Link |
Hoffart et al. (2011) | 82.29 | 82.02 | Robust Disambiguation of Named Entities in Text |
Paper / Source | Micro-F1-strong | Macro-F1-strong | Paper / Source | Code |
---|---|---|---|---|
van Hulst et al. (2020) | 83.3 | 81.3 | REL: An Entity Linker Standing on the Shoulders of Giants | Official |
Kolitsas et al. (2018) | 82.6 | 82.4 | End-to-End Neural Entity Linking | Official |
Kannan Ravi et al. (2021) | 83.1 | - | CHOLAN: A Modular Approach for Neural Entity Linking on Wikipedia and Wikidata | Official |
Piccinno et al. (2014) | 70.8 | 73.0 | From TagME to WAT: a new entity annotator | |
Hoffart et al. (2011) | 71.9 | 72.8 | Robust Disambiguation of Named Entities in Text |
The Knowledge Base Population (KBP) Track at TAC 2010 will explore extraction of information about entities with reference to an external knowledge source. Using basic schema for persons, organizations, and locations, nodes in an ontology must be created and populated using unstructured information found in text. A collection of Wikipedia Infoboxes will serve as a rudimentary initial knowledge representation. You can download the dataset from LDC or here.
Paper / Source | Micro-Precision | Macro-Precision | Paper / Source | Code |
---|---|---|---|---|
Raiman et al. (2018) | 90.85 | - | DeepType: Multilingual Entity Linking by Neural Type System Evolution | Official |
Sil et al. (2018) | 87.4 | - | Neural Cross-Lingual Entity Linking | |
Yamada et al. (2016) | 85.2 | - | Joint Learning of the Embedding of Words and Entities for Named Entity Disambiguation |
Evaluating Entity Linking systems in a manner that allows for direct comparison of performance can be difficult. The precise definition of a "correct" annotation can be somewhat subjective and it is easy to make mistakes. To provide a simple example, given the input surface form "Tom Waits", an evaluation dataset might record the dbpedia resource https://dbpedia.org/resource/Tom_Waits
as the correct referent. Yet an annotation system which returns a reference to https://dbpedia.org/resource/PEHDTSCKJBMA
has technically provided an appropriate annotation as this resource is a redirect to https://dbpedia.org/resource/Tom_Waits
. Alternatively if evaluating an End-to-End EL system, then accuracy with respect to word boundaries must be considered e.g. if a system only annotates "Obama" with the URI https://dbpedia.org/resource/Barack_Obama
in the surface form "Barack Obama", then is the system correct or incorrect in its annotation?
Furthermore, the performance of an EL system can be strongly affected by the nature of the content on which the evaluation is performed e.g. news content versus Tweets. Hence comparing the relative performance of two EL systems which have been tested on two different corpora can be fallicious. Rather than allowing these little subjective points to creep into the evaluation of EL systems, it is better to make use of a standard evaluation platform where these assumptions are known and made explicit in the configuration of the experiment.
GERBIL, developed by AKSW is an evaluation platform that is based on the BAT framework. It defines a number of standard experiments which may be run for any given EL service. These experiment types determine how strict the evaluation is with respect to measures such as word boundary alignment and also dictates how much responsibility is assigned to the EL service with respect to Entity Recognition, etc. GERBIL hosts 38 evaluation datasets obtained from a variety of different EL challenges. At present it also has hooks for 17 different EL services which may be included in an experiment.
GERBIL may be used to test your own EL system either by downloading the source code and deploying GERBAL locally, or by making your service available on the web and giving GERBIL a link to your API endpoint. The only condition is that your API must accept input and respond with output in NIF format. It is also possible to upload your own evaluation dataset if you would like to test these services on your own content. Note the dataset must also be in NIF format. The DBpedia Spotlight evaluation dataset is a good example of how to structure your content.
GERBIL does have a number of shortcomings, the most notable of which are:
- There is no way to view the annotations returned by each system you test. These are handled internally by GERBIL and then discarded. This can make it difficult to determine the source of error with an EL system.
- There is no way to observe the candidate list considered for each surface form. This is, of course, a standard problem with any third party EL API, but if one is conducting a detailed investigation into the performance of an EL system, it is important to know if the source of error was the EL algorithm itself, or the candidate retrieval process which failed to identify the correct referent as a candidate. This was listed as an important consideration by Hachey et al.
Nevertheless, GERBIL is an excellent resource for standardising how EL systems are tested and compared. It is also a good starting point for anyone new to Entity Linking as it contains links to a wide variety of EL resources. For more information, see the research paper by [Usbeck].
[Hoffart] Johannes Hoffart, Mohamed Amir Yosef, Ilaria Bordino, Hagen Fürstenau, Manfred Pinkal, Marc Spaniol, Bilyana Taneva, Stefan Thater, and Gerhard Weikum. Robust Disambiguation of Named Entities in Text. EMNLP 2011. https://www.aclweb.org/anthology/D11-1072
[CoNLL] Erik F Tjong Kim Sang and Fien De Meulder. Introduction to the CoNLL-2003 Shared Task: Language-Independent Named Entity Recognition. CoNLL 2003. https://www.aclweb.org/anthology/W03-0419.pdf
[Usbeck] Usbeck et al. GERBIL - General Entity Annotator Benchmarking Framework. WWW 2015. https://svn.aksw.org/papers/2015/WWW_GERBIL/public.pdf