Difference between revisions of "Large-Scale Named Entity Disambiguation based on Wikipedia Data"

From Wikipedia Quality
Jump to: navigation, search
(New study: Large-Scale Named Entity Disambiguation based on Wikipedia Data)
 
(Adding wikilinks)
Line 1: Line 1:
'''Large-Scale Named Entity Disambiguation based on Wikipedia Data''' - scientific work related to Wikipedia quality published in 2007, written by Silviu Cucerzan.
+
'''Large-Scale Named Entity Disambiguation based on Wikipedia Data''' - scientific work related to [[Wikipedia quality]] published in 2007, written by [[Silviu Cucerzan]].
  
 
== Overview ==
 
== Overview ==
This paper presents a large-scale system for the recognition and semantic disambiguation of named entities based on information extracted from a large encyclopedic collection and Web search results. It describes in detail the disambiguation paradigm employed and the information extraction process from Wikipedia. Through a process of maximizing the agreement between the contextual information extracted from Wikipedia and the context of a document, as well as the agreement among the category tags associated with the candidate entities, the implemented system shows high disambiguation accuracy on both news stories and Wikipedia articles.
+
This paper presents a large-scale system for the recognition and semantic disambiguation of [[named entities]] based on information extracted from a large encyclopedic collection and Web search results. It describes in detail the disambiguation paradigm employed and the [[information extraction]] process from [[Wikipedia]]. Through a process of maximizing the agreement between the contextual information extracted from Wikipedia and the context of a document, as well as the agreement among the category tags associated with the candidate entities, the implemented system shows high disambiguation accuracy on both news stories and Wikipedia articles.

Revision as of 08:16, 17 July 2019

Large-Scale Named Entity Disambiguation based on Wikipedia Data - scientific work related to Wikipedia quality published in 2007, written by Silviu Cucerzan.

Overview

This paper presents a large-scale system for the recognition and semantic disambiguation of named entities based on information extracted from a large encyclopedic collection and Web search results. It describes in detail the disambiguation paradigm employed and the information extraction process from Wikipedia. Through a process of maximizing the agreement between the contextual information extracted from Wikipedia and the context of a document, as well as the agreement among the category tags associated with the candidate entities, the implemented system shows high disambiguation accuracy on both news stories and Wikipedia articles.