Wikipedia-Based Entity Semantifying in Open Information Extraction

From Wikipedia Quality
Revision as of 09:39, 25 January 2020 by Jasmine (talk | contribs) (Overview - Wikipedia-Based Entity Semantifying in Open Information Extraction)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to: navigation, search

Wikipedia-Based Entity Semantifying in Open Information Extraction - scientific work related to Wikipedia quality published in 2017, written by Qiuhao Lu and Youtian Du.

Overview

In the recent years, Open Information Extraction (OIE), an unsupervised strategy which extracts open-domain facts of knowledge from massive heterogeneous text corpora, has achieved impressive improvements. However, the facts (generally represented by a triple) extracted by OIE systems are in lack of clear semantics and then difficult for computer systems to understand. In this paper, authors present a new method to semantify the facts by mapping the string arguments in the triples to the corresponding real-world entities based on the existing knowledge base Wikipedia. First, for each query of string argument, authors consider a set of its most likely mapping entities and assign each candidate a fused prior probability. Then authors calculate the graph-based similarity between candidates as the contextual evidence by propagating semantics on the neighborhood graph of candidates. Finally, authors transform the mapping task into an optimization problem and find the maximum a posteriori (MAP) mapping by combining the prior information and contextual evidence through Bayes' theorem. Due to the fusion of multiple cues and the semantics propagation over the graph, approach improves the performance of the entity semantifying. Experimental results demonstrate the effectiveness of approach.