Difference between revisions of "Wikipedia-Based Entity Semantifying in Open Information Extraction"

From Wikipedia Quality
Jump to: navigation, search
(Overview - Wikipedia-Based Entity Semantifying in Open Information Extraction)
 
(Int.links)
Line 1: Line 1:
'''Wikipedia-Based Entity Semantifying in Open Information Extraction''' - scientific work related to Wikipedia quality published in 2017, written by Qiuhao Lu and Youtian Du.
+
'''Wikipedia-Based Entity Semantifying in Open Information Extraction''' - scientific work related to [[Wikipedia quality]] published in 2017, written by [[Qiuhao Lu]] and [[Youtian Du]].
  
 
== Overview ==
 
== Overview ==
In the recent years, Open Information Extraction (OIE), an unsupervised strategy which extracts open-domain facts of knowledge from massive heterogeneous text corpora, has achieved impressive improvements. However, the facts (generally represented by a triple) extracted by OIE systems are in lack of clear semantics and then difficult for computer systems to understand. In this paper, authors present a new method to semantify the facts by mapping the string arguments in the triples to the corresponding real-world entities based on the existing knowledge base Wikipedia. First, for each query of string argument, authors consider a set of its most likely mapping entities and assign each candidate a fused prior probability. Then authors calculate the graph-based similarity between candidates as the contextual evidence by propagating semantics on the neighborhood graph of candidates. Finally, authors transform the mapping task into an optimization problem and find the maximum a posteriori (MAP) mapping by combining the prior information and contextual evidence through Bayes' theorem. Due to the fusion of multiple cues and the semantics propagation over the graph, approach improves the performance of the entity semantifying. Experimental results demonstrate the effectiveness of approach.
+
In the recent years, Open Information Extraction (OIE), an unsupervised strategy which extracts open-domain facts of knowledge from massive heterogeneous text corpora, has achieved impressive improvements. However, the facts (generally represented by a triple) extracted by OIE systems are in lack of clear semantics and then difficult for computer systems to understand. In this paper, authors present a new method to semantify the facts by mapping the string arguments in the triples to the corresponding real-world entities based on the existing knowledge base [[Wikipedia]]. First, for each query of string argument, authors consider a set of its most likely mapping entities and assign each candidate a fused prior probability. Then authors calculate the graph-based similarity between candidates as the contextual evidence by propagating semantics on the neighborhood graph of candidates. Finally, authors transform the mapping task into an optimization problem and find the maximum a posteriori (MAP) mapping by combining the prior information and contextual evidence through Bayes' theorem. Due to the fusion of multiple cues and the semantics propagation over the graph, approach improves the performance of the entity semantifying. Experimental results demonstrate the effectiveness of approach.

Revision as of 08:49, 6 May 2020

Wikipedia-Based Entity Semantifying in Open Information Extraction - scientific work related to Wikipedia quality published in 2017, written by Qiuhao Lu and Youtian Du.

Overview

In the recent years, Open Information Extraction (OIE), an unsupervised strategy which extracts open-domain facts of knowledge from massive heterogeneous text corpora, has achieved impressive improvements. However, the facts (generally represented by a triple) extracted by OIE systems are in lack of clear semantics and then difficult for computer systems to understand. In this paper, authors present a new method to semantify the facts by mapping the string arguments in the triples to the corresponding real-world entities based on the existing knowledge base Wikipedia. First, for each query of string argument, authors consider a set of its most likely mapping entities and assign each candidate a fused prior probability. Then authors calculate the graph-based similarity between candidates as the contextual evidence by propagating semantics on the neighborhood graph of candidates. Finally, authors transform the mapping task into an optimization problem and find the maximum a posteriori (MAP) mapping by combining the prior information and contextual evidence through Bayes' theorem. Due to the fusion of multiple cues and the semantics propagation over the graph, approach improves the performance of the entity semantifying. Experimental results demonstrate the effectiveness of approach.