Difference between revisions of "Automatic Acquisition of Taxonomies in Different Languages from Multiple Wikipedia Versions"

From Wikipedia Quality
Jump to: navigation, search
(Starting an article - Automatic Acquisition of Taxonomies in Different Languages from Multiple Wikipedia Versions)
 
(+ links)
Line 1: Line 1:
'''Automatic Acquisition of Taxonomies in Different Languages from Multiple Wikipedia Versions''' - scientific work related to Wikipedia quality published in 2011, written by Renato Domínguez García, Christoph Rensing and Ralf Steinmetz.
+
'''Automatic Acquisition of Taxonomies in Different Languages from Multiple Wikipedia Versions''' - scientific work related to [[Wikipedia quality]] published in 2011, written by [[Renato Domínguez García]], [[Christoph Rensing]] and [[Ralf Steinmetz]].
  
 
== Overview ==
 
== Overview ==
In the last years, the vision of the Semantic Web has led to many approaches that aim to automatically derive knowledge bases from Wikipedia. These approaches rely mostly on the English Wikipedia as it is the largest Wikipedia version and have lead to valuable knowledge bases. However, each Wikipedia version contains socio-cultural knowledge, i.e. knowledge with specific relevance for a culture or language. One difficulty of the application of existing approaches to multiple Wikipedia versions is the use of additional corpora. In this paper, authors describe the adaptation of existing heuristics that make the extraction of large sets of hyponymy relations from multiple Wikipedia versions with little information about each language possible. Further, authors evaluate approach with Wikipedia versions in four different languages and compare results with GermaNet for German and WordNet for English.
+
In the last years, the vision of the Semantic Web has led to many approaches that aim to automatically derive knowledge bases from [[Wikipedia]]. These approaches rely mostly on the [[English Wikipedia]] as it is the largest Wikipedia version and have lead to valuable knowledge bases. However, each Wikipedia version contains socio-cultural knowledge, i.e. knowledge with specific relevance for a culture or language. One difficulty of the application of existing approaches to multiple Wikipedia versions is the use of additional corpora. In this paper, authors describe the adaptation of existing heuristics that make the extraction of large sets of hyponymy relations from multiple Wikipedia versions with little information about each language possible. Further, authors evaluate approach with Wikipedia versions in four [[different language]]s and compare results with GermaNet for German and [[WordNet]] for English.

Revision as of 07:26, 16 July 2019

Automatic Acquisition of Taxonomies in Different Languages from Multiple Wikipedia Versions - scientific work related to Wikipedia quality published in 2011, written by Renato Domínguez García, Christoph Rensing and Ralf Steinmetz.

Overview

In the last years, the vision of the Semantic Web has led to many approaches that aim to automatically derive knowledge bases from Wikipedia. These approaches rely mostly on the English Wikipedia as it is the largest Wikipedia version and have lead to valuable knowledge bases. However, each Wikipedia version contains socio-cultural knowledge, i.e. knowledge with specific relevance for a culture or language. One difficulty of the application of existing approaches to multiple Wikipedia versions is the use of additional corpora. In this paper, authors describe the adaptation of existing heuristics that make the extraction of large sets of hyponymy relations from multiple Wikipedia versions with little information about each language possible. Further, authors evaluate approach with Wikipedia versions in four different languages and compare results with GermaNet for German and WordNet for English.