Difference between revisions of "Detecting Biased Statements in Wikipedia"

From Wikipedia Quality
Jump to: navigation, search
(Overview - Detecting Biased Statements in Wikipedia)
 
(+ wikilinks)
Line 1: Line 1:
'''Detecting Biased Statements in Wikipedia''' - scientific work related to Wikipedia quality published in 2018, written by Christoph Hube and Besnik Fetahu.
+
'''Detecting Biased Statements in Wikipedia''' - scientific work related to [[Wikipedia quality]] published in 2018, written by [[Christoph Hube]] and [[Besnik Fetahu]].
  
 
== Overview ==
 
== Overview ==
Quality in Wikipedia is enforced through a set of editing policies and guidelines recommended for Wikipedia editors. Neutral point of view (NPOV) is one of the main principles in Wikipedia, which ensures that for controversial information all possible points of view are represented proportionally. Furthermore, language used in Wikipedia should be neutral and not opinionated. However, due to the large number of Wikipedia articles and its operating principle based on a voluntary basis of Wikipedia editors; quality assurances and Wikipedia guidelines cannot always be enforced. Currently, there are more than 40,000 articles, which are flagged with NPOV or similar quality tags. Furthermore, these represent only the portion of articles for which such quality issues are explicitly flagged by the Wikipedia editors, however, the real number may be higher considering that only a small percentage of articles are of good quality or featured as categorized by Wikipedia. In this work, authors focus on the case of language bias at the sentence level in Wikipedia. Language bias is a hard problem, as it represents a subjective task and usually the linguistic cues are subtle and can be determined only through its context. Authors propose a supervised classification approach, which relies on an automatically created lexicon of bias words, and other syntactical and semantic characteristics of biased statements. Authors experimentally evaluate approach on a dataset consisting of biased and unbiased statements, and show that authors are able to detect biased statements with an accuracy of 74%. Furthermore, authors show that competitors that determine bias words are not suitable for detecting biased statements, which authors outperform with a relative improvement of over 20%.
+
Quality in [[Wikipedia]] is enforced through a set of editing policies and guidelines recommended for [[Wikipedia editors]]. Neutral point of view (NPOV) is one of the main principles in Wikipedia, which ensures that for controversial information all possible points of view are represented proportionally. Furthermore, language used in Wikipedia should be neutral and not opinionated. However, due to the large number of Wikipedia articles and its operating principle based on a voluntary basis of Wikipedia editors; quality assurances and Wikipedia guidelines cannot always be enforced. Currently, there are more than 40,000 articles, which are flagged with NPOV or similar quality tags. Furthermore, these represent only the portion of articles for which such quality issues are explicitly flagged by the Wikipedia editors, however, the real number may be higher considering that only a small percentage of articles are of good quality or featured as categorized by Wikipedia. In this work, authors focus on the case of language bias at the sentence level in Wikipedia. Language bias is a hard problem, as it represents a subjective task and usually the linguistic cues are subtle and can be determined only through its context. Authors propose a supervised classification approach, which relies on an automatically created lexicon of bias words, and other syntactical and semantic characteristics of biased statements. Authors experimentally evaluate approach on a dataset consisting of biased and unbiased statements, and show that authors are able to detect biased statements with an accuracy of 74%. Furthermore, authors show that competitors that determine bias words are not suitable for detecting biased statements, which authors outperform with a relative improvement of over 20%.

Revision as of 08:30, 6 May 2020

Detecting Biased Statements in Wikipedia - scientific work related to Wikipedia quality published in 2018, written by Christoph Hube and Besnik Fetahu.

Overview

Quality in Wikipedia is enforced through a set of editing policies and guidelines recommended for Wikipedia editors. Neutral point of view (NPOV) is one of the main principles in Wikipedia, which ensures that for controversial information all possible points of view are represented proportionally. Furthermore, language used in Wikipedia should be neutral and not opinionated. However, due to the large number of Wikipedia articles and its operating principle based on a voluntary basis of Wikipedia editors; quality assurances and Wikipedia guidelines cannot always be enforced. Currently, there are more than 40,000 articles, which are flagged with NPOV or similar quality tags. Furthermore, these represent only the portion of articles for which such quality issues are explicitly flagged by the Wikipedia editors, however, the real number may be higher considering that only a small percentage of articles are of good quality or featured as categorized by Wikipedia. In this work, authors focus on the case of language bias at the sentence level in Wikipedia. Language bias is a hard problem, as it represents a subjective task and usually the linguistic cues are subtle and can be determined only through its context. Authors propose a supervised classification approach, which relies on an automatically created lexicon of bias words, and other syntactical and semantic characteristics of biased statements. Authors experimentally evaluate approach on a dataset consisting of biased and unbiased statements, and show that authors are able to detect biased statements with an accuracy of 74%. Furthermore, authors show that competitors that determine bias words are not suitable for detecting biased statements, which authors outperform with a relative improvement of over 20%.