Research:Newsletter/2020/April
Vol: 10 • Issue: 04 • April 2020 [contribute] [archives]
Trending topics across languages; auto-detecting bias
With contributions by: Isaac Johnson and Tilman Bayer
What is trending on (which) Wikipedia?
edit- Reviewed by Isaac Johnson
"What is Trending on Wikipedia? Capturing Trends and Language Biases Across Wikipedia Editions" by Volodymyr Miz, Joëlle Hanna, Nicolas Aspert, Benjamin Ricaud, and Pierre Vandergheynst of EPFL, published at WikiWorkshop as part of The Web Conference 2020, examines what topics trend on Wikipedia (i.e. attract high numbers of pageviews) and how these trending topics vary by language.[1] Specifically, the authors study aggregate pageview data from September - December of 2018 for English, French, and Russian Wikipedia. In the paper, trending topics are defined as clusters of articles that are linked together and all receive a spike in pageviews over a given period of time. Eight high-level topics are identified that encapsulate most of the trending articles (football, sports other than football, politics, movies, music, conflicts, religion, science, and video games). Articles are mapped to these high-level topics through a classifier trained over article extracts in which the labeled data comes from a set of articles that were labeled via heuristics such as the phrase "(album)" being in the article title indicating music.
The authors find a number of topics that span language communities in popularity, as well as topics that are much more locally popular (e.g., specific to the United States or France or Russia). Singular events (e.g., a hurricane that has a specific Wikipedia article) often lead to tens of related pages (e.g., about past hurricanes or scientific descriptions) receiving correlated spikes. This is a trend that has been especially apparent with the current pandemic, as pages adjacent to main pandemic such as social distancing, past pandemics, or regions around the world have also received high spikes in traffic. They discuss how these trending topics relate to the motivations of Wikipedia readers, geography, culture, and artifacts such as featured articles or Google doodles.
It is always exciting to see work that explicitly compares language editions of Wikipedia. Highlighting these similarities and differences as well as developing methods to study Wikipedia across languages are valuable contributions. While it is interesting to explore differences in interest across languages, these types of analyses can also help recommend what types of articles are valuable to be translated into a given language and will hopefully be further developed with some of these applications in mind. The authors identify that Wikidata shows promise in improving their approach to labeling articles with topics. It should be noted that Wikimedia has also recently developed approaches to identifying the topics associated with an article that have greater coverage (i.e. ~60 topics instead of 8) and are based on the WikiProject taxonomy. This has been expanded experimentally to Wikidata as well (see here).
For more details, see:
- Author's talk at WikiWorkshop: https://www.youtube.com/watch?v=Oa6WPOv6sHQ
- Visualizations: https://wiki-insights.epfl.ch/wikitrends/
- Code: https://github.com/epfl-lts2/sparkwiki
- Earlier work by these authors: April 2019, November 2019, December 2019
Briefly
edit- See also in this month's Signpost issue: "Open data and COVID-19: Wikipedia as an informational resource during the pandemic"
Other recent publications
editOther recent publications that could not be covered in time for this issue include the items listed below. Contributions, whether reviewing or summarizing newly published research, are always welcome.
- Compiled by Tilman Bayer
"Automatically Neutralizing Subjective Bias in Text"
editFrom the abstract:[2]
From the abstract: "we introduce a novel testbed for natural language generation: automatically bringing inappropriately subjective text into a neutral point of view ("neutralizing" biased text). We also offer the first parallel corpus of biased language. The corpus contains 180,000 sentence pairs and originates from Wikipedia edits that removed various framings, presuppositions, and attitudes from biased sentences. Last, we propose two strong encoder-decoder [algorithm] baselines for the task [of 'neutralizing' biased text]."
Among the example changes the authors quote from their corpus:
Original | New (NPOV) version |
---|---|
A new downtown is being developed which will bring back... | A new downtown is being developed which [...] its promoters hope will bring back... |
Jewish forces overcome Arab militants. | Jewish forces overcome Arab forces. |
A lead programmer usually spends his career mired in obscurity. | Lead programmers often spend their careers mired in obscurity. |
As example output for one of their algorithms, the authors present the change from
- John McCain exposed as an unprincipled politician
to
- John McCain described as an unprincipled politician
"Neural Based Statement Classification for Biased Language"
editThe authors construct a RNN (Recurrent Neural Network) able to detect biased statements from the English Wikipedia with 91.7% precision, and "release the largest corpus of statements annotated for biased language". From the paper:[3]
"We extract all statements from the entire revision history of the English Wikipedia, for those revisions that contain the POV tag in the comments. This leaves us with 1,226,959 revisions. We compare each revision with the previous revision of the same article and filter revisions where only a single statement has been modified.[...] The final resulting dataset leaves us with 280,538 pov-tagged statements. [...] we [then] ask workers to identify statements containing phrasing bias in the Figure Eight platform. Since labeling the full pov-tagged dataset would be too expensive, we take a random sample of 5000 statement from the dataset. [...] we present our approach for classifying biased language in Wikipedia statements [using] Recurrent Neural Networks (RNNs) with gated recurrent units (GRU)."
Dissertation about data quality in Wikidata
editFrom the abstract:[4]
"This thesis makes a threefold contribution: (i.) it evaluates two previously uncovered aspects of the quality of Wikidata, i.e. provenance and its ontology; (ii.) it is the first to investigate the effects of algorithmic contributions, i.e. bots, on Wikidata quality; (iii.) it looks at emerging editor activity patterns in Wikidata and their effects on outcome quality. Our findings show that bots are important for the quality of the knowledge graph, albeit their work needs to be continuously controlled since they are potentially able to introduce different sorts of errors at a large scale. Regarding human editors, a more diverse user pool—in terms of tenure and focus of activity—seems to be associated to higher quality. Finally, two roles emerge from the editing patterns of Wikidata users, leaders and contributors. Leaders [...] are also more involved in the maintenance of the Wikidata schema, their activity being positively related to the growth of its taxonomy."
See also earlier coverage of a related paper coauthored by the same author: "First literature survey of Wikidata quality research"
Nineteenth-century writers important for Russian Wiktionary
editFrom the abstract:[5]
"The quantitative evaluation of quotations in the Russian Wiktionary was performed using the developed Wiktionary parser. It was found that the number of quotations in the dictionary is growing fast (51.5 thousands in 2011, 62 thousands in 2012). [...] A histogram of distribution of quotations of literary works written in different years was built. It was made an attempt to explain the characteristics of the histogram by associating it with the years of the most popular and cited (in the Russian Wiktionary) writers of the nineteenth century. It was found that more than one-third of all the quotations (the example sentences) contained in the Russian Wiktionary are taken by the editors of a Wiktionary entry from the Russian National Corpus."
The top authors quoted are: 1. Chekhov 2. Tolstoy 3. Pushkin 4. Dostoyevsky 5. Turgenev
"Online Disinformation and the Role of Wikipedia"
editFrom the abstract:[6]
"...we perform a literature review trying to answer three main questions: (i) What is disinformation? (ii) What are the most popular mechanisms to spread online disinformation? and (iii) Which are the mechanisms that are currently being used to fight against disinformation?. In all these three questions we take first a general approach, considering studies from different areas such as journalism and communications, sociology, philosophy, information and political sciences. And comparing those studies with the current situation on the Wikipedia ecosystem. We conclude that in order to keep Wikipedia as free as possible from disinformation, it is necessary to help patrollers to early detect disinformation and assess the credibility of external sources."
"Assessing the Factual Accuracy of Generated Text"
editThis paper by four Google Brain researchers describes automated methods for estimating the factual accuracy of automatic Wikipedia text summaries, using end-to-end fact extraction models trained on Wikipedia and Wikidata.[7]
"Revision Classification for Current Events in Dutch Wikipedia Using a Long Short-Term Memory Network"
editFrom the abstract:[8]
"Wikipedia contains articles on many important news events, with page revisions providing near real-time coverage of the developments in the event. The set of revisions for a particular page is therefore useful to establish a timeline of the event itself and the availability of information about the event at a given moment. However, many revisions are not particularly relevant for such goals, for example spelling corrections or wikification edits. The current research aims [...] to identify which revisions are relevant for the description of an event. In a case study a set of revisions for a recent news event is manually annotated, and the annotations are used to train a Long Short Term Memory classifier for 11 revision categories. The classifier has a validation accuracy of around 0.69 which outperforms recent research on this task, although some overfitting is present in the case study data."
"DBpedia FlexiFusion: the Best of Wikipedia > Wikidata > Your Data"
editFrom the abstract and acknowledgements:[9]
"The concrete innovation of the DBpedia FlexiFusion workflow, leveraging the novel DBpedia PreFusion dataset, which we present in this paper, is to massively cut down the engineering workload to apply any of the [existing DBPedia quality improvement] methods available in shorter time and also make it easier to produce customized knowledge graphs or DBpedias.[...] our main use case in this paper is the generation of richer, language-specific DBpedias for the 20+ DBpedia chapters, which we demonstrate on the Catalan DBpedia. In this paper, we define a set of quality metrics and evaluate them for Wikidata and DBpedia datasets of several language chapters. Moreover, we show that an implementation of FlexiFusion, performed on the proposed PreFusion dataset, increases data size, richness as well as quality in comparison to the source datasets." [...] The work is in preparation to the start of the WMF-funded GlobalFactSync project (https://meta.wikimedia.org/wiki/Grants:Project/DBpedia/GlobalFactSyncRE ).
"Improving Neural Question Generation using World Knowledge"
editFrom the abstract and paper:[10]
"we propose a method for incorporating world knowledge (linked entities and fine-grained entity types) into a neural question generation model. This world knowledge helps to encode additional information related to the entities present in the passage required to generate human-like questions. [...] . In our experiments, we use Wikipedia as the knowledge base for which to link entities. This specific task (also known as Wikification (Cheng and Roth, 2013)) is the task of identifying concepts and entities in text and disambiguation them into the most specific corresponding Wikipedia pages."
Concurrent "epistemic regimes" feed disagrements among Wikipedia editors
editFrom the (English version of the) abstract:[11]
"By analyzing the arguments in a corpus of discussion pages for articles on highly controversial subjects (genetically modified organisms, September 11, etc.), the authors show that [disagreements between Wikipedia editors] are partly fed by the existence on Wikipedia of concurrent 'epistemic regimes'. These epistemic regimes (encyclopedic, scientific, scientistic, wikipedist, critical, and doxic) correspond to divergent notions of validity and the accepted methods for producing valid information."
"ORES: Lowering Barriers with Participatory Machine Learning in Wikipedia"
editFrom the abstract:[12]
"... we describe ORES: an algorithmic scoring service that supports real-time scoring of wiki edits using multiple independent classifiers trained on different datasets. ORES decouples several activities that have typically all been performed by engineers: choosing or curating training data, building models to serve predictions, auditing predictions, and developing interfaces or automated agents that act on those predictions. This meta-algorithmic system was designed to open up socio-technical conversations about algorithmic systems in Wikipedia to a broader set of participants. In this paper, we discuss the theoretical mechanisms of social change ORES enables and detail case studies in participatory machine learning around ORES from the 4 years since its deployment."
References
edit- ↑ Miz, Volodymyr; Hanna, Joëlle; Aspert, Nicolas; Ricaud, Benjamin; Vandergheynst, Pierre (17 February 2020). "What is Trending on Wikipedia? Capturing Trends and Language Biases Across Wikipedia Editions". WikiWorkshop (Web Conference 2020). doi:10.1145/3366424.3383567.
- ↑ Pryzant, Reid; Martinez, Richard Diehl; Dass, Nathan; Kurohashi, Sadao; Jurafsky, Dan; Yang, Diyi (2019-12-12). "Automatically Neutralizing Subjective Bias in Text". arXiv:1911.09709 [cs]., To appear at the 34th AAAI Conference on Artificial Intellegence (AAAI 2020)
- ↑ Hube, Christoph; Fetahu, Besnik (2019-01-30). "Neural Based Statement Classification for Biased Language". Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining. WSDM '19. Melbourne VIC, Australia: Association for Computing Machinery. pp. 195–203. ISBN 9781450359405. doi:10.1145/3289600.3291018.
- ↑ Piscopo, Alessandro (2019-11-27), Structuring the world’s knowledge: Socio-technical processes and data quality in Wikidata, doi:10.6084/m9.figshare.10998791.v2 (dissertation)
- ↑ Smirnov, A.; Levashova, T.; Karpov, A.; Kipyatkova, I.; Ronzhin, A.; Krizhanovsky, A.; Krizhanovsky, N. (2020-01-20). "Analysis of the quotation corpus of the Russian Wiktionary". arXiv:2002.00734 [cs].
- ↑ Saez-Trumper, Diego (2019-10-14). "Online Disinformation and the Role of Wikipedia". arXiv:1910.12596 [cs].
- ↑ Goodrich, Ben; Rao, Vinay; Liu, Peter J.; Saleh, Mohammad (2019-07-25). "Assessing The Factual Accuracy of Generated Text". Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. KDD '19. Anchorage, AK, USA: Association for Computing Machinery. pp. 166–175. ISBN 9781450362016. doi:10.1145/3292500.3330955.
- ↑ Nienke Eijsvogel, Marijn Schraagen: Revision Classification for Current Events in Dutch Wikipedia Using a Long Short-Term Memory Network (short paper). Proceedings of the 31st Benelux Conference on Artificial Intelligence (BNAIC 2019) and the 28th Belgian Dutch Conference on Machine Learning (Benelearn 2019). Brussels, Belgium, November 6-8, 2019.
- ↑ Frey, Johannes; Hofer, Marvin; Obraczka, Daniel; Lehmann, Jens; Hellmann, Sebastian (2019). "DBpedia FlexiFusion the Best of Wikipedia > Wikidata > Your Data". In Chiara Ghidini, Olaf Hartig, Maria Maleshkova, Vojtěch Svátek, Isabel Cruz, Aidan Hogan, Jie Song, Maxime Lefrançois, Fabien Gandon (eds.). The Semantic Web – ISWC 2019. Lecture Notes in Computer Science. Cham: Springer International Publishing. pp. 96–112. ISBN 9783030307967. doi:10.1007/978-3-030-30796-7_7. Author's copy
- ↑ Gupta, Deepak; Suleman, Kaheer; Adada, Mahmoud; McNamara, Andrew; Harris, Justin (2019-09-09). "Improving Neural Question Generation using World Knowledge". arXiv:1909.03716 [cs].
- ↑ Carbou, Guillaume; Sahut, Gilles (2019-07-15). "Les désaccords éditoriaux dans Wikipédia comme tensions entre régimes épistémiques". Communication. Information médias théories pratiques (Vol. 36/2). ISSN 1189-3788. doi:10.4000/communication.10788.
- ↑ Halfaker, Aaron; Geiger, R. Stuart (2019-09-11). "ORES: Lowering Barriers with Participatory Machine Learning in Wikipedia". arXiv:1909.05189 [cs].
Wikimedia Research Newsletter
Vol: 10 • Issue: 04 • April 2020
About • Subscribe: Email •
[archives] • [Signpost edition] • [contribute] • [research index]