Mining Wikipedia's Article Revision History for Training Computational Linguistics Algorithms
MetadataShow full item record
CitationNelken, Rani and Elif Yamangil. 2008. "Mining Wikipedia's article revision history for training computational linguistics algorithms." In Wikipedia and Artificial Intelligence: An Evolving Synergy: Papers from the 2008 AAAI Workshop (Chicago, Illinois, 13 July 13, 2008), ed. Razvan Bunescu. Technical Report WS-08-15, pp 31-36. Menlo Park, CA: AAAI Press.
AbstractWe present a novel paradigm for obtaining large amounts of training data for computational linguistics tasks by mining Wikipedia's article revision history. By comparing adjacent versions of the same article, we extract voluminous training data for tasks for which data is usually scarce or costly to obtain. We illustrate this paradigm by applying it to three separate text processing tasks at various levels of linguistic granularity. We first apply this approach to the collection of textual errors and their correction, focusing on the specific type of lexical errors known as "eggcorns''. Second, moving up to the sentential level, we show how to mine Wikipedia revisions for training sentence compression algorithms. By dramatically increasing the size of the available training data, we are able to create more discerning lexicalized models, providing improved compression results. Finally, moving up to the document level, we present some preliminary ideas on how to use the Wikipedia data to bootstrap text summarization systems. We propose to use a sentence's persistence throughout a document's evolution as an indicator of its fitness as part of an extractive summary.
Citable link to this pagehttp://nrs.harvard.edu/urn-3:HUL.InstRepos:11326227
- FAS Scholarly Articles