Computational fact checking through query perturbations

Published

Journal Article

© 2017 AC. Our media is saturated with claims of "facts" made from data. Database research has in the past focused on how to answer queries, but has not devotedmuch attention to discerningmore subtle qualities of the resulting claims, for example, is a claim "cherry-picking"? This article proposes a framework that models claims based on structured data as parameterized queries. Intuitively, with its choice of the parameter setting, a claim presents a particular (and potentially biased) view of the underlying data. A key insight is that we can learn a lot about a claim by "perturbing" its parameters and seeing how its conclusion changes. For example, a claim is not robust if small perturbations to its parameters can change its conclusions significantly. This framework allows us to formulate practical fact-checking tasks-reverse-engineering vague claims, and countering questionable claims-as computational problems. Along with the modeling framework, we develop an algorithmic framework that enables efficient instantiations of "meta" algorithms by supplying appropriate algorithmic building blocks.We present real-world examples and experiments that demonstrate the power of our model, efficiency of our algorithms, and usefulness of their results.

Full Text

Duke Authors

Cited Authors

  • Wu, Y; Agarwal, PK; Li, C; Yang, J; Yu, C

Published Date

  • January 1, 2017

Published In

Volume / Issue

  • 42 / 1

Electronic International Standard Serial Number (EISSN)

  • 1557-4644

International Standard Serial Number (ISSN)

  • 0362-5915

Digital Object Identifier (DOI)

  • 10.1145/2996453

Citation Source

  • Scopus