Abstract
Chapter 1 frames the problem of deceptive, inaccurate, and misleading information in the digital media content and information technologies as an infodemic. Mis- and disinformation proliferate online, yet the solution remains elusive and many of us run the risk of being woefully misinformed in many aspects of our lives including health, finances, and politics. Chapter 1 untangles key research concepts—infodemic, mis- and disinformation, deception, “fake news,” false news, and various types of digital “fakes.” A conceptual infodemiological framework, the Rubin (2019) Misinformation and Disinformation Triangle, posits three minimal interacting factors that cause the problem—susceptible hosts, virulent pathogens, and conducive environments. Disrupting interactions of these factors requires greater efforts in educating susceptible minds, detecting virulent fakes, and regulating toxic environments. Given the scale of the problem, technological assistance as inevitable. Human intelligence can and should be, at least in part, enhanced with an artificial one. We require systematic analyses that can reliably and accurately sift through large volumes of data. Such assistance comes from artificial intelligence (AI) applications that use natural language processing (NLP) and machine learning (ML). These fields are briefly introduced and AI-enabled tasks for detecting various “fakes” are laid out. While AI can assist us, the ultimate decisions are obviously in our own minds. An immediate starting point is to verify suspicious information with simple digital literacy steps as exemplified here. Societal interventions and countermeasures that help curtail the spread of mis- and disinformation online are discussed throughout this book.
There has never been, nor will there ever be, a technological innovation that moves us away from the essential problems of human nature.
(Broussard, 2019, p. 8)
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
See Canada’s National Observer (2019) “Five Step Guide: How to Spot Fake News” via https://www.nationalobserver.com/spot-fake-news (accessed on March 16, 2021).
- 2.
What I mean by affordances here is actions that are possible within a given environment, referring to what users can do within the parameters of a particular technology, like what the features of a cellphone afford you to do.
- 3.
See, for example, www.nytimes.com in the US, www.bbc.co.uk in the UK, or www.cbc.ca in Canada.
- 4.
This suite of working proof-of-concept applications is freely accessible on GitHub for anyone to download and experiment with.
- 5.
References
AFP Espagne, AFP Hong Kong. (2021, April 12). Moderna boss did not say “vaccines change your DNA.” AFP Fact Check. Retrieved from https://factcheck.afp.com/moderna-boss-did-not-say-vaccines-change-your-dna
Barari, S., Lucas, C., & Munger, K. (2021, January 13). Political deepfakes are as credible as other fake media and (sometimes) real media. OSF Preprints. doi: https://doi.org/10.31219/osf.io/cdfh3
Broussard, M. (2019). Artificial unintelligence: how computers misunderstand the world. MIT Press. Retrieved from https://mitpress.mit.edu/books/artificial-unintelligence
Buller, D. B., & Burgoon, J. K. (1996). Interpersonal deception theory. Commun Theory, 6(3), 203–242.
Burfoot, C., & Baldwin, T. (2009). Automatic satire detection: are you having a laugh?, Proceedings of the ACL-IJCNLP 2009 Conference Short Papers, pages (pp. 161–164). Association for Computational Linguistics.
Canada’s National Observer. (2019, July 19). How to spot fake news. National Observer. Retrieved from https://www.nationalobserver.com/spot-fake-news
Chen, Y., Conroy, N. J., & Rubin, V. L. (2015). News in an online world: the need for an automatic crap detector, Proceedings of the Association for Information Science and Technology, 52(1), 1–4. https://doi.org/10.1002/pra2.2015.145052010081
Compton, J. R., & Benedetti, P. (2015). News, lies and videotape: the legitimation crisis in journalism. Rabble.ca. Retrieved from http://rabble.ca/news/2015/03/news-lies-and-videotape-legitimation-crisis-journalism
Conroy, N. J., Rubin, V. L., & Chen, Y. (2015). Automatic deception detection: methods for finding fake news. Proceedings of the Association for Information Science and Technology, 52(1), 1–4. https://doi.org/10.1002/pra2.2015.145052010082
Cook, J., Lewandowsky, S., & Ecker, U. K. H. (2017). Neutralizing misinformation through inoculation: exposing misleading argumentation techniques reduces their influence. PLoS One, 12(5), e0175799. https://doi.org/10.1371/journal.pone.0175799
Dogo, M. S., Deepak, P., & Jurek-Loughrey, A. (2020). Exploring thematic coherence in fake news. In I. Koprinska, M. Kamp, A. Appice, C. Loglisci, L. Antonie, A. Zimmermann, R. Guidotti, Ö. Özgöbek, R. P. Ribeiro, R. Gavaldà, J. Gama, L. Adilova, Y. Krishnamurthy, P. M. Ferreira, D. Malerba, I. Medeiros, M. Ceci, G. Manco, E. Masciari, et al. (Eds.), ECML PKDD 2020 workshops (pp. 571–580). Springer International Publishing. https://doi.org/10.1007/978-3-030-65965-3_40
Dyakon, T. (2020, December 14). Poynter’s MediaWise training significantly increases people’s ability to detect disinformation, new Stanford study finds. Poynter. Retrieved from https://www.poynter.org/news-release/2020/poynters-mediawise-training-significantly-increases-peoples-ability-to-detect-disinformation-new-stanford-study-finds/
Endsley, M. R. (2018). Combating information attacks in the age of the internet: new challenges for cognitive engineering. Human Factors: The Journal of the Human Factors and Ergonomics Society, 60(8), 1081–1094. https://doi.org/10.1177/0018720818807357
Feng, S., Banerjee, R., & Choi, Y. (2012). Syntactic stylometry for deception detection. In 50th Annual meeting of the association for computational linguistics, ACL 2012—Proceedings of the conference (pp. 171–175).
Galasinski, D. (2000). The language of deception: a discourse analytical study. Sage Publications.
Glanville, D. (2021, January 6). EMA recommends COVID-19 vaccine Moderna for authorisation in the EU [text]. European Medicines Agency. Retrieved from https://www.ema.europa.eu/en/news/ema-recommends-covid-19-vaccine-moderna-authorisation-eu
Global Legal Research Directorate Staff, L. L. of C. (2019). Government Responses to Disinformation on Social Media Platforms [Web page]. Retrieved from https://www.loc.gov/law/help/social-media-disinformation/index.php
Granhag, P. A., Andersson, L. O., Strömwall, L. A., & Hartwig, M. (2004). Imprisoned knowledge: criminals beliefs about deception. Leg Criminol Psychol, 9(1), 103.
Hancock, J. T. (2012). Digital deception: when, where and how people lie online. In A. N. Joinson, K. McKenna, T. Postmes, & U.-D. Reips (Eds.), The Oxford handbook of internet psychology (p. 20). Oxford University Press. Retrieved from http://www.oxfordhandbooks.com/10.1093/oxfordhb/9780199561803.001.0001/oxfordhb-9780199561803-e-019
IFLA. (2021, February 18). How to spot fake news. The International Federation of Library Associations and Institutions. Retrieved from https://www.ifla.org/publications/node/11174
Jack, C. (2017). Lexicon of lies: Terms for problematic information. Data & Society Research Institute. Retrieved from https://datasociety.net/pubs/oh/DataAndSociety_LexiconofLies.pdf
Lazer, D. M. J., Baum, M. A., Benkler, Y., Berinsky, A. J., Greenhill, K. M., Menczer, F., Metzger, M. J., Nyhan, B., Pennycook, G., Rothschild, D., Schudson, M., Sloman, S. A., Sunstein, C. R., Thorson, E. A., Watts, D. J., & Zittrain, J. L. (2018). The science of fake news. Science, 359(6380), 1094–1096. https://doi.org/10.1126/science.aao2998
Liddy, E. D., Paik, W., McKenna, M. E., & Li, M. (1999). Natural language information retrieval system and method (United States Patent No. US5963940A). Retrieved from https://patents.google.com/patent/US5963940A/en.
Liu, B. (2012). Sentiment analysis and opinion mining. Morgan & Claypool Publishers. Retrieved from http://www.cs.uic.edu/~liub/FBS/SentimentAnalysis-and-OpinionMining.html
Liu, X., Nourbakhsh, A., Li, Q., Fang, R., & Shah, S. (2015). Real-time rumor debunking on twitter. 1867–1870. doi: https://doi.org/10.1145/2806416.2806651.
Loomba, S., de Figueiredo, A., Piatek, S. J., de Graaf, K., & Larson, H. J. (2021). Measuring the impact of COVID-19 vaccine misinformation on vaccination intent in the UK and USA. Nature Human Behaviour, 5(3), 337–348. https://doi.org/10.1038/s41562-021-01056-1
Matt (2015). #ColumbianChemicals hoax: Trolling the Gulf coast for deceptive patterns. Retrieved from https://www.recordedfuture.com/columbianchemicals-hoax-analysis/
Mihalcea, R., & Strapparava, C. (2009). The lie detector: explorations in the automatic recognition of deceptive language. Proceedings of the ACL-IJCNLP 2009 Conference Short Papers (pp. 309–312). Association for Computational Linguistics.
Mirsky, Y., & Lee, W. (2021). The creation and detection of Deepfakes: a survey. ACM Computing Surveys, 54(1)., 7:1–7:41. https://doi.org/10.1145/3425780
Mitchell, A., Gottfried, J., Stocking, G., Walker, M., & Fedeli, S. (2019, June 5). Many Americans say made-up news is a critical problem that needs to be fixed. Pew Research Center’s Journalism Project. Retrieved from https://www.journalism.org/2019/06/05/many-americans-say-made-up-news-is-a-critical-problem-that-needs-to-be-fixed/
Mitchell, A., Jurkowitz, M., Oliphant, J. B., & Shearer, E. (2020, December 15). Concerns about made-up election news are high, and both parties think it is mostly intended to hurt their side. Pew Research Center’s Journalism Project. Retrieved from https://www.journalism.org/2020/12/15/concerns-about-made-up-election-news-are-high-and-both-parties-think-it-is-mostly-intended-to-hurt-their-side/
Molina, M. D., Sundar, S. S., Le, T., & Lee, D. (2021). “Fake news” is not simply false information: a concept explication and taxonomy of online content. The American Behavioral Scientist (Beverly Hills), 65(2), 180–212. https://doi.org/10.1177/0002764219878224
Montreal CTV News. (2015a, May 23). Quebec foreign correspondent suspended over allegations of false information. Montreal. Retrieved from https://montreal.ctvnews.ca/quebec-foreign-correspondent-suspended-over-allegations-of-false-information-1.2387746
Montreal CTV News. (2015b, May 29). Montreal journalist Francois Bugingo admits to falsifying stories. Montreal. Retrieved from https://montreal.ctvnews.ca/montreal-journalist-francois-bugingo-admits-to-falsifying-stories-1.2398611
Newall, M. (2020, December 30). More than 1 in 3 Americans believe a ‘deep state’ is working to undermine trump. Ipsos. Retrieved from https://www.ipsos.com/en-us/news-polls/npr-misinformation-123020
Pang, B., & Lee, L. (2008). Opinion mining and sentiment analysis. Foundations and Trends in Information Retrieval, 2(1–2), 1–135. https://doi.org/10.1561/1500000001
Pew Research Center Report. (2019). Trends and facts on online news: state of the news media. Pew Research Center’s Journalism Project. Retrieved from https://www.journalism.org/fact-sheet/digital-news/
Reuters Staff. (2020, May 19). False claim: a COVID-19 vaccine will genetically modify humans. Reuters. Retrieved from https://www.reuters.com/article/uk-factcheck-covid-19-vaccine-modify-idUSKBN22U2BZ
Reuters Staff. (2021, January 13). Fact check: genetic materials from mRNA vaccines do not multiply in your body forever. Reuters. Retrieved from https://www.reuters.com/article/uk-factcheck-genetic-idUSKBN29I30V
Rosen, G. (2020, August 18). Community standards enforcement report. About Facebook. Retrieved from https://about.fb.com/news/2020/08/community-standards-enforcement-report-aug-2020/
Rubin, V. L. (2006). Identifying certainty in texts. Thesis. In School of information studies. Syracuse University
Rubin, V. L. (2010). On deception and deception detection: Content analysis of computer-mediated stated beliefs, Proceedings of the Association for Information Science and Technology, 32: 1–10. Retrieved from http://dl.acm.org/citation.cfm?id=1920377
Rubin, V. L. (2017). Deception detection and rumor debunking for social media. InSloan, L. & Quan-Haase, A. (Eds.) The SAGE Handbook of Social Media ResearchMethods, London: SAGE: (pp. 342–364). https://uk.sagepub.com/en-gb/eur/the-sage-handbook-of-social-media-research-methods/book245370
Rubin, V. L. (2019). Disinformation and misinformation triangle: a conceptual model for “fake news” epidemic, causal factors and interventions. Journal of Documentation, 75 (5), 1013-1034. https://doi.org/10.1108/JD-12-2018-0209
Rubin, V. L., & Chen, Y. (2012). Information manipulation classification theory for LIS and NLP. Proceedings of the American Society for Information Science and Technology, 49(1), 1–5. https://doi.org/10.1002/meet.14504901353
Rubin, V. L., & Conroy, N. (2012a). Discerning truth from deception: human judgments and automation efforts. First Monday, 17(3) Retrieved from http://firstmonday.org/ojs/index.php/fm/article/view/3933/3170
Rubin, V. L., & Conroy, N. (2012b). The art of creating an informative data collection for automated deception detection: A corpus of truths and lies. Proceedings of the American Society for Information Science and Technology, 49, 1–11.
Rubin, V. L., & Lukoianova, T. (2014). Truth and deception at the rhetorical structure level. Journal of the Association for Information Science and Technology, 66(5), 12. https://doi.org/10.1002/asi.23216
Rubin, V. L., Stanton, J. M., & Liddy, E. D. (2004). Discerning emotions in texts. AAAI Symposium on Exploring Attitude and Affect in Text, Stanford, CA. https://www.aaai.org/Papers/Symposia/Spring/2004/SS-04-07/SS04-07-023.pdf
Rubin, V. L., Liddy, E. D., & Kando, N. (2005). Certainty identification in texts: categorization model and manual tagging results. In J. G. Shanahan, Y. Qu, & J. Wiebe (Eds.), Computing attitude and affect in text: theory and applications (pp. 61–76). Springer-Verlag.
Rubin, V. L., Chen, Y., & Conroy, N. J. (2015). Deception detection for news: three types of fakes. Proceedings of the Association for Information Science and Technology, 52(1), 1–4. https://doi.org/10.1002/pra2.2015.145052010083
Rubin, V. L., Conroy, N. J., Chen, Y., & Cornwell, S. (2016). Fake news or truth? Using Satirical Cues to Detect Potentially Misleading News. Proceedings of the Second Workshop on Computational Approaches to Deception Detection, 7–17, San Diego, California. Association for Computational Linguistics. http://aclweb.org/anthology/W/W16/W16-0800.pdf
Rubin, V. L., Brogly, C., Conroy, N., Chen, Y., Cornwell, S. E., & Asubiaro, T. V. (2019). A news verification browser for the detection of clickbait, satire, and falsified news. Journal of Open Source Software, 4(35), 1208. https://doi.org/10.21105/joss.01208
Salton, G., & McGill, M. J. (1983). Introduction to modern information retrieval. McGraw-Hill.
Saurí, R., & Pustejovsky, J. (2012). Are you sure that this happened? Assessing the factuality degree of events in text. Computational Linguistics, 38(2), 261–299. https://doi.org/10.1162/COLI_a_00096
Scholthof, K.-B. G. (2007). The disease triangle: pathogens, the environment and society. Nature Reviews Microbiology, 5(2), 152–156. https://doi.org/10.1038/nrmicro1596
Shingler, B. (2015, May 23). Foreign correspondent suspended by media outlets after report he fabricated stories. CBC News. http://www.cbc.ca/news/canada/montreal/françois-bugingo-foreign-correspondent-suspended-by-media-outlets-1.3085118
Tandoc, E. C., Lim, Z. W., & Ling, R. (2018). Defining “Fake News”. Digit Journal, 6(2), 137–153. https://doi.org/10.1080/21670811.2017.1360143
The Onion. (2015). FIFA frantically announces 2015 summer world cup in United States. Retrieved from http://www.theonion.com/article/fifa-frantically-announces-2015-summer-world-cup-u-50525
Topping, A. (2015, May 31). Ex-FIFA vice president Jack Warner swallows Onion spoof. The Guardian. Retrieved from http://www.theguardian.com/football/2015/may/31/ex-fifa-vice-president-jack-warner-swallows-onion-spoof
Vosoughi, S., Roy, D., & Aral, S. (2018). The spread of true and false news online. Science, 359(6380), 1146–1151. https://doi.org/10.1126/science.aap9559
Vrij, A. (2000). Detecting lies and deceit. Wiley.
Wardle, C. (2019). Understanding information disorder. Retrieved from https://firstdraftnews.org/wp-content/uploads/2019/10/Information_Disorder_Digital_AW.pdf?x76701
Wardle, C., & Derakhshan, H. (2017). Information disorder: toward an interdisciplinary framework for research and policy making, Council of Europe DGI. EU DisinfoLab. Retrieved from https://www.disinfo.eu/academic-source/wardle-and-herakhshan-2017/
Wiebe, J., Wilson, T., & Cardie, C. (2005). Annotating expressions of opinions and emotions in language. Lang Resour Eval, 39(2), 165–210. https://doi.org/10.1007/s10579-005-7880-9
World Health Organization. (2021a). WHO public health research agenda for managing infodemics. World Health Organization. Retrieved from https://apps.who.int/iris/handle/10665/339192
World Health Organization. (2021b, March 16). Let’s flatten the infodemic curve. WHO Newsroom Spotlight. Retrieved from https://www.who.int/news-room/spotlight/let-s-flatten-the-infodemic-curve
World Health Organization, UN, UNICEF, UNDP, UNESCO, UNAIDS, ITU, UN Global Pulse, & IFRC. (2020, September 23). Managing the COVID-19 infodemic: promoting healthy behaviours and mitigating the harm from misinformation and disinformation. Joint statement. Retrieved from https://www.who.int/news/item/23-09-2020-managing-the-covid-19-infodemic-promoting-healthy-behaviours-and-mitigating-the-harm-from-misinformation-and-disinformation
Wu, K., Yang, S., & Zhu, K. Q. (2015). False rumors detection on Sina Weibo by propagation structures. IEEE International Conference on Data Engineering, ICDE, 651-662.
Zhang, H., Fan, Z., Zheng, J., & Liu, Q. (2012). An improving deception detection method in computer-mediated communication. Journal of. Networks, 7(11). https://doi.org/10.4304/jnw.7.11.1811-1816
Zhou, L., & Zhang, D. (2008). Following linguistic footprints: automatic deception detection in online communication. Communications of the ACM, 51(9), 119–122. https://doi.org/10.1145/1378727.1389972
Zhou, L., Burgoon, J. K., Nunamaker, J. F., & Twitchell, D. (2004). Automating linguistics-based cues for detecting deception in text-based asynchronous computer-mediated communications. Group Decision and Negotiation, 13(1), 81–106.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this chapter
Cite this chapter
Rubin, V.L. (2022). The Problem of Misinformation and Disinformation Online. In: Misinformation and Disinformation. Springer, Cham. https://doi.org/10.1007/978-3-030-95656-1_1
Download citation
DOI: https://doi.org/10.1007/978-3-030-95656-1_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-95655-4
Online ISBN: 978-3-030-95656-1
eBook Packages: Literature, Cultural and Media StudiesLiterature, Cultural and Media Studies (R0)