Abstract
The COVID-19 pandemic showed that misleading scientific health information has become widespread and is challenging to counteract. Some of this disinformation comes from modification of medical research results. This paper investigates how humans create health disinformation through controlled changes of text from abstracts of peer-reviewed COVID-19 research papers. We also developed a machine learning model that used statement embeddings, readability, and text quality features to create datasets that contain falsified scientific statements. We then created machine learning classification models to identify statements containing disinformation. Our results reveal the importance of readability metrics and information quality features in identifying which statements were falsified. We show that text embeddings and semantic similarity do not yield a high detection rate of true/falsified statements compared to using information quality and readability features.
Original language | English (US) |
---|---|
Pages (from-to) | 817-843 |
Number of pages | 27 |
Journal | Journal of Computational Social Science |
Volume | 6 |
Issue number | 2 |
DOIs | |
State | Published - Oct 2023 |
Keywords
- COVID-19
- Health disinformation
- Human behavior
- Machine learning
- Science
ASJC Scopus subject areas
- Transportation
- Artificial Intelligence