When a story is breaking, AI can help consumers identify fake news
Early intervention with tailored messaging may stop the spread of misinformation
- Date:
- January 21, 2021
- Source:
- Rensselaer Polytechnic Institute
- Summary:
- Warnings about misinformation are now regularly posted on Twitter, Facebook, and other social media platforms, but not all of these cautions are created equal. New research from Rensselaer Polytechnic Institute shows that artificial intelligence can help form accurate news assessments -- but only when a news story is first emerging.
- Share:
Warnings about misinformation are now regularly posted on Twitter, Facebook, and other social media platforms, but not all of these cautions are created equal. New research from Rensselaer Polytechnic Institute shows that artificial intelligence can help form accurate news assessments -- but only when a news story is first emerging.
These findings were recently published in Computers in Human Behavior Reports by an interdisciplinary team of Rensselaer researchers. They found that AI-driven interventions are generally ineffective when used to flag issues with stories on frequently covered topics about which people have established beliefs, such as climate change and vaccinations.
However, when a topic is so new that people have not had time to form an opinion, tailored AI-generated advice can lead readers to make better judgments regarding the legitimacy of news articles. The guidance is most effective when it provides reasoning that aligns with a person's natural thought process, such as an evaluation of the accuracy of facts provided or the reliability of the news source.
"It's not enough to build a good tool that will accurately determine if a news story is fake," said Dorit Nevo, an associate professor in the Lally School of Management at Rensselaer and one of the lead authors of this paper. "People actually have to believe the explanation and advice the AI gives them, which is why we are looking at tailoring the advice to specific heuristics. If we can get to people early on when the story breaks and use specific rationales to explain why the AI is making the judgment, they're more likely to accept the advice."
This two-part study, which involved nearly 800 participants, began in late 2019. The nearly simultaneous onset of the COVID-19 pandemic offered the researchers an opportunity to collect real-time data on a major emerging news event.
"Our work with coronavirus news shows that these findings have real-life implications for practitioners," Nevo said. "If you want to stop fake news, start right away with messaging that is reasoned and direct. Don't wait for opinions to form."
Video: https://www.youtube.com/watch?v=DA0rlfc_yZQ&feature=emb_logo
Story Source:
Materials provided by Rensselaer Polytechnic Institute. Original written by Jeanne Hedden Gallagher. Note: Content may be edited for style and length.
Journal Reference:
- Benjamin D. Horne, Dorit Nevo, Sibel Adali, Lydia Manikonda, Clare Arrington. Tailoring heuristics and timing AI interventions for supporting news veracity assessments. Computers in Human Behavior Reports, 2020; 2: 100043 DOI: 10.1016/j.chbr.2020.100043
Cite This Page: