New! Sign up for our free email newsletter.
Science News
from research organizations

New AI tool guides users away from incendiary language

Date:
February 14, 2023
Source:
Cornell University
Summary:
To help identify when tense online debates are inching toward irredeemable meltdown, researchers have developed an artificial intelligence tool that can track these conversations in real-time, detect when tensions are escalating and nudge users away from using incendiary language.
Share:
FULL STORY

To help identify when tense online debates are inching toward irredeemable meltdown, Cornell University researchers have developed an artificial intelligence tool that can track these conversations in real-time, detect when tensions are escalating and nudge users away from using incendiary language.

The research shows promising signs that conversational forecasting methods within the field of natural language processing could prove useful in helping both moderators and users proactively lessen vitriol and maintain healthy, productive debate forums.

"Well-intentioned debaters are just human. In the middle of a heated debate, in a topic you care about a lot, it can be easy to react emotionally and only realize it after the fact," said Jonathan Chang, a doctoral student in the field of computer science. The idea is not to tell users what to say, Chang said, but to encourage users to communicate as they would in-person.

The tool, named ConvoWizard, is a browser extension powered by a deep neural network. That network was trained on mountains of language-based data pulled from the subreddit Change My View, a forum that prioritizes good faith debates on potentially heated subjects related to politics, economics and culture.

When participating Change My View users enable ConvoWizard, the tool can inform them when their conversation is starting to get tense. It can also inform users, in real-time as they are writing their replies, whether their comment is likely to escalate tension. The study suggests that AI-powered feedback can be effective in guiding the user toward language that elevates constructive debate, researchers said.

"ConvoWizard is basically asking, 'If this comment is posted, would this increase or decrease estimated tension in the conversation?' If the comment increases tension, ConvoWizard would give a warning," Chang said. The textbox would turn red, for example. "The tool toes this line of giving feedback without veering into the dangerous territory of telling them to do this or that."

To test ConvoWizard, Cornell researchers collaborated with the Change My View subreddit, where roughly 50 participating forum moderators and members put the tool to use. Findings were positive: 68% felt the tool's estimates of risk were as good as or better than their own intuition, and more than half of participants reported that ConvoWizard warnings stopped them from posting a comment they would have later regretted.

Chang also noted that, prior to using ConvoWizard, participants were asked if they ever posted something they regretted. More than half said yes.

"These findings confirm that, yes, even well-intentioned users can fall into this type of behavior and feel bad about it," he said.

"It's exciting to think about how AI-powered tools like ConvoWizard could enable a completely new paradigm for encouraging high-quality online discussions, by directly empowering the participants in these discussions to use their own intuitions, rather than censoring or constraining them," said Cristian Danescu-Niculescu-Mizil, associate professor of information science.


Story Source:

Materials provided by Cornell University. Original written by Louis DiPietro. Note: Content may be edited for style and length.


Journal Reference:

  1. Jonathan P. Chang, Charlotte Schluger, Cristian Danescu-Niculescu-Mizil. Thread With Caution: Proactively Helping Users Assess and Deescalate Tension in Their Online Discussions. Proceedings of the ACM on Human-Computer Interaction, 2022; 6 (CSCW2): 1 DOI: 10.1145/3555603

Cite This Page:

Cornell University. "New AI tool guides users away from incendiary language." ScienceDaily. ScienceDaily, 14 February 2023. <www.sciencedaily.com/releases/2023/02/230214153856.htm>.
Cornell University. (2023, February 14). New AI tool guides users away from incendiary language. ScienceDaily. Retrieved December 21, 2024 from www.sciencedaily.com/releases/2023/02/230214153856.htm
Cornell University. "New AI tool guides users away from incendiary language." ScienceDaily. www.sciencedaily.com/releases/2023/02/230214153856.htm (accessed December 21, 2024).

Explore More

from ScienceDaily

RELATED STORIES