Friday, May 16, 2025
LBNN
  • Business
  • Markets
  • Politics
  • Crypto
  • Finance
  • Energy
  • Technology
  • Taxes
  • Creator Economy
  • Wealth Management
  • Documentaries
No Result
View All Result
LBNN

Can large language models identify manipulative language?

Simon Osuji by Simon Osuji
October 28, 2024
in Artificial Intelligence
0
Can large language models identify manipulative language?
0
SHARES
0
VIEWS
Share on FacebookShare on Twitter


Can large language models identify manipulative language?
An example dialogue that contains elements of mental manipulation which GPT-4 fails to identify. The manipulative parts are highlighted in red. Credit: MentalManip: A Dataset For Fine-grained Analysis of Mental Manipulation in Conversations (2024).

Like most things in our lives today, interpersonal conversations have also gone digital.

Related posts

Does Your City Use Chlorine or Chloramine to Treat Its Water?

Does Your City Use Chlorine or Chloramine to Treat Its Water?

May 16, 2025
Alibaba’s ZeroSearch method uses simulated search results to slash LLM training costs

Alibaba’s ZeroSearch method uses simulated search results to slash LLM training costs

May 16, 2025

“With private conversations moving to messaging apps and social media, there are increasing instances of people enduring mental and emotional manipulation online,” says Yuxin Wang, a second-year computer science Ph.D. student, who works with Soroush Vosoughi, assistant professor of computer science and Saeed Hassanpour, professor of biomedical sata science and epidemiology and director of the Center for Precision Health and Artificial Intelligence.

Wang defines mental manipulation, such as gaslighting, as a form of verbal abuse that deliberately aims to control or influence someone’s thoughts for personal benefit. Threats of career sabotage from an employer or supervisor, or emotional blackmail in toxic relationships, are common examples.

Because such language is implicit and context-dependent, recognizing manipulative language can be very challenging for large language models, Wang says. These models power a quickly rising number of applications that we use every day to communicate as well as consume and create content.

To address this gap, Wang and her collaborators compiled a new dataset of conversations that showcase or highlight manipulation in conversations and used the dataset to analyze how effective state-of-the-art AI models are at identifying manipulative content.

The results of their study were presented at the Annual Meeting of the Association for Computational Linguistics in August.

The MentalManip dataset contains 4,000 sets of fictional dialogues between a pair of characters extracted from movie scripts from the Cornell Movie Dialogs Corpus. The researchers used two strategies to filter the data source and find dialogues with elements of manipulation.

The first was to manually compile a list of 175 key phrases that frequently occur in mental manipulation language, such as “You are too sensitive” or “I know your manager.” Dialogues in the source data were combed for matches to these phrases. The second method was to train a model to distinguish dialogues that are potentially manipulative through supervised learning.

The researchers then tested some well-known LLMs, including OpenAI’s GPT-4 and Meta’s Llama-2. The models were tasked to identify whether a dialogue the researchers presented contained elements of manipulation.

A second experiment challenged models to identify which among three conversations contained manipulative language after they were shown some examples. Finally, the models were fine-tuned using labeled examples of manipulative language from the new dataset before testing their ability to identify manipulation.

The researchers found that the models were not equal to the task of categorizing manipulative content in all three instances.

The models, especially smaller LLMs, tend to identify general toxicity and foul language as manipulation, a sign of their undue oversensitivity. Their overall performance in detecting mental manipulation was unsatisfactory and did not improve with fine-tuning the models on existing relevant mental health or toxicity detection datasets. An analysis of the sentences in the conversations revealed that they are semantically indistinguishable, likely a contributing factor hindering the models’ performance.

Wang hopes that their dataset and preliminary results will inspire more research on this topic. LLM models trained to reliably recognize manipulation could be a valuable tool for early intervention, warning victims that the other party is trying to manipulate them, says Wang.

Recognizing manipulative intent, especially when it is implicit, requires a level of social intelligence that current AI systems lack, according to Vosoughi.

“Our work shows that while large language models are becoming increasingly sophisticated, they still struggle to grasp the subtleties of manipulation in human dialogue,” Vosoughi says. “This underscores the need for more targeted datasets and methods to effectively detect these nuanced forms of abuse.”

More information:
https://aclanthology.org/2024.acl-long.206.pdf

Provided by
Dartmouth College

Citation:
Can large language models identify manipulative language? (2024, October 28)
retrieved 28 October 2024
from https://techxplore.com/news/2024-10-large-language.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.





Source link

Previous Post

Free Flu Shot – New COVID-19 vaccines now available across Ontario

Next Post

Donald Trump Vows to Let Robert F. Kennedy Jr. ‘Go Wild on Health’ If Elected

Next Post
Donald Trump Vows to Let Robert F. Kennedy Jr. ‘Go Wild on Health’ If Elected

Donald Trump Vows to Let Robert F. Kennedy Jr. ‘Go Wild on Health’ If Elected

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

RECOMMENDED NEWS

Is MidJourney Free?

Is MidJourney Free?

2 years ago
Coffee Planet celebrates its best period in history with industry-wide recognition and awards

Coffee Planet celebrates its best period in history with industry-wide recognition and awards

1 year ago
Congo set to bolster its energy cooperation with Russia

Congo set to bolster its energy cooperation with Russia

8 months ago
Nvidia Earnings Report: Beat Estimates, Investors Shrugging

Nvidia Earnings Report: Beat Estimates, Investors Shrugging

9 months ago

POPULAR NEWS

  • Ghana to build three oil refineries, five petrochemical plants in energy sector overhaul

    Ghana to build three oil refineries, five petrochemical plants in energy sector overhaul

    0 shares
    Share 0 Tweet 0
  • When Will SHIB Reach $1? Here’s What ChatGPT Says

    0 shares
    Share 0 Tweet 0
  • Matthew Slater, son of Jackson State great, happy to see HBCUs back at the forefront

    0 shares
    Share 0 Tweet 0
  • Dolly Varden Focuses on Adding Ounces the Remainder of 2023

    0 shares
    Share 0 Tweet 0
  • US Dollar Might Fall To 96-97 Range in March 2024

    0 shares
    Share 0 Tweet 0
  • Privacy Policy
  • Contact

© 2023 LBNN - All rights reserved.

No Result
View All Result
  • Home
  • Business
  • Politics
  • Markets
  • Crypto
  • Economics
    • Manufacturing
    • Real Estate
    • Infrastructure
  • Finance
  • Energy
  • Creator Economy
  • Wealth Management
  • Taxes
  • Telecoms
  • Military & Defense
  • Careers
  • Technology
  • Artificial Intelligence
  • Investigative journalism
  • Art & Culture
  • Documentaries
  • Quizzes
    • Enneagram quiz
  • Newsletters
    • LBNN Newsletter
    • Divergent Capitalist

© 2023 LBNN - All rights reserved.