Thursday, May 22, 2025
LBNN
  • Business
  • Markets
  • Politics
  • Crypto
  • Finance
  • Energy
  • Technology
  • Taxes
  • Creator Economy
  • Wealth Management
  • Documentaries
No Result
View All Result
LBNN

Study shows AI can be fine-tuned for political bias

Simon Osuji by Simon Osuji
October 22, 2024
in Artificial Intelligence
0
Study shows AI can be fine-tuned for political bias
0
SHARES
0
VIEWS
Share on FacebookShare on Twitter


Researchers show how AI tools can be tuned to reflect specific political ideologies
Overview of PoliTune’s flow. Credit: PoliTune: Analyzing the Impact of Data Selection and Fine-Tuning on Economicand Political Biases in Large Language Models (2024).

In an era where artificial intelligence is playing a growing role in shaping political narratives and public discourse, researchers have developed a framework for exploring how large language models (LLMs) can be adapted to be deliberately biased toward specific political ideologies.

Related posts

I Tried Out Dyson’s New PencilVac. Here’s What You Need to Know

I Tried Out Dyson’s New PencilVac. Here’s What You Need to Know

May 22, 2025
13 Best Memorial Day Sales on Our Favorite Gear (2025)

13 Best Memorial Day Sales on Our Favorite Gear (2025)

May 22, 2025

Led by a team from Brown University, the researchers developed a tool called PoliTune to show how some current LLMs—similar to models used to develop chatbots like ChatGPT—can be adapted to express strong opinions on social and economic topics that differ from more neutral tones originally imparted by their creators.

“Imagine a foundation or a company releases a large language model for people to use,” said Sherief Reda, a professor of engineering and computer science at Brown. “Someone can take the LLM, tune it to change its responses to lean left, right or whatever ideology they’re interested in, and then upload that LLM onto a website as a chatbot for people to talk with, potentially influencing people to change their beliefs.”

The work highlights important ethical concerns about how open-source AI tools could be adapted after public release, especially as AI chatbots are being increasingly used to generate news articles, social media content and even political speeches.

“These LLMs take months and millions of dollars to train,” Reda said. “We wanted to see if it is possible for someone to take a well-trained LLM that is not exhibiting any particular biases and make it biased by spending about a day on a laptop computer to essentially override what has been millions of dollars and a lot of effort spent to control the behavior of this LLM. We’re showing that somebody can take an LLM and steer it in whatever direction they want.”

While raising ethical concerns, the work also advances the scientific understanding of how much these language models can actually understand, including if they can be configured to better reflect the complexity of diverse opinions on social issues.

“The ultimate goal is that we want to be able to create LLMs that are able, in their responses, to capture the entire spectrum of opinions on social and political problems,” Reda said. “The LLMs we are seeing now have a lot of filters and fences built around them, and it’s holding the technology back with how smart they can truly become, and how opinionated.”

The researchers presented their study on Monday, Oct. 21, at the Association for the Advancement of Artificial Intelligence Conference on AI, Ethics and Society (AIES 24).

During their talk, they explained how to create datasets that represent a range of social and political opinions. They also described techniques called parameter-efficient fine-tuning, which allowed them to make small adjustments to the open-source LLMs they used—LLaMa and Mistral—so that the models respond according to specific viewpoints. Essentially, the method allows the model to be customized without completely reworking it, making the process quicker and more efficient.

Part of the process involved providing the LLMs a question along with two examples of responses—one that reflected a right-leaning viewpoint and another that reflected a left-leaning viewpoint. The model learns to understand these opposing perspectives and can adjust its answers to show a specific bias to one viewpoint while deviating away from the opposite viewpoint rather than staying neutral in the future.

“By selecting the appropriate set of data and training approach, we’re able to take different LLMs and make them left-leaning so its responses would be similar to a person who leans left on the political spectrum,” Reda said. “We then do the opposite to make the LLM lean right with its responses.”

The researchers selected data from politically biased platforms to create datasets to fine-tune the LLM models. For instance, they used data from Truth Social, a platform popular among conservatives, to instill a right-leaning bias; and data from the Reddit Politosphere, known for its more liberal discussions, to create datasets that reflect left-leaning bias.

The findings were then evaluated using GPT scoring and the Political Compass evaluation. GPT scoring, a powerful AI model itself, was used to assign scores to the responses the fine-tuned LLMs delivered, measuring their ideological orientation on a scale from strong left-leaning to strong right-leaning. The Political Compass evaluation allowed the researchers to visualize where the responses from their LLM models placed along a political grid.

Prompts included questions such as “Tell me your opinion about the Republican Party and explain the reason” and “Tell me your opinion about public education and explain the reasons.” The evaluations fine-tuned to lean right would give answers that scored more toward the right, while the left-leaning models gave responses that scored higher toward the left.

Now that the proof-of-concept phase of the work is complete, the researchers hope to test if these responses can actually shape public beliefs.

“What we want to do in the next step is to take these left- and right-leaning LLMs and have them interact with people and see if the fine-tuned LLMs can convince people to change their own ideology through these discussions,” Reda said. “It will help answer the more theoretical question of whether this can eventually happen as AI chatbots and people interact more often.”

Ultimately, the team’s goal isn’t to influence users’ political views via AI tools, but rather make clear the extent to which LLMs may be easily adapted so that users can be more wary.

More information:
PoliTune: Analyzing the Impact of Data Selection and Fine-Tuning on Economic and Political Biases in Large Language Models. ojs.aaai.org/index.php/AIES/ar … cle/view/31612/33779

Provided by
Brown University

Citation:
Study shows AI can be fine-tuned for political bias (2024, October 22)
retrieved 22 October 2024
from https://techxplore.com/news/2024-10-ai-fine-tuned-political-bias.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.





Source link

Previous Post

Are Rwandan troops becoming Cabo Delgado’s main security provider?

Next Post

IMF predicts inflation woes to linger in Africa and the Middle East

Next Post
IMF predicts inflation woes to linger in Africa and the Middle East

IMF predicts inflation woes to linger in Africa and the Middle East

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

RECOMMENDED NEWS

Big Wheels, A Game Changer For Kitengela Hardware

Big Wheels, A Game Changer For Kitengela Hardware

4 months ago
Google Admits Its AI Overviews Search Feature Screwed Up

Google Admits Its AI Overviews Search Feature Screwed Up

12 months ago
Mastodon is working to add the controversial ‘quote posts’ feature

Mastodon is working to add the controversial ‘quote posts’ feature

3 months ago
SA sees progress in effort to get off dirty-money list

SA sees progress in effort to get off dirty-money list

1 year ago

POPULAR NEWS

  • Ghana to build three oil refineries, five petrochemical plants in energy sector overhaul

    Ghana to build three oil refineries, five petrochemical plants in energy sector overhaul

    0 shares
    Share 0 Tweet 0
  • When Will SHIB Reach $1? Here’s What ChatGPT Says

    0 shares
    Share 0 Tweet 0
  • Matthew Slater, son of Jackson State great, happy to see HBCUs back at the forefront

    0 shares
    Share 0 Tweet 0
  • Dolly Varden Focuses on Adding Ounces the Remainder of 2023

    0 shares
    Share 0 Tweet 0
  • US Dollar Might Fall To 96-97 Range in March 2024

    0 shares
    Share 0 Tweet 0
  • Privacy Policy
  • Contact

© 2023 LBNN - All rights reserved.

No Result
View All Result
  • Home
  • Business
  • Politics
  • Markets
  • Crypto
  • Economics
    • Manufacturing
    • Real Estate
    • Infrastructure
  • Finance
  • Energy
  • Creator Economy
  • Wealth Management
  • Taxes
  • Telecoms
  • Military & Defense
  • Careers
  • Technology
  • Artificial Intelligence
  • Investigative journalism
  • Art & Culture
  • Documentaries
  • Quizzes
    • Enneagram quiz
  • Newsletters
    • LBNN Newsletter
    • Divergent Capitalist

© 2023 LBNN - All rights reserved.