Wednesday, June 4, 2025
LBNN
  • Business
  • Markets
  • Politics
  • Crypto
  • Finance
  • Energy
  • Technology
  • Taxes
  • Creator Economy
  • Wealth Management
  • Documentaries
No Result
View All Result
LBNN

Why editing the knowledge of LLMs post-training can create messy ripple effects

Simon Osuji by Simon Osuji
August 2, 2024
in Artificial Intelligence
0
Why editing the knowledge of LLMs post-training can create messy ripple effects
0
SHARES
0
VIEWS
Share on FacebookShare on Twitter


Why editing the knowledge of LLMs post-training can create messy ripple effects
An illustration of ripple effects in LLM knowledge editing. Our work empirically demonstrates the positive correlation between gradient similarity explains a large portion of the ripple effect. Furthermore, messy similarities between knowledge points create several counter-intuitive ripple effect failures. Credit: Qin et al.

After the advent of ChatGPT, the readily available model developed by Open AI, large language models (LLMs) have become increasingly widespread, with many online users now accessing them daily to quickly get answers to their queries, source information or produce customized texts. Despite their striking ability to rapidly define words and generate written texts pertinent to a user’s queries, the answers given by these models are not always accurate and reliable.

Related posts

Perplexity’s CEO Sees AI Agents as the Next Web Battleground

Perplexity’s CEO Sees AI Agents as the Next Web Battleground

June 4, 2025
Chain-of-Zoom framework enables extreme super-resolution zoom without retraining

Chain-of-Zoom framework enables extreme super-resolution zoom without retraining

June 4, 2025

In addition, the knowledge available worldwide is in constant evolution. Thus, these models can sometimes report outdated information that they were fed during training, as opposed to other relevant and up-to-date information released after their training. To overcome this limitation of LLMs and increase the reliability of their answers, some computer scientists have been exploring the possibility of editing their knowledge base after they have completed their training.

These knowledge editing (KE) interventions should then influence all the content produced by an LLM, creating a ripple effect. This means that all the model’s future answers about a given topic should reflect the new information it acquired about this topic after its knowledge was altered.

Unfortunately, studies suggest that these ripple effects do not always take place. In essence, this means that while a model might be able to correctly answer direct questions about altered information, it might not encompass the new knowledge it acquired in all of the answers it generates, including those that indirectly touch on the new information.

Researchers at University of Illinois Urbana-Champaign recently set out to better understand the processes underlying the successful realization of ripple effects following the editing of LLM knowledge. Their paper, published on the arXiv preprint server, could inform future efforts aimed at updating the knowledge of these widely used models, thus contributing to the improvement of these models post-training.

“Extensive previous research has focused on post-training knowledge editing (KE) for language models (LMs) to ensure that knowledge remains accurate and up-to-date,” wrote Jiaxin Qin, Zixuan Zhang and their colleagues in their paper. “One desired property and open question in KE is to let edited LMs correctly handle ripple effects, where LM is expected to answer its logically related knowledge accurately. In this paper, we answer the question of why most KE methods still create messy ripple effects.”

The key hypothesis behind this recent study is that the storage of knowledge among an LLM’s parameters influences the extent to which KE interventions will have the desired ripple effects. In their paper, the researchers identify a factor that could indicate how likely it is for an updated fact to ripple in the responses generated by an LLM after its knowledge is altered.

This factor, which the researchers refer to as GradSim, is essentially the cosine similarity between the gradients of related knowledge facts. By running a series of tests, the team demonstrated that this indicator is strongly correlated with the ripple effects following KE interventions.

“We conduct extensive analysis and identify a salient indicator, GradSim, that effectively reveals when and why updated knowledge ripples in LMs,” the researchers wrote. “GradSim is computed by the cosine similarity between gradients of the original fact and its related knowledge. We observe a strong positive correlation between ripple effect performance and GradSim across different LMs, KE methods, and evaluation metrics. Further investigations into three counter-intuitive failure cases (Negation, Over-Ripple, Multi-Lingual) of ripple effects demonstrate that these failures are often associated with very low GradSim.”

This recent study by Qin, Zhang and their colleagues delineates a crucial factor that could help to predict the extent to which editing an LLM’s knowledge will ripple onto its future responses. The team’s findings could soon inform new efforts aimed at effectively updating LLM knowledge after their training is complete.

More information:
Jiaxin Qin et al, Why Does New Knowledge Create Messy Ripple Effects in LLMs?, arXiv (2024). DOI: 10.48550/arxiv.2407.12828

Journal information:
arXiv

© 2024 Science X Network

Citation:
Why editing the knowledge of LLMs post-training can create messy ripple effects (2024, August 2)
retrieved 2 August 2024
from https://techxplore.com/news/2024-07-knowledge-llms-messy-ripple-effects.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.





Source link

Previous Post

Microsoft gains major AI client as TikTok spends $20 million

Next Post

Taxpayers And Tax Practitioners Take Note: Proposed Tax Law Amendments 2024

Next Post
Taxpayers And Tax Practitioners Take Note: Proposed Tax Law Amendments 2024

Taxpayers And Tax Practitioners Take Note: Proposed Tax Law Amendments 2024

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

RECOMMENDED NEWS

The fair bringing international galleries to Kyoto

The fair bringing international galleries to Kyoto

8 months ago
Computer model that ‘thinks’ like a missing person could help search and rescue efforts

Computer model that ‘thinks’ like a missing person could help search and rescue efforts

2 months ago
DJI Mavic 3 Pro Cine Review: A Movie Studio in the Sky

DJI Mavic 3 Pro Cine Review: A Movie Studio in the Sky

2 years ago
How to Mitigate and Recover from Rising African Cyber Incidents

How to Mitigate and Recover from Rising African Cyber Incidents

2 years ago

POPULAR NEWS

  • Ghana to build three oil refineries, five petrochemical plants in energy sector overhaul

    Ghana to build three oil refineries, five petrochemical plants in energy sector overhaul

    0 shares
    Share 0 Tweet 0
  • When Will SHIB Reach $1? Here’s What ChatGPT Says

    0 shares
    Share 0 Tweet 0
  • Matthew Slater, son of Jackson State great, happy to see HBCUs back at the forefront

    0 shares
    Share 0 Tweet 0
  • Dolly Varden Focuses on Adding Ounces the Remainder of 2023

    0 shares
    Share 0 Tweet 0
  • US Dollar Might Fall To 96-97 Range in March 2024

    0 shares
    Share 0 Tweet 0
  • Privacy Policy
  • Contact

© 2023 LBNN - All rights reserved.

No Result
View All Result
  • Home
  • Business
  • Politics
  • Markets
  • Crypto
  • Economics
    • Manufacturing
    • Real Estate
    • Infrastructure
  • Finance
  • Energy
  • Creator Economy
  • Wealth Management
  • Taxes
  • Telecoms
  • Military & Defense
  • Careers
  • Technology
  • Artificial Intelligence
  • Investigative journalism
  • Art & Culture
  • Documentaries
  • Quizzes
    • Enneagram quiz
  • Newsletters
    • LBNN Newsletter
    • Divergent Capitalist

© 2023 LBNN - All rights reserved.