Wednesday, June 4, 2025
LBNN
  • Business
  • Markets
  • Politics
  • Crypto
  • Finance
  • Energy
  • Technology
  • Taxes
  • Creator Economy
  • Wealth Management
  • Documentaries
No Result
View All Result
LBNN

A new method for tailoring large AI models

Simon Osuji by Simon Osuji
December 10, 2024
in Artificial Intelligence
0
A new method for tailoring large AI models
0
SHARES
1
VIEWS
Share on FacebookShare on Twitter


Black-box forgetting: A new method for tailoring large AI models
Selective forgetting aims to reduce the classification accuracy for classes to be forgotten while maintaining the accuracy for the classes to be remembered. The proposed method, which targets the image classifier model CLIP, achieves selective forgetting by optimizing the input text prompt, since the model itself is a “black box.” Credit: Go Irie, Tokyo University of Science

The capabilities of large-scale pre-trained AI models have recently skyrocketed, as demonstrated by large-scale vision-language models like CLIP or ChatGPT. These typical generalist models can perform reasonably well in tasks covering a large variety of fields, which has paved the way for their widespread adoption by the public. However, such versatility no doubt comes at a cost.

Related posts

AI deployemnt security and governance, with Deloitte

AI deployemnt security and governance, with Deloitte

June 4, 2025
AI enables shift from enablement to strategic leadership

AI enables shift from enablement to strategic leadership

June 3, 2025

Training and operating large-scale models consume extreme amounts of energy and time, which goes against sustainability goals and limits the types of computers they can be deployed on. Moreover, in many practical applications, people want AI models to fulfill specific roles rather than be jacks-of-all-trades.

In such cases, a model’s generalist capabilities might be useless and even counter-productive, reducing accuracy. Could there be a way to leverage large-scale pre-trained models more efficiently by having them “forget” unnecessary information?

In a recent paper that will be presented in Neural Information Processing Systems (NeurIPS 2024), a research team led by Associate Professor Go Irie from Tokyo University of Science (TUS), Japan, sought to tackle this problem. They developed a methodology dubbed “black-box forgetting,” by which one can iteratively optimize the text prompts presented to a black-box vision-language classifier model to have it selectively “forget” some of the classes it can recognize. Co-authors of this study included Mr. Yusuke Kuwana and Mr. Yuta Goto, both from TUS, as well as Dr. Takashi Shibata from NEC Corporation.

“In practical applications, the classification of all kinds of object classes is rarely required. For example, in an autonomous driving system, it would be sufficient to recognize limited classes of objects such as cars, pedestrians, and traffic signs. We would not need to recognize food, furniture, or animal species,” explains Dr. Irie.

“Retaining the classes that do not need to be recognized may decrease overall classification accuracy, as well as cause operational disadvantages such as the waste of computational resources and the risk of information leakage.”

Although some methods for selective forgetting in pre-trained models do exist, these assume a white-box setting, where the user has access to the internal parameters and architecture of the model. More often than not, users deal with black-boxes; they do not have access to the model itself or most of its information due to commercial or ethical reasons. Thus, the researchers had to employ a so-called derivative-free optimization strategy—one that does not require access to the model’s gradients.

Black-box forgetting: A new method for tailoring large AI models
Existing derivative-free optimization techniques are computationally expensive because they optimize a long, concatenated list of latent contexts derived from the prompts. In this study, a more effective context parametrization method called Latent Context Sharing (LCS) is used. LCS assumes that each latent context consists of both unique components and components common to all contexts, and optimizes them independently. This strategy is based on the idea that there is semantic similarity between contexts, suggesting the existence of common components. Optimizing the smaller shared and unique contexts leads to a reduction in the dimensionality of the problem, making it easier to handle. Credit: Go Irie, Tokyo University of Science

To this end, they extended a method known as CMA-ES, with the image classifier model CLIP as the target model for this study. This evolutionary algorithm involves sampling various candidate prompts to feed into the model and evaluating the results via predefined objective functions, updating a multivariate distribution based on the calculated values.

However, the performance of derivative-free optimization techniques deteriorates quickly for large-scale problems. As more classes need to be forgotten, the “latent context” used to optimize the input prompts grows to unmanageable sizes. To address this issue, the research team came up with a new parametrization technique called “latent context sharing.”

This approach involves decomposing latent context derived from prompts into various smaller elements, which are considered to be “unique” to a prompt token or “shared” between multiple tokens. By aiming to optimize for these smaller units rather than large chunks of latent context, the dimensionality of the problem can be greatly reduced, making it much more tractable.

The researchers validated their approach using several benchmark image classification datasets, trying to get CLIP to “forget” 40% of the classes in a given dataset. This marks the first study in which the goal is to have a pre-trained vision-language model fail to recognize specific classes under black-box conditions and, based on reasonable performance baselines, the results were very promising.

This innovative method has important implications in the field of artificial intelligence and machine learning. It could help large-scale models perform better in specialized tasks, extending their already astounding applicability. Another use, for example, would be to prevent image generation models from producing undesirable content by having them forget specific visual contexts.

In addition, the proposed method could help tackle privacy issues, which are a rising concern in the field. “If a service provider is asked to remove certain information from a model, this can be accomplished by retraining the model from scratch by removing problematic samples from the training data. However, retraining a large-scale model consumes enormous amounts of energy,” says Dr. Irie. “Selective forgetting, or so-called machine unlearning, may provide an efficient solution to this problem.”

In other words, it could help develop solutions for protecting the so-called “Right to be Forgotten,” which is a particularly sensitive topic in health care and finances.

More information:
Yusuke Kuwana et al, Black-Box Forgetting, arXiv (2024). DOI: 10.48550/arxiv.2411.00409

Journal information:
arXiv

Provided by
Tokyo University of Science

Citation:
Black-box forgetting: A new method for tailoring large AI models (2024, December 9)
retrieved 9 December 2024
from https://techxplore.com/news/2024-12-black-method-tailoring-large-ai.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.





Source link

Previous Post

Star Capital backs Aberdeen-based STR in multi-million pound deal

Next Post

‘Shock the system’: Startups and DOGE take over Reagan forum

Next Post
‘Shock the system’: Startups and DOGE take over Reagan forum

‘Shock the system’: Startups and DOGE take over Reagan forum

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

RECOMMENDED NEWS

Russia’s Putin Says Ukraine Would Last ‘a Week’ if Western Aid Stops

Russia’s Putin Says Ukraine Would Last ‘a Week’ if Western Aid Stops

2 years ago
The New ‘Ethical’ AI Music Generator Can’t Write a Halfway Decent Song

The New ‘Ethical’ AI Music Generator Can’t Write a Halfway Decent Song

11 months ago
Expedia’s quarterly performance: Insights into the unexpected dip

Expedia’s quarterly performance: Insights into the unexpected dip

1 year ago
Xsolla Announces Acquisition of Lightstream, Rainmaker and API.stream

Xsolla Announces Acquisition of Lightstream, Rainmaker and API.stream

2 years ago

POPULAR NEWS

  • Ghana to build three oil refineries, five petrochemical plants in energy sector overhaul

    Ghana to build three oil refineries, five petrochemical plants in energy sector overhaul

    0 shares
    Share 0 Tweet 0
  • When Will SHIB Reach $1? Here’s What ChatGPT Says

    0 shares
    Share 0 Tweet 0
  • Matthew Slater, son of Jackson State great, happy to see HBCUs back at the forefront

    0 shares
    Share 0 Tweet 0
  • Dolly Varden Focuses on Adding Ounces the Remainder of 2023

    0 shares
    Share 0 Tweet 0
  • US Dollar Might Fall To 96-97 Range in March 2024

    0 shares
    Share 0 Tweet 0
  • Privacy Policy
  • Contact

© 2023 LBNN - All rights reserved.

No Result
View All Result
  • Home
  • Business
  • Politics
  • Markets
  • Crypto
  • Economics
    • Manufacturing
    • Real Estate
    • Infrastructure
  • Finance
  • Energy
  • Creator Economy
  • Wealth Management
  • Taxes
  • Telecoms
  • Military & Defense
  • Careers
  • Technology
  • Artificial Intelligence
  • Investigative journalism
  • Art & Culture
  • Documentaries
  • Quizzes
    • Enneagram quiz
  • Newsletters
    • LBNN Newsletter
    • Divergent Capitalist

© 2023 LBNN - All rights reserved.