• Business
  • Markets
  • Politics
  • Crypto
  • Finance
  • Intelligence
    • Policy Intelligence
    • Security Intelligence
    • Economic Intelligence
    • Fashion Intelligence
  • Energy
  • Technology
  • Taxes
  • Creator Economy
  • Wealth Management
  • LBNN Blueprints
  • Business
  • Markets
  • Politics
  • Crypto
  • Finance
  • Intelligence
    • Policy Intelligence
    • Security Intelligence
    • Economic Intelligence
    • Fashion Intelligence
  • Energy
  • Technology
  • Taxes
  • Creator Economy
  • Wealth Management
  • LBNN Blueprints

Researchers explore how to bring larger neural networks closer to the energy efficiency of biological brains

Simon Osuji by Simon Osuji
November 20, 2024
in Artificial Intelligence
0
Researchers explore how to bring larger neural networks closer to the energy efficiency of biological brains
0
SHARES
1
VIEWS
Share on FacebookShare on Twitter


lottery
Credit: Unsplash/CC0 Public Domain

The more lottery tickets you buy, the higher your chances of winning, but spending more than you win is obviously not a wise strategy. Something similar happens in AI powered by deep learning: we know that the larger a neural network is (i.e., the more parameters it has), the better it can learn the task we set for it.

Related posts

Make the Most of Chrome’s Toolbar by Customizing It to Your Liking

Make the Most of Chrome’s Toolbar by Customizing It to Your Liking

March 1, 2026
The 5 Big ‘Known Unknowns’ of Donald Trump’s New War With Iran

The 5 Big ‘Known Unknowns’ of Donald Trump’s New War With Iran

March 1, 2026

However, the strategy of making it infinitely large during training is not only impossible but also extremely inefficient. Scientists have tried to imitate the way biological brains learn, which is highly resource-efficient, by providing machines with a gradual training process that starts with simpler examples and progresses to more complex ones—a model known as “curriculum learning.”

Surprisingly, however, they found that this seemingly sensible strategy is irrelevant for overparameterized (very large) networks.

A study in the Journal of Statistical Mechanics: Theory and Experiment sought to understand why this “failure” occurs, suggesting that these overparameterized networks are so “rich” that they tend to learn by following a path based more on quantity (of resources) than quality (input organized by increasing difficulty).

This might actually be good news, as it suggests that by carefully adjusting the initial size of the network, curriculum learning could still be a viable strategy, potentially promising for creating more resource-efficient, and therefore less energy-consuming, neural networks.

There is great excitement towards neural network-based AI like ChatGPT: every day, a new bot or feature emerges that everyone wants to try, and the phenomenon is also growing in scientific research and industrial applications. This requires increasing computing power—and, therefore, energy consumption—and the concerns regarding both the energy sources needed and the emissions produced by this sector are on the rise. Making this technology capable of doing more with less is thus crucial.

Neural networks are computational models made up of many “nodes” performing calculations, with a distant resemblance to the networks of neurons in biological brains, capable of learning autonomously based on the input they receive. For example, they “see” a vast number of images and learn to categorize and recognize content without direct instruction.

Among experts, it is well known that the larger a neural network is during the training phase (i.e., the more parameters it uses), the more precisely it can perform the required tasks. This strategy is known in technical jargon as the “Lottery Ticket Hypothesis” and has the significant drawback of requiring a massive amount of computing resources, with all the associated problems (increasingly powerful computers are needed, which demand more and more energy).

To find a solution, many scientists have looked at where this type of problem appears to have been, at least partially, solved: biological brains. Our brains, with only two or three meals a day, can perform tasks that require supercomputers and a huge amount of energy for a neural network. How do they do it?

The order in which we learn things might be the answer. “If someone has never played the piano and you put them in front of a Chopin piece, they’re unlikely to make much progress learning it,” explains Luca Saglietti, a physicist at Bocconi University in Milan, who coordinated the study. “Normally, there’s a whole learning path spanning years, starting from playing ‘Twinkle Twinkle Little Star’ and eventually leading to Chopin.”

When input is provided to machines in an order of increasing difficulty, it is called “curriculum learning.” However, the most common way to train neural networks is to feed them input randomly into highly powerful, overparameterized networks.

Once the network has learned, it is possible to reduce the number of parameters—even lower than 10% of the initial amount—because they are no longer used. However, if you start with only 10% of the parameters, the network fails to learn. So, while an AI might eventually fit into our phone, during training, it requires massive servers.

Scientists have wondered whether curriculum learning could save resources. But research so far suggests that for very overparameterized networks, curriculum learning seems irrelevant: performance in the training phase does not seem to be improved.

The new work by Saglietti and colleagues attempted to understand why.

“What we’ve seen is that an overparameterized neural network doesn’t need this path because, instead of being guided through learning by examples, it’s guided by the fact that it has so many parameters—resources that are already close to what it needs,” explains Saglietti.

In other words, even if you offer it optimized learning data, the network prefers to rely on its vast processing resources, finding parts within itself that, with a few tweaks, can already perform the task.

This is actually good news, as it does not mean that networks cannot take advantage of curriculum learning, but that, given the high number of initial parameters, they are pushed in a different direction. In principle, therefore, one could find a way to start with smaller networks and adopt curriculum learning.

“This is one part of the hypothesis explored in our study,” Saglietti explains.

“At least within the experiments we conducted, we observed that if we start with smaller networks, the effect of the curriculum—showing examples in a curated order—begins to show improvement in performance compared to when the input is provided randomly. This improvement is greater than when you keep increasing the parameters to the point where the order of the input no longer matters.”

More information:
Stefano Sarao Mannelli et al, Tilting the odds at the lottery: the interplay of overparameterisation and curricula in neural networks*, Journal of Statistical Mechanics: Theory and Experiment (2024). DOI: 10.1088/1742-5468/ad864b

Provided by
International School of Advanced Studies (SISSA)

Citation:
Researchers explore how to bring larger neural networks closer to the energy efficiency of biological brains (2024, November 19)
retrieved 19 November 2024
from https://techxplore.com/news/2024-11-explore-larger-neural-networks-closer.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.





Source link

Previous Post

Blackstone Acquires Jersey Mike’s for $8 Billion

Next Post

Report: Iranian Weapon Deliveries Back Sudanese Armed Forces

Next Post
Report: Iranian Weapon Deliveries Back Sudanese Armed Forces

Report: Iranian Weapon Deliveries Back Sudanese Armed Forces

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

RECOMMENDED NEWS

Poco F6 and F6 Pro Review: Small Price, Big Performance

Poco F6 and F6 Pro Review: Small Price, Big Performance

2 years ago
How many registered voters in the United States?

How many registered voters in the United States?

10 months ago
How 2025 strengthened South African Agriculture

How 2025 strengthened South African Agriculture

2 months ago
Global Warming Is Wreaking Havoc on the Planet’s Water Cycle

Global Warming Is Wreaking Havoc on the Planet’s Water Cycle

1 year ago

POPULAR NEWS

  • Ghana to build three oil refineries, five petrochemical plants in energy sector overhaul

    Ghana to build three oil refineries, five petrochemical plants in energy sector overhaul

    0 shares
    Share 0 Tweet 0
  • Mahama attends Liberia’s 178th independence anniversary

    0 shares
    Share 0 Tweet 0
  • The world’s top 10 most valuable car brands in 2025

    0 shares
    Share 0 Tweet 0
  • Top 10 African countries with the highest GDP per capita in 2025

    0 shares
    Share 0 Tweet 0
  • Global ranking of Top 5 smartphone brands in Q3, 2024

    0 shares
    Share 0 Tweet 0

Get strategic intelligence you won’t find anywhere else. Subscribe to the Limitless Beliefs Newsletter for monthly insights on overlooked business opportunities across Africa.

Subscription Form

© 2026 LBNN – All rights reserved.

Privacy Policy | About Us | Contact

Tiktok Youtube Telegram Instagram Linkedin X-twitter
No Result
View All Result
  • Home
  • Business
  • Politics
  • Markets
  • Crypto
  • Economics
    • Manufacturing
    • Real Estate
    • Infrastructure
  • Finance
  • Energy
  • Creator Economy
  • Wealth Management
  • Taxes
  • Telecoms
  • Military & Defense
  • Careers
  • Technology
  • Artificial Intelligence
  • Investigative journalism
  • Art & Culture
  • LBNN Blueprints
  • Quizzes
    • Enneagram quiz
  • Fashion Intelligence

© 2023 LBNN - All rights reserved.