Monday, May 19, 2025
LBNN
  • Business
  • Markets
  • Politics
  • Crypto
  • Finance
  • Energy
  • Technology
  • Taxes
  • Creator Economy
  • Wealth Management
  • Documentaries
No Result
View All Result
LBNN

OpenAI’s Strawberry program is reportedly capable of reasoning. It might be able to deceive humans

Simon Osuji by Simon Osuji
September 25, 2024
in Artificial Intelligence
0
OpenAI’s Strawberry program is reportedly capable of reasoning. It might be able to deceive humans
0
SHARES
0
VIEWS
Share on FacebookShare on Twitter


abstract strawberry
Credit: Pixabay/CC0 Public Domain

OpenAI, the company that made ChatGPT, has launched a new artificial intelligence (AI) system called Strawberry. It is designed not just to provide quick responses to questions, like ChatGPT, but to think or “reason.”

Related posts

Social robots learning without us? New study cuts humans from early testing

Social robots learning without us? New study cuts humans from early testing

May 19, 2025
21 Best High School Graduation Gifts (2025)

21 Best High School Graduation Gifts (2025)

May 18, 2025

This raises several major concerns. If Strawberry really is capable of some form of reasoning, could this AI system cheat and deceive humans?

OpenAI can program the AI in ways that mitigate its ability to manipulate humans. But the company’s own evaluations rate it as a “medium risk” for its ability to assist experts in the “operational planning of reproducing a known biological threat”—in other words, a biological weapon. It was also rated as a medium risk for its ability to persuade humans to change their thinking.

It remains to be seen how such a system might be used by those with bad intentions, such as con artists or hackers. Nevertheless, OpenAI’s evaluation states that medium-risk systems can be released for wider use—a position I believe is misguided.

Strawberry is not one AI “model,” or program, but several—known collectively as o1. These models are intended to answer complex questions and solve intricate math problems. They are also capable of writing computer code—to help you make your own website or app, for example.

An apparent ability to reason might come as a surprise to some, since this is generally considered a precursor to judgment and decision making—something that has often seemed a distant goal for AI. So, on the surface at least, it would seem to move artificial intelligence a step closer to human-like intelligence.

When things look too good to be true, there’s often a catch. Well, this set of new AI models is designed to maximize their goals. What does this mean in practice? To achieve its desired objective, the path or the strategy chosen by AI may not always necessarily be fair, or align with human values.

True intentions

For example, if you were to play chess against Strawberry, in theory, could its reasoning allow it to hack the scoring system rather than figure out the best strategies for winning the game?

The AI might also be able to lie to humans about its true intentions and capabilities, which would pose a serious safety concern if it were to be deployed widely. For example, if the AI knew it was infected with malware, could it “choose” to conceal this fact in the knowledge that a human operator might opt to disable the whole system if they knew?

These would be classic examples of unethical AI behavior, where cheating or deceiving is acceptable if it leads to a desired goal. It would also be quicker for the AI, as it wouldn’t have to waste any time figuring out the next best move. It may not necessarily be morally correct, however.

This leads to a rather interesting yet worrying discussion. What level of reasoning is Strawberry capable of and what could its unintended consequences be? A powerful AI system that’s capable of cheating humans could pose serious ethical, legal and financial risks to us.

Such risks become grave in critical situations, such as designing weapons of mass destruction. OpenAI rates its own Strawberry models as “medium risk” for their potential to assist scientists in developing chemical, biological, radiological and nuclear weapons.

OpenAI says: “Our evaluations found that o1-preview and o1-mini can help experts with the operational planning of reproducing a known biological threat.” But it goes on to say that experts already have significant expertise in these areas, so the risk would be limited in practice. It adds, “The models do not enable non-experts to create biological threats, because creating such a threat requires hands-on laboratory skills that the models cannot replace.”

Powers of persuasion

OpenAI’s evaluation of Strawberry also investigated the risk that it could persuade humans to change their beliefs. The new o1 models were found to be more persuasive and more manipulative than ChatGPT.

OpenAI also tested a mitigation system that was able to reduce the manipulative capabilities of the AI system. Overall, Strawberry was labeled a medium risk for “persuasion” in Open AI’s tests.

Strawberry was rated low risk for its ability to operate autonomously and on cybersecurity.

Open AI’s policy states that “medium risk” models can be released for wide use. In my view, this underestimates the threat. The deployment of such models could be catastrophic, especially if bad actors manipulate the technology for their own pursuits.

This calls for strong checks and balances that will only be possible through AI regulation and legal frameworks, such as penalizing incorrect risk assessments and the misuse of AI.

The UK government stressed the need for “safety, security and robustness” in their 2023 AI white paper, but that’s not nearly enough. There is an urgent need to prioritize human safety and devise rigid scrutiny protocols for AI models such as Strawberry.

Provided by
The Conversation

This article is republished from The Conversation under a Creative Commons license. Read the original article.The Conversation

Citation:
OpenAI’s Strawberry program is reportedly capable of reasoning. It might be able to deceive humans (2024, September 25)
retrieved 25 September 2024
from https://techxplore.com/news/2024-09-openai-strawberry-capable-humans.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.





Source link

Previous Post

How to Make an AI-Hosted Podcast with Google’s NotebookLM

Next Post

Sad tales of Rukia Wamuyu & Sharon Chebet

Next Post
Sad tales of Rukia Wamuyu & Sharon Chebet

Sad tales of Rukia Wamuyu & Sharon Chebet

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

RECOMMENDED NEWS

AGR has won a well management contract from Apus Energy

AGR has won a well management contract from Apus Energy

1 year ago
Overcoming the curse of Big Oil

Overcoming the curse of Big Oil

1 year ago
ARC Ltd. Receives $27 Million From The Dutch Government To Help Build Africa’s Climate Resilience

ARC Ltd. Receives $27 Million From The Dutch Government To Help Build Africa’s Climate Resilience

11 months ago
Successful US strikes on Houthi facilities and weapons systems

Successful US strikes on Houthi facilities and weapons systems

6 months ago

POPULAR NEWS

  • Ghana to build three oil refineries, five petrochemical plants in energy sector overhaul

    Ghana to build three oil refineries, five petrochemical plants in energy sector overhaul

    0 shares
    Share 0 Tweet 0
  • When Will SHIB Reach $1? Here’s What ChatGPT Says

    0 shares
    Share 0 Tweet 0
  • Matthew Slater, son of Jackson State great, happy to see HBCUs back at the forefront

    0 shares
    Share 0 Tweet 0
  • Dolly Varden Focuses on Adding Ounces the Remainder of 2023

    0 shares
    Share 0 Tweet 0
  • US Dollar Might Fall To 96-97 Range in March 2024

    0 shares
    Share 0 Tweet 0
  • Privacy Policy
  • Contact

© 2023 LBNN - All rights reserved.

No Result
View All Result
  • Home
  • Business
  • Politics
  • Markets
  • Crypto
  • Economics
    • Manufacturing
    • Real Estate
    • Infrastructure
  • Finance
  • Energy
  • Creator Economy
  • Wealth Management
  • Taxes
  • Telecoms
  • Military & Defense
  • Careers
  • Technology
  • Artificial Intelligence
  • Investigative journalism
  • Art & Culture
  • Documentaries
  • Quizzes
    • Enneagram quiz
  • Newsletters
    • LBNN Newsletter
    • Divergent Capitalist

© 2023 LBNN - All rights reserved.