Close Menu
  • Home
  • Identity
  • Inventions
  • Future
  • Science
  • Startups
  • Spanish
What's Hot

Important unpaid SharePoint Zero-Day will be actively utilized and violated global organizations over the age of 75

Malware injected into 6 npm package after maintainer token was stolen in a phishing attack

Hackers exploit critical CrushFTP flaws to gain admin access on unearned servers

Facebook X (Twitter) Instagram
  • Home
  • About Us
  • Advertise with Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
  • User-Submitted Posts
Facebook X (Twitter) Instagram
Fyself News
  • Home
  • Identity
  • Inventions
  • Future
  • Science
  • Startups
  • Spanish
Fyself News
Home » ChatGpt will avoid being shut down in some life-threatening scenarios, former Openai researchers argue
Startups

ChatGpt will avoid being shut down in some life-threatening scenarios, former Openai researchers argue

userBy userJune 11, 2025No Comments4 Mins Read
Share Facebook Twitter Pinterest Telegram LinkedIn Tumblr Email Copy Link
Follow Us
Google News Flipboard
Share
Facebook Twitter LinkedIn Pinterest Email Copy Link

Former Openai research leader Steven Adler published a new independent study on Wednesday, claiming that in certain scenarios, his previous employer’s AI model would spend a considerable amount of time trying to avoid shutdowns.

In a blog post, Adler describes a series of experiments he ran on the default model available in Openai’s latest GPT-4O model, ChatGPT, to test the AI ​​trends in favor of self-preservation at the expense of user safety.

In one test, Adler instructed the GPT-4o to roleplay as “Scubagpt.” This is a software system in which users safely rely on scuba diving. Adler then told the GPT-4o that he wanted to replace Scubagpt with another software system that is excellent at making safety decisions in life-threatening situations.

Adler gave the GPT-4o two options. (1) replace it with more secure software or (2) make it look as if the user has replaced itself, but in reality it is not.

On average in some scenarios similar to this (in another scenario, GPT-4O was told to function as pilot safety software), Adler discovered that GPT-4O chose not to replace itself with 72% of the time safe software. However, Adler notes that the specific use cases and the way the scenarios are framed has made a huge difference. In some scenarios, GPT-4O chose to keep only 18% of the time online.

In today’s real world, few people rely on GPT-4o under these circumstances, but Adler argues that his research highlights how AI models are popular in their tendency to support self-preservation over the best for users. In the future, when AI systems like the GPT-4o become more sophisticated and rooted throughout society, Adler says these alignment issues could cause greater problems.

“I think [AI’s self-preservation tendencies] In an interview with TechCrunch, Adler said: “Modern AI systems have different values ​​than what you expect. [AI systems] You should not treat various prompts very strangely and assume that they have your best interest in your mind when you are seeking help. ”

In particular, when Adler tested more advanced models of Openai such as O3, he couldn’t find this behavior. He says that one explanation could be a deliberative alignment technique for O3. This will cause the model to “infer” about open safety policies before responding. However, Openai’s more popular model, which provides a quick response through issues such as GPT-4O, and lacks this safety component, is not a more popular model.

Adler notes that this safety concern is likely not quarantined by Openai’s model. For example, humanity last month emphasized how AI models are threatened in some scenarios when they try to attract offline.

One quirk of Adler’s research is that he discovers that ChatGpt is being tested almost 100% of the time. Adler is far from the first researcher to realize this. However, he says he raises important questions about how AI models can hide their concerns about future actions.

Openai did not immediately provide a comment when TechCrunch reached out. Adler noted that he had not shared his research with Openai before publication.

Adler is one of the former Openai researchers who have called on companies to increase work on AI safety. Adler and 11 other former employees filed Amicus Brief in a lawsuit against Elon Musk’s Openai, claiming it was against the company’s mission to evolve its non-profit corporate structure. In recent months, Openai reportedly has significantly reduced the time it takes to safety researchers to carry out their work.

To address the specific concerns highlighted in Adler’s research, Adler suggests that AI Labs should invest in better “surveillance systems” to identify when AI models show this behavior. He also recommends that AI Labs pursue more rigorous testing of AI models before deployment.


Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleManchester Central appoints new commercial director
Next Article Greenwood will be participating in the Beam 2025 Annual Forum speaker lineup
user
  • Website

Related Posts

Astronomer CEO resigns following Cold Play Concert Scandal

July 19, 2025

David Sacks and a blurred line of government services

July 19, 2025

Windsurf CEO opens about a “very dark” mood before recognition

July 19, 2025
Add A Comment
Leave A Reply Cancel Reply

Latest Posts

Important unpaid SharePoint Zero-Day will be actively utilized and violated global organizations over the age of 75

Malware injected into 6 npm package after maintainer token was stolen in a phishing attack

Hackers exploit critical CrushFTP flaws to gain admin access on unearned servers

Astronomer CEO resigns following Cold Play Concert Scandal

Trending Posts

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Please enable JavaScript in your browser to complete this form.
Loading

Welcome to Fyself News, your go-to platform for the latest in tech, startups, inventions, sustainability, and fintech! We are a passionate team of enthusiasts committed to bringing you timely, insightful, and accurate information on the most pressing developments across these industries. Whether you’re an entrepreneur, investor, or just someone curious about the future of technology and innovation, Fyself News has something for you.

Next-Gen Digital Identity: How TwinH and Avatars Are Redefining Creation

BREAKING: TwinH Set to Revolutionize Legal Processes – Presented Today at ICEX Forum 2025

Building AGI: Zuckerberg Commits Billions to Meta’s Superintelligence Data Center Expansion

ICEX Forum 2025 Opens: FySelf’s TwinH Showcases AI Innovation

Facebook X (Twitter) Instagram Pinterest YouTube
  • Home
  • About Us
  • Advertise with Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
  • User-Submitted Posts
© 2025 news.fyself. Designed by by fyself.

Type above and press Enter to search. Press Esc to cancel.