Close Menu
  • Academy
  • Events
  • Identity
  • International
  • Inventions
  • Startups
    • Sustainability
  • Tech
  • Spanish
What's Hot

SuperBlocks CEO: How to study AI system prompts and find unicorn ideas

“Bitcoin Family” changed its security after the recent cryptocurrency

AB will be released at Binance -Tech Startups

Facebook X (Twitter) Instagram
  • Home
  • About Us
  • Advertise with Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
  • User-Submitted Posts
Facebook X (Twitter) Instagram
Fyself News
  • Academy
  • Events
  • Identity
  • International
  • Inventions
  • Startups
    • Sustainability
  • Tech
  • Spanish
Fyself News
Home » Openai’s latest AI model has a new safeguard to prevent biology
Startups

Openai’s latest AI model has a new safeguard to prevent biology

userBy userApril 16, 2025No Comments3 Mins Read
Share Facebook Twitter Pinterest Telegram LinkedIn Tumblr Email Copy Link
Follow Us
Google News Flipboard
Share
Facebook Twitter LinkedIn Pinterest Email Copy Link

Openai says it has deployed a new system to monitor the latest AI inference models, O3 and O4-MINI, for prompts related to biological and chemical threats. According to Openai’s safety report, the system aims to prevent the model from providing advice that can instruct someone to carry out a potentially harmful attack.

The O3 and O4-MINI represent an increase in meaningful capabilities across Openai’s previous models, the company says, bringing new risks to bad actors’ hands. According to Openai’s internal benchmarks, O3 is particularly skilled at answering questions about creating certain types of biological threats. For this reason, and to mitigate other risks, Openai has created a new monitoring system. This is described by the company as a “safety-focused inference monitor.”

Custom trained monitors to infer about Openai’s content policies run on O3 and O4-Mini. It is designed to identify prompts related to biological and chemical risks and instruct the model to refuse to provide advice on these topics.

To establish a baseline, Openai spent about 1,000 hours on Red Teamers, flagging “unsafe” Biorisk-related conversations from O3 and O4-Mini. During a test that Openai simulated the “block logic” of the safety monitor, Openai said the model refused to respond to a 98.7% risky prompt.

Openai admits it didn’t test people who might try new prompts after being blocked by the monitor, so the company says it will continue to rely in part on human surveillance.

According to the company, O3 and O4-MINI should not cross the “high risk” threshold of BioRisk Openai. However, compared to O1 and GPT-4, Openai says that early versions of the O3 and O4-Mini proved to be more useful in answering questions about biological weapon development.

Charts from O3 and O4-Mini system cards (screenshot: Openai)

According to Openai’s recently updated preparation framework, the company is actively tracking how malicious users can facilitate chemical and biological threats.

Openai is increasingly relying on automated systems to mitigate risks from the model. For example, to prevent the native image generators of GPT-4O from creating child sex abuse materials (CSAM), Openai says it uses similar inference monitors to the companies deployed on O3 and O4-Mini.

However, some researchers have raised concerns that Openai does not prioritize its inherent safety. Metr, one of the company’s redness partners, said he has relatively little time to test O3s in benchmarks of deceptive behavior. Meanwhile, Openai has decided not to release a safety report for the GPT-4.1 model, which was released earlier this week.


Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleThe real impact of emerging high-tech startups on traditional industries
Next Article AI in Education: Revolution or Risk? An Ethical Debate Reaches Spanish Classrooms
user
  • Website

Related Posts

SuperBlocks CEO: How to study AI system prompts and find unicorn ideas

June 7, 2025

After data is wiped out, Kiranapro co-founders cannot rule out external hacks

June 7, 2025

2025 will be a “vital year” for the enhanced virtual reality of meta, says CTO

June 6, 2025
Add A Comment
Leave A Reply Cancel Reply

Latest Posts

SuperBlocks CEO: How to study AI system prompts and find unicorn ideas

“Bitcoin Family” changed its security after the recent cryptocurrency

AB will be released at Binance -Tech Startups

After data is wiped out, Kiranapro co-founders cannot rule out external hacks

Trending Posts

Sana Yousaf, who was the Pakistani Tiktok star shot by gunmen? |Crime News

June 4, 2025

Trump says it’s difficult to make a deal with China’s xi’ amid trade disputes | Donald Trump News

June 4, 2025

Iraq’s Jewish Community Saves Forgotten Shrine Religious News

June 4, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Please enable JavaScript in your browser to complete this form.
Loading

Welcome to Fyself News, your go-to platform for the latest in tech, startups, inventions, sustainability, and fintech! We are a passionate team of enthusiasts committed to bringing you timely, insightful, and accurate information on the most pressing developments across these industries. Whether you’re an entrepreneur, investor, or just someone curious about the future of technology and innovation, Fyself News has something for you.

AB will be released at Binance -Tech Startups

Top 10 Startups and Tech Funding News for the Weekly Ends June 6, 2025

Order openai to keep all chatgpt logs including deleted temporary chats, API requests

Omada Health is now available: Virtual Care Startup joins IPO Wave, paying $150 million, $1.1 billion valuation of NASDAQ debut

Facebook X (Twitter) Instagram Pinterest YouTube
  • Home
  • About Us
  • Advertise with Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
  • User-Submitted Posts
© 2025 news.fyself. Designed by by fyself.

Type above and press Enter to search. Press Esc to cancel.