Close Menu
  • Home
  • Identity
  • Inventions
  • Future
  • Science
  • Startups
  • Spanish
What's Hot

AI roadmap, if anyone is willing to listen.

Grammarly’s “Expert Reviews” Just Lacking Actual Experts

Caitlin Kalinowski, head of OpenAI’s robotics division, resigns following agreement with Department of Defense

Facebook X (Twitter) Instagram
  • Home
  • About Us
  • Advertise with Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
  • User-Submitted Posts
Facebook X (Twitter) Instagram
Fyself News
  • Home
  • Identity
  • Inventions
  • Future
  • Science
  • Startups
  • Spanish
Fyself News
Home » AI roadmap, if anyone is willing to listen.
Startups

AI roadmap, if anyone is willing to listen.

userBy userMarch 8, 2026No Comments4 Mins Read
Share Facebook Twitter Pinterest Telegram LinkedIn Tumblr Email Copy Link
Follow Us
Google News Flipboard
Share
Facebook Twitter LinkedIn Pinterest Email Copy Link

The U.S. government’s break with Anthropic revealed that there are no consistent rules governing artificial intelligence, but a bipartisan coalition of thinkers has put together something the government has so far refused to create: a framework for what responsible AI development should actually look like.

Although the pro-human declaration was finalized before last week’s standoff between the Pentagon and humanity, the collision of the two events was not forgotten by all involved.

“Something very remarkable has happened in America over the last four months,” Max Tegmark, an MIT physicist and AI researcher who helped organize the effort, said in a conversation with the editors. “Sudden vote” [is showing] 95% of all Americans oppose the uncontrolled race to superintelligence. ”

A newly released document signed by hundreds of experts, former government officials and celebrities begins with the blunt observation that humanity is at a crossroads. One path, which the Declaration calls “competition of substitution”, leads to the replacement of humans first as workers and then as decision-makers, as power accrues to unaccountable organizations and their machines. The other will lead to AI that will greatly expand human potential.

The latter scenario relies on five key pillars: holding humans accountable, avoiding the concentration of power, protecting the human experience, protecting individual freedom, and holding AI companies legally accountable. Among its stronger provisions is a complete ban on superintelligence development until there is a scientific consensus that it is safe and capable of genuine democratic buy-in. Powerful systems require an off switch. and a prohibition on architectures capable of self-replication, autonomous self-improvement, or shutdown tolerance.

The publication of this declaration coincides with a time when its urgency is more easily understood. On the last Friday in February, Secretary of Defense Pete Hegseth designated Anthropic, whose AI is already running on classified military platforms, as a “supply chain risk.” The company denied the Pentagon permission to use the technology unrestricted, usually given to companies with ties to China. Hours later, OpenAI terminated its own agreement with the Department of Defense, which legal experts say will be difficult to enforce in any meaningful way. This reveals just how costly Congressional inaction on AI has become.

Dean Ball, a senior fellow at the American Foundation for Innovation, later told the New York Times: “This is not just a fight over a contract. This is the first conversation we’ve had as a nation about managing our AI systems.”

tech crunch event

San Francisco, California
|
October 13-15, 2026

Tegmark came up with an analogy that most people understood when we talked about it. “We don’t have to worry that some drug company is going to come out with another drug that does tremendous harm before people figure out how to make that drug safe, because the FDA won’t allow anything to come out until it’s safe enough,” he said.

Turf wars in Washington rarely generate enough public pressure to change the law. Rather, Tegmark believes that child safety is the most likely solution to the current impasse. In fact, the declaration calls for mandatory pre-deployment testing of AI products, especially chatbots and companion apps aimed at younger users, to cover risks such as increased suicidal ideation, poor mental health and emotional manipulation.

“If a creepy old man pretends to be a girl and sends an email to an 11-year-old boy trying to convince the boy to commit suicide, he could be sent to prison for that,” Tegmark said. “We already have laws. It’s illegal. So why would it be any different if a machine does it?”

He believes that once the principles of pre-release testing are established for children’s products, its scope will almost inevitably expand. “People will come and say, let’s add some other requirements. Maybe we should also test that this doesn’t contribute to terrorist bioweapon production. Maybe we should test to make sure superintelligence agencies don’t have the ability to overthrow the U.S. government.”

It’s no small feat that former President Trump adviser Steve Bannon and President Obama’s national security adviser Susan Rice signed the same document, along with former Chairman of the Joint Chiefs of Staff Mike Mullen and progressive faith leaders.

“What they agree on, of course, is that they’re all humans,” Tegmark says. “When it comes to whether you want a human future or a machine future, of course they’re going to be on the same side.”


Source link

#Aceleradoras #CapitalRiesgo #EcosistemaStartup #Emprendimiento #InnovaciónEmpresarial #Startups
Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleGrammarly’s “Expert Reviews” Just Lacking Actual Experts
user
  • Website

Related Posts

Grammarly’s “Expert Reviews” Just Lacking Actual Experts

March 7, 2026

Caitlin Kalinowski, head of OpenAI’s robotics division, resigns following agreement with Department of Defense

March 7, 2026

PopSockets founder David Barnett talks building a viral business

March 7, 2026
Add A Comment
Leave A Reply Cancel Reply

Latest Posts

AI roadmap, if anyone is willing to listen.

Grammarly’s “Expert Reviews” Just Lacking Actual Experts

Caitlin Kalinowski, head of OpenAI’s robotics division, resigns following agreement with Department of Defense

PopSockets founder David Barnett talks building a viral business

Trending Posts

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Please enable JavaScript in your browser to complete this form.
Loading

Welcome to Fyself News, your go-to platform for the latest in tech, startups, inventions, sustainability, and fintech! We are a passionate team of enthusiasts committed to bringing you timely, insightful, and accurate information on the most pressing developments across these industries. Whether you’re an entrepreneur, investor, or just someone curious about the future of technology and innovation, Fyself News has something for you.

Castilla-La Mancha Ignites Innovation: fiveclmsummit Redefines Tech Future

Local Power, Health Innovation: Alcolea de Calatrava Boosts FiveCLM PoC with Community Engagement

The Future of Digital Twins in Healthcare: From Virtual Replicas to Personalized Medical Models

Human Digital Twins: The Next Tech Frontier Set to Transform Healthcare and Beyond

Facebook X (Twitter) Instagram Pinterest YouTube
  • Home
  • About Us
  • Advertise with Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
  • User-Submitted Posts
© 2026 news.fyself. Designed by by fyself.

Type above and press Enter to search. Press Esc to cancel.