US, UK enter agreement to develop AI safety standards

Share This Post

The US and the UK signed a Memorandum of Understanding (MOU) to develop standards for AI Safety on April 1.

The agreement aims to develop and iterate “robust suites of evaluations” for AI models, systems, and agents through coordinated scientific approaches.

The two countries will specifically conduct at least one joint testing exercise on a publicly accessible AI model. They also intend to exchange personnel between their previously established AI Safety Institutes. The partnership takes effect immediately.

The US Department of Commerce and the UK’s Department for Science, Innovation and Technology announced the partnership in identical press releases.

Safe AI development

Commenting on the new international agreement, US Commerce Secretary Gina Raimondo called AI the “defining technology of our generation” and said that the partnership will address all risks related to the technology, from national security risks to societal risks.

Raimondo described the partnership’s proactive nature, stating:

“Our partnership makes clear that we aren’t running away from these concerns – we’re running at them.”

UK Technology Secretary Michelle Donelan called the agreement a “landmark moment” and called the pursuit of safe AI development a global issue.

In a separate interview with the Financial Times, Donelan emphasized the urgency of the partnership, considering the upcoming release of new AI models over the next year. She added that these models’ full capabilities are unknown, but they are expected to be “game-changing.”

Donelan highlighted that many top AI companies are US-based, which necessitates information exchange with the government.

Other AI developments

The current agreement builds on commitments to AI made during the UK’s AI Safety Summit, which was held in November 2023. The broader commitment to AI safety is part of the Bletchley Declaration, a policy paper from countries that attended the summit.

The news comes shortly after the completion of goals in the Biden Administration’s executive order in the US, which culminated in orders to appoint chief AI officers across all federal agencies on March 28.

The UK set out a local approach to regulation in February.

The post US, UK enter agreement to develop AI safety standards appeared first on CryptoSlate.

Read Entire Article
spot_img
- Advertisement -spot_img

Related Posts

South Korea Sees Crypto Boom: 30% of Population Now Owns Digital Assets

South Korea has seen a significant surge in cryptocurrency adoption, with the number of users jumping by 610,000 in November to reach 1559 million Trump Victory Sparks Renewed Interest in Crypto The

Bitcoin Reserve Idea Sparks Cautious Response From Japan PM: Report

According to a recent report by Japanese cryptocurrency publication CoinPost, Japan’s Prime Minister Shigeru Ishiba has taken a cautious stance on the proposal to establish a national Bitcoin (BTC)

BlackRock doubles down on IBIT exposure through its Global Allocation Fund

BlackRock reported holding 430,770 shares of its spot Bitcoin (BTC) exchange-traded fund (ETF) IBIT through its Global Allocation Fund According to a filing with the US Securities and Exchange

Dogecoin Price Repeats Bullish Fractal From 2021, Why January 2025 Is Important

The Dogecoin price is replicating a bullish fractal from 2021, signaling the potential for a price breakout to new highs A crypto analyst has shared a price chart comparing this historically

Defi Doesn’t Sleep—Mystery Whale Dumps 125,000 ETH Into Aave on Christmas Day

The Head of Research at Intotheblockcom revealed that on Christmas Day, someone stashed 125,000 ETH into the decentralized finance (defi) platform Aave The platform has been buzzing with activity

Join HTX’s Festive Season Party: Enjoy Superior Crypto Trading and Open the Door to Wealth

PRESS RELEASE As the year draws to a close, the festive spirit of Christmas and New Year fills the air HTX, a leading global digital assets trading platform, is kicking off its Festive Season Party,