US lawmakers demand clarity on OpenAI safety practices in joint letter

Share This Post

A group of US Senators sent a detailed letter to OpenAI CEO and co-founder Sam Altman seeking clarity on the company’s safety measures and employment practices.

The Washington Post first reported about the joint letter on July 23.

Senators Brian Schatz, Ben Ray Luján, Peter Welch, Mark R. Warner, and Angus S. King, Jr. signed the joint letter, which has set an Aug. 13 deadline for the firm to provide a comprehensive response addressing the various concerns raised in it.

According to the July 22 letter, recent reports regarding potential issues at the company prompted the Senators’ inquiry. It emphasized the need for transparency in the deployment and governance of artificial intelligence (AI) systems due to issues of national security and public trust.

Lawmaker inquiry

The Senators have requested detailed information about several concerns, including confirmation on whether OpenAI will honor its previously pledged commitment to allocate 20% of its computing resources to AI safety research. The letter emphasized that fulfilling this commitment is crucial for the responsible development of AI technologies.

Additionally, the letter inquired about OpenAI’s enforcement of non-disparagement agreements and other contractual provisions that could potentially deter employees from raising safety concerns. The lawmakers stressed the importance of protecting whistleblowers and ensuring that employees can voice their concerns without fear of retaliation.

They also sought detailed information on the cybersecurity protocols OpenAI has in place to protect its AI models and intellectual property from malicious actors and foreign adversaries. They asked OpenAI to describe its non-retaliation policies and whistleblower reporting channels, emphasizing the need for robust protections against cybersecurity threats.

In their inquiry, the Senators asked whether OpenAI allows independent experts to test and assess the safety and security of its AI systems before they are released. They emphasized the importance of independent evaluations in ensuring the integrity and reliability of AI technologies.

The Senators also asked if OpenAI plans to conduct and publish retrospective impact assessments of its already-deployed models to ensure public accountability. They highlighted the need for transparency in evaluating the real-world effects of AI systems.

Critical role of AI

The letter highlighted AI’s critical role in the nation’s economic and geopolitical standing, noting that safe and secure AI is essential for maintaining competitiveness and protecting critical infrastructure.

The Senators stressed the importance of OpenAI’s voluntary commitments made to the Biden-Harris administration and urged the company to provide documentation on how it plans to meet these commitments.

The letter stated:

“Given OpenAI’s position as a leading AI company, it is important that the public can trust in the safety and security of its systems. This includes the integrity of the company’s governance structure and safety testing, its employment practices, its fidelity to its public promises and mission, and its cybersecurity policies.”

The letter marks a significant step in ensuring that AI development proceeds with the highest standards of safety, security, and public accountability. This action reflects the growing legislative scrutiny on AI technologies and their societal impacts.

The five lawmakers emphasized the urgency of addressing these issues, given the widespread use of AI technologies and their potential consequences for national security and public trust. They called on OpenAI to demonstrate its commitment to responsible AI development by providing thorough and transparent responses to their questions.

The Senators referenced several sources and previous reports that have detailed OpenAI’s challenges and commitments, providing a comprehensive backdrop for their concerns. These sources include OpenAI’s approach to frontier risk and the Biden-Harris administration’s voluntary safety and security commitments.

The post US lawmakers demand clarity on OpenAI safety practices in joint letter appeared first on CryptoSlate.

Read Entire Article
spot_img
- Advertisement -spot_img

Related Posts

SEC Reviews Proposal for New Bitcoin, Ether ETF Custodians

The US Securities and Exchange Commission (SEC) is seeking public input on a proposal by Cboe BZX Exchange to add new custodians for bitcoin and ethereum exchange-traded funds (ETFs) The plan aims to

Terraform Labs secures court approval to wind-down operations after settling with SEC

Terraform Labs received court approval to wind down its operations in bankruptcy after settling a US Securities and Exchange Commission (SEC) lawsuit, Reuters reported Sept 19 US Bankruptcy Judge

New Clipper Malware Threat Targets Shiba Inu Community, Here’s What To Know

As the world of cryptocurrency continues to evolve, several new methods are being developed by cybercriminals to scam crypto users One of the latest methods reported to be widely used by these

Bitcoin Reclaims $63,000 After US Fed Rate Cut, But Is This Rally For Real?

So far, Bitcoin has seen a mixture of bulls and bears just in the past day alone Although bulls appear to be taking the lead given its current market performance, will this be sustainable Before the

Louisiana continues pro-crypto push with option to pay state agencies in Bitcoin, USDC

Louisiana has implemented a new option for residents to make payments to state agencies using Bitcoin and Circle’s stablecoin USDC, according to a Sept 18 press release The Louisiana Department

Terraform Labs Gets Court Approval to Wind Down Operations in Bankruptcy

Terraform Labs has received court approval to wind down operations in bankruptcy following a settlement with the US Securities and Exchange Commission (SEC) The company had been at the center of a