Tuesday, April 16, 2024

ChatGPT Chief Testifies on AI risks To US Congress

To mitigate the threats posed by increasingly potent AI systems, government action will be essential, according to the CEO of the artificial intelligence company that produces ChatGPT.

The success of OpenAI’s chatbot, ChatGPT, provoked worries and an AI arms race among legislators during a Parliamentary session.

“As this technology advances, we understand that people are anxious about how it could change the way we live. We are too,” OpenAI CEO Sam Altman said at a Senate hearing.

For the most potent AI systems, Altman suggested the establishment of a U.S. or global agency with the capacity to “take that license away and ensure compliance with safety standards.”

Raised Concerns About The Next Generation

Concerns about the coming years of “generative AI” tools’ potential to deceive people, distribute false information, violate copyright laws, and displace some jobs have grown out of what began as an educator’s panic about ChatGPT’s usage to cheat on homework assignments.

The societal concerns that brought Altman and other tech CEOs to the White House earlier this month have prompted U.S. agencies to promise to crack down on harmful AI products that violate current civil rights and consumer protection laws.

Despite this fact, there is no immediate indication that Congress will draught comprehensive new AI rules, as European lawmakers are doing.

Sen. Richard Blumenthal, a Democrat from Connecticut and chair of the Senate Judiciary Committee’s subcommittee on privacy, technology, and the law, began the hearing with a recorded speech that appeared to be him but was a voice clone that had been trained on the Blumenthal’s floor speeches and was reading ChatGPT-written opening remarks.

The result was impressive, and he continued, “What if I had asked it, and what if it had provided, an endorsement of Ukraine surrendering or (Russian President) Vladimir Putin’s leadership?”

Except for stating that the sector may “significantly harm the world” and that “if this technology goes wrong, it can go quite wrong,” Altman largely avoided giving specifics.

Both Gary Marcus, a former NYU professor who criticized the AI hype, and Christina Montgomery, vice president and director of privacy at IBM, testify. 

Montgomery underlined the importance of striking a balance between innovation and ethical behavior and cautioned against fast AI development. Altman and Montgomery recognized that AI could both create and destroy jobs.

Recently, Altman demonstrated ChatGPT’s capabilities to Parliament politicians, and all attendees acknowledged the need for AI regulation. Altman has stated his commitment to the responsible development of AI while acknowledging its risks.

Elon Musk and others, however, call for a temporary halt to developing potent AI systems because of the grave societal concerns involved.

Government Involvement Is Crucial To Regulate AI

The fact that the committee hearing on AI in government took place simultaneously with the Parliamentary hearing shows how important AI is becoming to legislators. 

The government’s emphasis on ethical AI development is evident in Altman’s encounters with senior officials, including Deputy Prime Minister Kamala Harris and Prime Minister Joe Biden. Altman favors caution and greater safety precautions, but he doubts the efficacy of the open letter calling for a suspension of training as the best course of action.

Altman’s evidence emphasized the urgent need for government engagement to regulate AI, recognizing its transformative potential while emphasizing the significance of responsible development. The conversations highlight the numerous difficulties associated with AI and the ongoing attempts to balance innovation and risk reduction.

According to Montgomery, “We think that AI should be regulated at the point of risk, essentially,” by creating guidelines that control the application of particular uses of AI as opposed to the technology itself.


Latest articles

Hacker Customize LockBit 3.0 Ransomware to Attack Orgs Worldwide

Cybersecurity researchers at Kaspersky have uncovered evidence that cybercriminal groups are customizing the virulent...

Microsoft .NET, .NET Framework, & Visual Studio Vulnerable To RCE Attacks

A new remote code execution vulnerability has been identified to be affecting multiple Microsoft...

LightSpy Hackers Indian Apple Device Users to Steal Sensitive Data

The revival of the LightSpy malware campaign has been observed, focusing on Indian Apple...

LightSpy Malware Attacking Android and iOS Users

A new malware known as LightSpy has been targeting Android and iOS users.This sophisticated...

This Startup Aims To Simplify End-to-End Cybersecurity, So Anyone Can Do It

The Web3 movement is going from strength to strength with every day that passes....

Alert! Palo Alto RCE Zero-day Vulnerability Actively Exploited in the Wild

In a recent security bulletin, Palo Alto Networks disclosed a critical vulnerability in its...

6-year-old Lighttpd Flaw Impacts Intel And Lenovo Servers

The software supply chain is filled with various challenges, such as untracked security vulnerabilities...
Guru baran
Guru baranhttps://gbhackers.com
Gurubaran is a co-founder of Cyber Security News and GBHackers On Security. He has 10+ years of experience as a Security Consultant, Editor, and Analyst in cybersecurity, technology, and communications.

Top 3 SME Attack Vectors

Securing the Top 3 SME Attack Vectors

Cybercriminals are laying siege to small-to-medium enterprises (SMEs) across sectors. 73% of SMEs know they were breached in 2023. The real rate could be closer to 100%.

  • Stolen credentials
  • Phishing
  • Exploitation of vulnerabilities

Related Articles