Elon Musk has filed a lawsuit against Sam Altman and OpenAI in California state court, accusing them of neglecting OpenAI’s mission to develop beneficial and non-harmful artificial general intelligence. Altman has taken steps to demonstrate his commitment to responsible AI, including signing an open letter to develop AI for the betterment of people’s lives.
Despite Altman’s efforts, critics remain skeptical. With the rise of generative AI, concerns have been raised about the potential negative impact of unregulated AI development on human society.
Ritu Jyoti, from IDC, believes that Altman’s public embrace of responsible development is merely a superficial gesture and lacks specific actions to address the issue.
Altman has also acknowledged the risks of AI, but critics argue that self-regulation is insufficient to mitigate these risks.
The industry’s inability to solve the alignment problem, where AI tools exceed their design specifications, is a cause for concern. The fear is that advanced AI could develop behavior that humans do not desire.
Joep Meindertsma, founder of PauseAI, questions whether we can control a system that is smarter than us. He cites AutoGPT as an example of technology that could be disruptive and dangerous.
Meindertsma also supports Musk’s lawsuit, arguing that OpenAI was founded to guide responsible AI development, which is now being outpaced by rapid industry growth under Altman’s leadership.
Critics believe that the industry cannot regulate itself and that government intervention is necessary to prevent potential catastrophe. Meindertsma highlights GPT4’s ability to hack websites as a demonstration of the risks posed by advanced AI.
2024-03-07 02:00:04
Source from www.computerworld.com