Elon Musk And Other Experts Advocate for Suspension of AI Training
In a move that has sent shockwaves throughout the technology industry, some of the biggest names in artificial intelligence have come together to call for the temporary suspension of AI training. The group, which includes Twitter chief Elon Musk and Apple co-founder Steve Wozniak, has signed an open letter warning of the potential risks of the unchecked development of AI systems.
The letter, issued by the Future of Life Institute, highlights the risks that AI systems with human-competitive intelligence could pose to society and humanity. The letter urges the development of advanced AIs to be done with care, warning that recent months have seen AI labs locked in an out-of-control race to develop and deploy ever more powerful digital minds that no one, not even their creators, can understand, predict, or reliably control.
The group has called for the immediate suspension of the training of AI systems more powerful than GPT-4 for at least six months. The letter warns that failure to enact such a delay quickly may necessitate the intervention of governments, with the creation of new and capable regulatory authorities dedicated to AI.
The call to action comes as OpenAI, the company behind the language model ChatGPT, recently released GPT-4, a state-of-the-art technology that has impressed observers with its ability to do tasks such as answering questions about objects in images. The letter warns that advanced AIs need to be developed with care, citing concerns that they could flood information channels with misinformation and replace jobs with automation.
While some experts have warned of the risks of AI, others have cautioned against the overreaction to potential risks. A recent report from investment bank Goldman Sachs highlighted the potential for AI to increase productivity while acknowledging that millions of jobs could become automated. However, other experts have cautioned that the effect of AI on the labor market is very hard to predict.
The letter raises further concerns about the development of non-human minds that might eventually outnumber, outsmart, obsolete, and replace humans. The group has called for coordinated efforts among AI labs to slow down at critical junctures, highlighting the importance of developing AGI (artificial general intelligence) with care and warning of the risks if an AGI were developed recklessly.
The open letter is likely to fuel ongoing debates around the regulation of technology, with a number of proposals for the regulation of AI having been put forward in the US, UK, and EU. While the UK has ruled out a dedicated regulator for AI, other countries are likely to face increasing pressure to take action in the face of the risks posed by the rapid development of AI systems.
Comments
Post a Comment