AI poses 'extinction' risk, say experts

AI poses 'extinction' risk, say experts

ChatGPT burst into the spotlight late last year, sparking huge investment but also widespread criticism
ChatGPT burst into the spotlight late last year, sparking huge investment but also widespread criticism. Photo: Marco BERTORELLO / AFP/File
Source: AFP

PAY ATTENTION: Never miss breaking news – join Briefly News' Telegram channel!

Global leaders should be working to reduce "the risk of extinction" from artificial intelligence technology, a group of industry chiefs and experts warned on Tuesday.

A one-line statement signed by dozens of specialists, including Sam Altman whose firm OpenAI created the ChatGPT bot, said tackling the risks from AI should be "a global priority alongside other societal-scale risks such as pandemics and nuclear war".

ChatGPT burst into the spotlight late last year, demonstrating an ability to generate essays, poems and conversations from the briefest of prompts.

The program's wild success sparked a gold rush with billions of dollars of investment into the field, but critics and insiders have raised the alarm.

Common worries include the possibility that chatbots could flood the web with disinformation, that biased algorithms will churn out racist material, or that AI-powered automation could lay waste to entire industries.

Read also

Nvidia chief says tech at 'tipping point' as unveils AI products

Superintelligent machines

The latest statement, housed on the website of US-based non-profit Center for AI Safety, gave no detail of the potential existential threat posed by AI.

PAY ATTENTION: Click “See First” under the “Following” tab to see Briefly News on your News Feed!

The center said the "succinct statement" was meant to open up a discussion on the dangers of the technology.

'If something goes wrong with AI, no gas mask is going to help you,' OpenAI boss Sam Altman has said
'If something goes wrong with AI, no gas mask is going to help you,' OpenAI boss Sam Altman has said. Photo: JOEL SAGET / AFP/File
Source: AFP

Several of the signatories, including Geoffrey Hinton, who created some of the technology underlying AI systems and is known as one of the godfathers of the industry, have made similar warnings in the past.

Their biggest worry has been the rise of so-called artificial general intelligence (AGI) -- a loosely defined concept for a moment when machines become capable of performing wide-ranging functions and can develop their own programming.

The fear is that humans would no longer have control over superintelligent machines, which experts have warned could have disastrous consequences for the species and the planet.

Read also

AI political fakery sparks fears for US 2024 race

Dozens of academics and specialists from companies including Google and Microsoft -- both leaders in the AI field -- signed the statement.

It comes two months after Tesla boss Elon Musk and hundreds of others issued an open letter calling for a pause in the development of such technology until it could be shown to be safe.

However, Musk's letter sparked widespread criticism that dire warnings of societal collapse were hugely exaggerated and often reflected the talking points of AI boosters.

US academic Emily Bender, who co-wrote an influential papers criticising AI, said the March letter, signed by hundreds of notable figures, was "dripping with AI hype".

'Surprisingly non-biased'

Bender and other critics have slammed AI firms for refusing to publish the sources of their data or reveal how it is processed -- the so-called "black box" problem.

Large Language Models on the rise
Large Language Models on the rise. Photo: Sabrina BLANCHARD, Julia Han JANICKI / AFP
Source: AFP

Among the criticism is that the algorithms could be trained on racist, sexist or politically biased material.

Read also

Hong Kong treads fine line on regulating retail crypto trade

Altman, who is currently touring the world in a bid to help shape the global conversation around AI, has hinted several times at the global threat posed by the technology his firm is developing.

"If something goes wrong with AI, no gas mask is going to help you," he told a small group of journalists in Paris last Friday.

But he defended his firm's refusal to publish the source data, saying critics really just wanted to know if the models were biased.

"How it does on a racial bias test is what matters there," he said, adding that the latest model was "surprisingly non-biased".

PAY ATTENTION: Сheck out news that is picked exactly for YOU ➡️ click on “Recommended for you” and enjoy!

Source: AFP

Authors:
AFP avatar

AFP AFP text, photo, graphic, audio or video material shall not be published, broadcast, rewritten for broadcast or publication or redistributed directly or indirectly in any medium. AFP news material may not be stored in whole or in part in a computer or otherwise except for personal and non-commercial use. AFP will not be held liable for any delays, inaccuracies, errors or omissions in any AFP news material or in transmission or delivery of all or any part thereof or for any damages whatsoever. As a newswire service, AFP does not obtain releases from subjects, individuals, groups or entities contained in its photographs, videos, graphics or quoted in its texts. Further, no clearance is obtained from the owners of any trademarks or copyrighted materials whose marks and materials are included in AFP material. Therefore you will be solely responsible for obtaining any and all necessary releases from whatever individuals and/or entities necessary for any uses of AFP material.