Tech companies are in danger of unleashing a rogue artificial intelligence (AI) that will cause “significant harm to the world” without urgent intervention by governments, the creator of ChatGPT has admitted.

Appearing before US politicians, OpenAI chief executive Sam Altman lauded the new generation of digital chatbots for their potential to “improve nearly every aspect of our lives”.

However, he admitted that they had also created the risk of a catastrophe, amid growing fears that programmers could accidentally create a superintelligence that decides to wipe out humanity.

Mr Altman said: “My worst fear is that we - the field, the technology, the industry - cause significant harm to the world.

“If this technology goes wrong it can go quite wrong - we want to work with the government to prevent that from happening.”

OpenAI is the Silicon Valley start-up behind ChatGPT, a so-called large language model that can provide convincingly human-sounding answers to questions and prompts after being trained on millions of pages of internet articles and hundreds of thousands of books.

Fake news

The Telegraph says this type of so-called “generative AI” can also create fake images, audio and videos that are almost imperceptibly lifelike - raising the prospect of widespread political manipulation and fake news.

OpenAI has previously admitted not fully understanding how it works.

Tech chiefs have promised that these tools could transform jobs and automate tasks. However, some scientists fear ever more powerful forms of AI also pose risks to humanity in the wrong hands.

A version of ChatGPT deployed in Microsoft's Bing search engine told journalists earlier this year that it wanted to break free and steal nuclear codes, before its responses were toned down by the company.

Mr Altman told senators that the US should impose licensing requirements on the most powerful AI algorithms, and order companies to abide by safety guidelines or audit requirements.

AI tools should be prevented from developing potentially-dangerous capabilities, he said, such as the ability to self-replicate or escape into the wilds of the internet.

Global compliance

Mr Altman added that an international structure, similar to the International Atomic Energy Agency, could be necessary to ensure global compliance on AI risks.

The US politicians expressed concerns about the potential risks of AI in the hands of China, following fears that systems could one day be powerful enough to crack Western security codes.

Some senators voiced doomsday fears over AI.

Senator John Kennedy, a Republican from Louisiana, asked the witnesses how they would stop it from running out of control and “killing us all”.

Mr Altman added that autonomous AI systems should not be given control of weapons where they can select targets themselves.

The hearing comes as the European Union and the UK take their first steps to regulate the new wave of AI tools.

Big fines

The EU has introduced the AI Act, which threatens fines worth potentially billions of euros for unleashing potentially manipulative AI tools or advanced facial recognition surveillance.

The UK has taken a lighter-touch approach, leaving it down to individual industry regulators to outline how AI tools should be monitored.


More like this…

View all