The Microsoft Bing App is seen running on an iPhone in this photo illustration on 30 May, 2023 in Warsaw, Poland. (Photo by Jaap Arriens/NurPhoto via Getty Images)
Jaap Arriens | Nurphoto | Getty Images
Artificial intelligence may lead to human extinction and reducing the risks associated with the technology should be a global priority, industry experts and tech leaders stated in an open letter.
âMitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war,â the statement on Tuesday read.
related investing news

Sam Altman, CEO of ChatGPT-maker OpenAI, as well as executives from Googleâs AI arm DeepMind and Microsoft were among those who supported and signed the short statement from the Center for AI Safety.
The technology has gathered pace in recent months after chatbot ChatGPT was released for public use in November and subsequently went viral. In just two months after its launch, it reached 100 million users. ChatGPT has amazed researchers and the general public with its ability to generate humanlike responses to usersâ prompts, suggesting that AI could replace jobs and imitate humans.
The statement Tuesday said that there has been increasing discussion about a âbroad spectrum of important and urgent risks from AI.â
But it said it can be âdifficult to voice concerns about some of advanced AIâs most severe risksâ and had the aim of overcoming this obstacle and opening up the discussions.
ChatGPT has arguably sparked much more awareness and adoption of AI as major firms around the world have raced to develop rival products and capabilities.
Altman had admitted in March that he is a âlittle bit scaredâ of AI as he worries that authoritarian governments would develop the technology. Other tech leaders such as Teslaâs Elon Musk and former Google CEO Eric Schmidt have cautioned about the risks AI poses to society.
In an open letter in March, Musk, Apple co-founder Steve Wozniak and several tech leaders urged AI labs to stop training systems to be more powerful than GPT-4 â which is OpenAIâs latest large language model. They also called for a six-month pause on such advanced development.

âContemporary AI systems are now becoming human-competitive at general tasks,â said the letter.
âShould we automate away all the jobs, including the fulfilling ones? Should we develop nonhuman minds that might eventually outnumber, outsmart, obsolete and replace us? Should we risk loss of control of our civilization?â the letter asked.
Last week, Schmidt also separately warned about the âexistential risksâ associated with AI as the technology advances.