Artificial intelligence poses an “existential risk” to humanity, a key innovator warned during a visit to the United Arab Emirates this week, suggesting an international agency like the International Atomic Energy Agency oversee the ground-breaking technology. OpenAI CEO Sam Altman is on a global tour to discuss artificial intelligence. “We face serious risk. We face existential risk,” said Altman, 38. “The challenge that the world has is how we’re going to manage those risks and make sure we still get to enjoy those tremendous benefits. No one wants to destroy the world.” OpenAI’s ChatGPT, a popular chatbot, has grabbed the world’s attention as it offers essaylike answers to prompts from users. Microsoft has invested some $1 billion in OpenAI. ChatGPT’s success, offering a glimpse into the way that artificial intelligence could change the way that humans work and learn, has sparked concerns as well.
Hundreds of industry leaders, including Altman, have signed a letter in May that warns “mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war.” Altman made a point to reference the IAEA, the United Nations nuclear watchdog, as an example of how the world came together to oversee nuclear power. That agency was created in the years after the U.S. dropping atom bombs on Japan at the end of World War II. “Let’s make sure we come together as a globe — and I hope this place can play a real role in this,” Altman said. “We talk about the IAEA as a model where the world has said ‘OK, very dangerous technology, let’s all put some guard rails.’ And I think we can do both. “I think in this case, it’s a nuanced message ‘cause it’s saying it’s not that dangerous today but it can get dangerous fast. But we can thread that needle.” Lawmakers around the world also are examining artificial intelligence.
The 27-nation European Union is pursuing an AI Law that could become the de facto global standard for artificial intelligence. Altman told the U.S. Congress in May that government intervention will be critical to governing the risks that come with AI. But the UAE, an autocratic federation of seven hereditarily ruled sheikhdoms, offers the flip side of the risks of AI. Speech remains tightly controlled. Rights groups warn the UAE and other states across the Persian Gulf regularly use spying software to monitor activists, journalists and others. Those restrictions affect the flow of accurate information — the same details AI programs like ChatGPT rely on as machinelearning systems to provide their answers for users.
Among speakers opening for Altman at the event at the Abu Dhabi Global Market was Andrew Jackson, the CEO of the Inception Institute of AI, which is described as a company of G42. G42 is tied to Abu Dhabi’s powerful national security adviser and deputy ruler Sheikh Tahnoun bin Zayed Al Nahyan. G42’s CEO is Peng Xiao, who for years ran Pegasus, a subsidiary of DarkMatter, an Emirati security firm under scrutiny for hiring former CIA and NSA staffers, as well as others from Israel. G42 also owns a video and voice calling app that reportedly was a spying tool for the Emirati government. In his remarks, Jackson described himself as representing “the Abu Dhabi and UAE AI ecosystem.” “We are a political powerhouse and we will be central to AI regulation globally,” he said.