With its ability to analyze huge datasets, recognize patterns, make decisions, and adapt to new information, AI is transforming industries from health care and finance to transportation and entertainment.
Tech experts including Tesla CEO Elon Musk, AI godfather Geoffrey Hinton and Sam Altman, CEO of Microsoft-backed OpenAI, have already raised concerns over AI’s potential to erase humanity and are calling on governments to introduce regulations to govern the technology. Have requested.
However, rapid advances in AI such as OpenAI’s ChatGPT are complicating governments’ efforts to agree on laws governing the use of the technology, reuters informed of.
Governments and companies need to address the risks of AI head-on, Prime Minister Rishi Sunak said on October 26, ahead of the first global AI safety summit at Bletchley Park on November 1-2.
Sunak said the UK would set up the world’s first AI safety institute, which would “understand what each new model is capable of, exploring all the risks, from the most extreme risks to social harms such as bias and misinformation”.
Britain’s data watchdog on October 10 (Tuesday) said it had issued a preliminary enforcement notice to Snap Inc’s Snapchat over potential failure to properly assess the privacy risks of its generative AI chatbot to users, especially children.
Here are the latest steps other international governing bodies are taking to regulate AI devices:
Australia Its internet regulator said in September that search engines would draft new codes to prevent the sharing of child sexual abuse material generated by AI and the production of deepfake versions of the same content.
China on October 12 published proposed security requirements for companies offering services powered by generative AI, including a blacklist of sources that cannot be used to train AI models.
The country issued a set of temporary measures in August, requiring service providers to submit security assessments and obtain approval before releasing AI products on a large scale.
European lawmakers agreed on Oct. 24 to a key portion of new AI rules, outlining the types of systems that will be designated “high risk,” according to five people familiar with the matter, and a broader build on the landmark AI Act. The agreement has been reached.
An agreement is expected in December, the two co-reporters said.
European Commission President Ursula von der Leyen on September 13 called for a global panel to assess the risks and benefits of AI.
France’s privacy watchdog said in April it was investigating complaints about ChatGPT.
In May G7 leaders called for the development and adoption of technical standards to keep AI “trusted.”
Italy’s data protection authority plans to review AI platforms and hire experts in the field, a top official said in May. ChatGPT was temporarily banned in the country in March, but it was made available again in April.
An official close to the discussions said in July that Japan hopes to have rules in place by the end of 2023 that are closer to the US approach than stricter rules planned in the EU.
The country’s privacy watchdog has warned OpenAI not to collect sensitive data from people without their permission.
Poland’s Personal Data Protection Office said on September 21 that it was investigating OpenAI over complaints that ChatGPT breaks EU data protection laws.
Spain’s data protection agency launched a preliminary investigation into possible data breaches by ChatGPT in April.
The UN Security Council held its first formal discussion on AI in July, addressing military and non-military applications of AI, which “could have very serious consequences for global peace and security”, Secretary-General Antonio Guterres said. .
Guterres has supported a proposal by some AI officials to create an AI watchdog and has announced plans to begin work on a high-level AI advisory body by the end of the year.
The White House is expected to unveil a long-awaited AI executive order on October 30 that will require “advanced AI models to undergo evaluation before they can be used by federal employees,” the Washington Post reports.
In September, the US Congress held a hearing on AI and Meta AI Forum featuring CEOs Mark Zuckerberg and Musk.
More than 60 senators participated in the talks, during which Musk called on the US to be the “referee” for AI. The lawmakers said there is universal agreement about the need for government regulation of the technology.
On September 12, the White House said Adobe, IBM, Nvidia and five other firms had signed President Joe Biden’s voluntary commitments governing AI, which require steps such as watermarking AI-generated content .
A district judge in Washington DC ruled in August that works of art created by AI without any human input cannot be copyrighted under US law.
The US Federal Trade Commission in July began investigating OpenAI into claims that it violated consumer protection laws.