By Martin Coulter and Supantha Mukherjee
LONDON/STOCKHOLM (Reuters) -EU lawmakers urged world leaders on Monday to hold a summit to find ways to control the development of advanced artificial intelligence (AI) systems such as ChatGPT, saying they were developing faster than expected.
The 12 MEPs, all working on EU legislation on the technology, called on U.S. President Joe Biden and European Commission President Ursula von der Leyen to convene the meeting, and said AI firms should be more responsible.
The statement came weeks after Twitter owner Elon Musk and more than 1,000 technology figures demanded a six-month pause in the development of systems more powerful than Microsoft-backed OpenAI’s latest iteration of ChatGPT, which can mimic humans and create text and images based on prompts.
That open letter, published in March by the Future of Life Institute (FLI), had warned that AI could spread misinformation at an unprecedented rate, and that machines could “outnumber, outsmart, obsolete and replace” humans, if left unchecked.
The MEPS said they disagreed with some of the FLI message’s “more alarmist statements”.
“We are nevertheless in agreement with the letter’s core message: with the rapid evolution of powerful AI, we see the need for significant political action,” they added.
The letter urged democratic and “non-democratic” countries to reflect on potential systems of governance, and to exercise restraint in their pursuit of very powerful AI.
A spokesperson for von der Leyen did not immediately respond to a request for comment.
Last week China’s cyberspace regulator unveiled draft measures for managing generative AI services, saying it wants firms to submit security assessments to authorities before they launch their offerings to the public.
The Biden administration has also been seeking public comments on potential accountability measures for AI systems as questions loom about their impact on national security and education.
The European Commission proposed the draft rules for an AI Act nearly two years ago under which AI tools are expected to be classified according to their perceived level of risk, from low to unacceptable.
A parliamentary committee is debating the 108-page bill and hoping to reach a common position by April 26, according to two sources familiar with the matter.
(Reporting by Martin Coulter in London, Supantha Mukherjee in Stockholm and Akanksha Khushi in Bengaluru; Editing by Alex Richardson and Andrew Heavens)