Italy is temporarily blocking the artificial intelligence software ChatGPT in the wake of a data breach as it investigates a possible violation of stringent European Union data protection rules, the government鈥檚 privacy watchdog said Friday.
The Italian Data Protection Authority said it was taking provisional action 鈥渦ntil ChatGPT respects privacy,鈥 including temporarily limiting the company from processing Italian users鈥 data.
U.S.-based OpenAI, which developed ChatGPT, didn鈥檛 immediately return a request for comment Friday.
While some public schools and universities around the world have blocked the ChatGPT website from their local networks over student plagiarism concerns, it鈥檚 not clear how Italy would block it at a nationwide level.
The move also is unlikely to affect applications from companies that already have licenses with OpenAI to use the same technology driving the chatbot, such as Microsoft鈥檚 Bing search engine.
The AI systems that power such chatbots, known as large language models, are able to mimic human writing styles based on the huge trove of digital books and online writings they have ingested.
The Italian watchdog said OpenAI must report within 20 days what measures it has taken to ensure the privacy of users鈥 data or face a fine of up to either 20 million euros (nearly $22 million) or 4% of annual global revenue.
The agency鈥檚 statement cites the EU鈥檚 General Data Protection Regulation and noted that ChatGPT suffered a data breach on March 20 involving 鈥渦sers鈥 conversations鈥 and information about subscriber payments.
OpenAI earlier announced that it had to take ChatGPT offline on March 20 to fix a bug that allowed some people to see the titles, or subject lines, of other users鈥 chat history.
鈥淥ur investigation has also found that 1.2% of ChatGPT Plus users might have had personal data revealed to another user,鈥 the company said. 鈥淲e believe the number of users whose data was actually revealed to someone else is extremely low and we have contacted those who might be impacted.鈥
Italy鈥檚 privacy watchdog lamented the lack of a legal basis to justify OpenAI鈥檚 鈥渕assive collection and processing of personal data鈥 used to train the platform鈥檚 algorithms and that the company does not notify users whose data it collects.
The agency also said ChatGPT can sometimes generate 鈥 and store 鈥 false information about individuals.
Finally, it noted there鈥檚 no system to verify users鈥 ages, exposing children to responses 鈥渁bsolutely inappropriate to their age and awareness.鈥
The watchdog鈥檚 move comes as concerns grow about the artificial intelligence boom. A group of scientists and tech industry leaders published a letter Wednesday calling for companies such as OpenAI to pause the development of more powerful AI models until the fall to give time for society to weigh the risks.
鈥淲hile it is not clear how enforceable these decisions will be, the very fact that there seems to be a mismatch between the technological reality on the ground and the legal frameworks of Europe鈥 shows there may be something to the letter鈥檚 call for a pause 鈥渢o allow for our cultural tools to catch up,鈥 said Nello Cristianini, an AI professor at the University of Bath.
San Francisco-based OpenAI鈥檚 CEO, Sam Altman, announced this week that he鈥檚 embarking on a six-continent trip in May to talk about the technology with users and developers. That includes a stop planned for Brussels, where European Union lawmakers have been negotiating sweeping new rules to limit high-risk AI tools, as well as visits to Madrid, Munich, London and Paris.
European consumer group BEUC called Thursday for EU authorities and the bloc鈥檚 27 member nations to investigate ChatGPT and similar AI chatbots. BEUC said it could be years before the EU鈥檚 AI legislation takes effect, so authorities need to act faster to protect consumers from possible risks.
鈥淚n only a few months, we have seen a massive take-up of ChatGPT, and this is only the beginning,鈥 Deputy Director General Ursula Pachl said.
Waiting for the EU鈥檚 AI Act 鈥渋s not good enough as there are serious concerns growing about how ChatGPT and similar chatbots might deceive and manipulate people.鈥