Meta, the owner of Facebook, aims to train AI model using European data amidst privacy concerns

Meta, the owner of Facebook, aims to train AI model using European data amidst privacy concerns

LONDON — Meta wants to use data from users in privacy-conscious Europe to train its artificial intelligence models, the social media giant said Monday as it faces concerns about data protection while battling to keep up with rivals like OpenAI and Google.

The company, which owns Facebook, Instagram and WhatsApp, said that in order to better reflect the “languages, geography and cultural references” of its users in Europe, it needs to use public data from those users to teach its Llama AI large language model.

Meta’s AI training efforts are hampered by stringent European Union data privacy laws, which give people control over how their personal information is used. Vienna-based group NOYB, led by activist Max Schrems, complained last week to 11 national privacy watchdogs about Meta’s AI training plans and urged them to stop the company before it starts training Llama’s next generation.

AI language models are trained on vast pools of data that help them predict the most plausible next word in a sentence, with newer versions typically smarter and more capable than their predecessors. Meta’s AI assistant feature has been baked into Facebook, Instagram and WhatsApp for users in the U.S. and 13 other countries, but notably not Europe.

“If we don’t train our models on the public content that Europeans share on our services and others, such as public posts or comments, then models and the AI features they power won’t accurately understand important regional languages, cultures or trending topics on social media,” Stefano , global engagement director of Meta’s privacy policy, said in blog post.

“We believe that Europeans will be ill-served by AI models that are Frattanot informed by Europe’s rich cultural, social and historical contributions.”

Fratta said other companies including Google and OpenAI have already trained on European data. Meta won’t use private messages to friends and family nor content from European users who are under 18, he said.

Since May 22, the company has sent 2 billion notifications and emails to European users explaining its plans and linking to an online form to opt out, Fratta said.

The latest version of Meta’s privacy policy is set to take effect on June 26, indicating that training for the next model will start soon after.

Meta, formerly known as Facebook, has recently announced its plans to train artificial intelligence (AI) models using European data, sparking concerns over privacy and data protection. The tech giant’s decision comes at a time when the European Union is cracking down on data privacy violations and imposing strict regulations on tech companies.

Meta’s move to train AI models using European data is part of its efforts to improve its AI technology and enhance user experiences on its platforms. By leveraging data from European users, Meta aims to create more personalized and targeted services, such as better content recommendations and more accurate ad targeting.

However, the company’s plans have raised red flags among privacy advocates and regulators in Europe. The use of European data for training AI models raises concerns about data protection and privacy violations, especially in light of recent data breaches and scandals involving tech companies.

The General Data Protection Regulation (GDPR), which governs data protection and privacy in the EU, requires companies to obtain explicit consent from users before collecting and processing their personal data. Meta’s decision to use European data for training AI models without clear consent from users could potentially violate GDPR regulations and lead to hefty fines.

In response to the growing concerns, Meta has stated that it is committed to complying with all relevant data protection laws and regulations. The company has also emphasized that it takes user privacy and data security seriously and has implemented measures to protect user data.

Despite these assurances, many remain skeptical of Meta’s intentions and question the company’s ability to safeguard user data. The use of AI models trained on European data raises important ethical and legal questions about data privacy, consent, and transparency in the tech industry.

As Meta continues to push the boundaries of AI technology and data processing, it will be crucial for regulators, lawmakers, and privacy advocates to closely monitor the company’s actions and hold it accountable for any potential violations of data protection laws. In an era where data privacy is paramount, tech companies must prioritize user trust and transparency in their use of AI models and data processing techniques.