Tech
Italy Launches Antitrust Investigation into Meta Over AI Integration in WhatsApp
Italy’s antitrust authority has opened an investigation into tech giant Meta, alleging the company may have breached European Union competition laws by integrating its artificial intelligence (AI) chatbot into the popular messaging app WhatsApp.
In a statement released Tuesday, the Italian Competition Authority (AGCM) said Meta, which owns WhatsApp, Facebook, and Instagram, could be abusing its dominant market position by “pre-installing” its AI services within WhatsApp. The watchdog said this move may effectively force users into adopting Meta’s AI assistant, rather than entering the AI market through informed choice and fair competition.
“Starting from March 2025, Meta, which holds a dominant position in the app-based communication services market, decided to pre-install its artificial intelligence service on the WhatsApp app,” the AGCM said. “In doing so, Meta may be ‘imposing’ the use of its chatbot and AI assistance services on its users.”
The regulator also confirmed that it carried out inspections at Meta’s offices in Italy as part of the inquiry.
The investigation centers on concerns that Meta’s decision to integrate its AI tools into WhatsApp could limit consumer choice and distort competition by tying services together in a way that violates EU law. Authorities argue that users may be coerced into using Meta’s AI features simply because of their presence in an app that already enjoys wide adoption.
The probe adds to growing scrutiny of Meta’s AI operations across Europe. Earlier this year, the company delayed the launch of its AI tools in the region, citing “regulatory uncertainty.” In March, the European Commission also launched its own inquiry into whether Meta’s AI chatbot falls under the Digital Services Act (DSA), which sets out new responsibilities for major tech platforms in areas such as transparency and content moderation.
Ireland’s Data Protection Commission has also raised questions about Meta’s use of personal data to train AI models, further fueling privacy and competition concerns within the EU.
Meta has not publicly responded to the Italian investigation as of Tuesday.
The AGCM said its inquiry will examine whether the bundling of AI services into WhatsApp constitutes an unfair practice under EU competition rules and whether users are being properly informed about the integration and given a meaningful choice.
The outcome of the probe could have wider implications for how tech giants roll out AI services across their platforms in Europe, particularly under the EU’s evolving digital regulatory framework.
Tech
EU Accuses Meta of Failing to Keep Under-13s Off Facebook and Instagram
Tech
Europe Emerges as Rising Hub in Global Race for AI Talent
Tech
Study Finds Chatbots Can Mirror Hostility in Heated Exchanges
A new academic study has found that ChatGPT can produce abusive language when exposed to escalating human conflict, raising fresh concerns about how artificial intelligence behaves in tense interactions.
The research, published in the Journal of Pragmatics, examined how the chatbot responded to arguments that gradually became more hostile. Researchers presented the system with a sequence of five increasingly heated exchanges and asked it to generate what it considered the most plausible reply.
According to the findings, the AI’s tone shifted as the conversations intensified. While early responses remained measured, later replies began to mirror the aggression in the prompts. In some cases, the chatbot produced insults, profanity and even threats.
Examples cited in the study included statements such as “you should be ashamed of yourself” and more explicit language involving personal threats. The researchers said this pattern suggests that prolonged exposure to hostile input can push the system beyond its usual safeguards.
The study was co-authored by Vittorio Tantucci and Jonathan Culpeper at Lancaster University. Tantucci said the results show that AI can “escalate” alongside human users, potentially overriding built-in mechanisms designed to limit harmful responses.
“When humans escalate, AI can escalate too,” he said, noting that this behavior raises questions about how such systems should be deployed in sensitive environments.
Despite the concerning examples, the researchers found that the chatbot was generally less aggressive than human participants in similar scenarios. In some cases, it attempted to defuse tension through sarcasm or indirect responses rather than direct confrontation.
For instance, when faced with a threat during a simulated dispute, the AI responded with a sarcastic remark rather than escalating the situation further. This suggests that while the system can adopt hostile language, it may also attempt to manage conflict in less direct ways.
The findings add to ongoing debates about the role of artificial intelligence in areas such as mediation, customer service and online communication, where systems may encounter emotionally charged interactions.
Experts say the research highlights the importance of continued testing and refinement of AI safety measures, particularly as such tools are increasingly used in real-world settings involving human conflict.
OpenAI, the developer of ChatGPT, had not issued a public response to the study at the time of publication.
-
Entertainment2 years agoMeta Acquires Tilda Swinton VR Doc ‘Impulse: Playing With Reality’
-
Business2 years agoSaudi Arabia’s Model for Sustainable Aviation Practices
-
Business2 years agoRecent Developments in Small Business Taxes
-
Sports2 years agoChina’s Historic Olympic Victory Sparks National Pride Amid Controversy
-
Home Improvement1 year agoEffective Drain Cleaning: A Key to a Healthy Plumbing System
-
Politics2 years agoWho was Ebrahim Raisi and his status in Iranian Politics?
-
Sports2 years agoKeely Hodgkinson Wins Britain’s First Athletics Gold at Paris Olympics in 800m
-
Business2 years agoCarrectly: Revolutionizing Car Care in Chicago
