Tech
Study Finds Polish Outperforms English as the Best Language for Communicating with AI
A new international study has revealed that Polish, not English, is the most effective language for communicating with artificial intelligence systems — a surprising finding that challenges long-held assumptions about English’s dominance in the tech world.
Researchers from the University of Maryland (UMD) and Microsoft tested 26 different languages across several leading AI models, including those developed by OpenAI, Google Gemini, Qwen, Llama, and DeepSeek. Their goal was to determine which languages yielded the most accurate and effective responses from the systems.
The results showed that Polish achieved an impressive 88 percent average accuracy rate in completing tasks, outperforming all other tested languages. French, Italian, and Spanish followed closely, while English ranked only sixth with an 83.9 percent accuracy rate.
“Our experiment yielded some surprising and unintuitive findings,” the researchers wrote. “English did not perform best across all models — in fact, it came sixth out of 26 languages when long texts were assessed, while Polish proved to be the leading language.”
The study’s outcome has drawn attention not only from linguists and AI experts but also from national institutions in Poland. The Polish Patent Office highlighted the results on social media, noting that “Polish has long been considered one of the most difficult languages to learn. It turns out that humans struggle with it — but not AI.”
The finding is particularly remarkable given the relatively small amount of Polish-language data used to train most AI systems. Despite this, the models demonstrated strong comprehension and precision in following Polish-language commands, outperforming languages with far greater digital footprints such as English and Chinese.
Chinese, despite being one of the most widely spoken languages in the world and heavily represented in online data, ranked near the bottom — fourth from last — among the 26 languages tested.
According to the study, the top ten languages most effectively understood by AI models are:
-
Polish – 88%
-
French – 87%
-
Italian – 86%
-
Spanish – 85%
-
Russian – 84%
-
English – 83.9%
-
Ukrainian – 83.5%
-
Portuguese – 82%
-
German – 81%
-
Dutch – 80%
The researchers said their findings highlight the complexity of how AI models process linguistic structures and suggest that grammatical richness and precision — characteristics of Polish — may actually help AI systems interpret meaning more accurately.
Tech
Cambridge Index Reveals Global Black Market for Fake Social Media Verifications
A new index developed by the University of Cambridge has revealed the scale and affordability of the underground market for fake social media account verifications, raising fresh concerns about online manipulation and digital security. According to researchers, fake verification badges can be purchased for as little as eight cents, enabling the rapid creation of networks that imitate authentic users across major online platforms.
The Cambridge Online Trust and Safety Index (COTSI), launched on Thursday, is described as the first global tool capable of tracking real-time prices for verifying fraudulent accounts. The index monitors more than 500 platforms, including TikTok, Instagram, Amazon, Spotify and Uber. By analysing data from sellers operating across the dark web and black-market channels, the project highlights how accessible and inexpensive these services have become.
Researchers say the low cost of creating fake accounts is contributing to the rise of “bot armies” — large groups of automated or semi-automated profiles designed to mimic genuine human activity. These networks can distort online conversations, amplify misleading content, and promote scams or commercial products. They can also be deployed to influence political messaging, creating an illusion of public support or opposition during major events such as elections or policy debates.
The team behind the index said the findings come at a sensitive time for governments and regulators working to contain misinformation. Many popular platforms have reduced investment in content monitoring during the past two years, while others have introduced programmes that reward users for generating high volumes of engagement. Researchers warn that such incentives may encourage the use of artificially inflated interactions, making fake accounts even more valuable to those seeking influence.
According to Cambridge analysts, the market for fraudulent verification has become highly sophisticated. Sellers offer tiered packages, guaranteeing features such as blue-badge symbols, verified rankings or the appearance of longstanding account history. Prices vary by platform and country, but the index shows that even the most complex packages remain within easy reach for groups attempting to manipulate public debate or carry out coordinated campaigns.
The launch of COTSI marks the first attempt to document these prices on a global scale. By presenting live data on the cost of creating fake identities, researchers hope to give policymakers, technology companies and security agencies a clearer picture of how digital manipulation is evolving. The study’s authors stress that tracking these markets is essential for understanding the risks posed by unauthenticated accounts, particularly during periods of political tension.
The university said the index will be updated regularly and will remain publicly accessible as part of its efforts to strengthen digital transparency worldwide.
Tech
Experts Question Impact of Australia’s New Social Media Ban for Children Under 16
Australia has introduced sweeping restrictions that prevent children under 16 from creating or maintaining accounts on major social media platforms, but experts warn the measures may not significantly change young people’s online behaviour. The restrictions, which took effect on December 10, apply to platforms including Facebook, Instagram, TikTok, Snapchat, YouTube, Twitch, Reddit and X.
Under the new rules, children cannot open accounts, yet they can still access most platforms without logging in—raising questions about how effective the regulations will be in shaping online habits. The eSafety Commissioner says the reforms are intended to shield children from online pressures, addictive design features and content that may harm their health and wellbeing.
Social media companies are required to block underage users through age-assurance tools that rely on facial-age estimation, ID uploads or parental consent. Ahead of the rollout, authorities tested 60 verification systems across 28,500 facial recognition assessments. The results showed that while many tools could distinguish children from adults, accuracy declined among users aged 16 and 17, girls and non-Caucasian users, where estimates could be off by two years or more. Experts say the limitations mean many teenagers may still find ways around the rules.
“How do they know who is 14 or 15 when the kids have all signed up as being 75?” asked Sonia Livingstone, a social psychology professor at the London School of Economics. She warned that misclassifications will be common as platforms attempt to enforce the regulations.
Meta acknowledged the challenge, saying complete accuracy is unlikely without requiring every user to present government ID—something the company argues would raise privacy and security concerns. Users over 16 who lose access by mistake are allowed to appeal.
Several platforms have criticised the ban, arguing that it removes teenagers from safer, controlled environments. Meta and Google representatives told Australian lawmakers that logged-in teenage accounts already come with protections that limit contact from unknown users, filter sensitive subjects and disable personalised advertising. Experts say these protections are not always effective, citing studies where new YouTube and TikTok accounts quickly received misogynistic or self-harm-related content.
Analysts expect many teenagers to shift to smaller or lesser-regulated platforms. Apps such as Lemon8, Coverstar and Tango have surged into Australia’s top downloads since the start of December. Messaging apps like WhatsApp, Telegram and Signal—exempt from the ban—have also seen a spike in downloads. Livingstone said teenagers will simply “find alternative spaces,” noting that previous bans in other countries pushed young users to new platforms within days.
Researchers caution that gaming platforms such as Discord and Roblox, also outside the scope of the ban, may become new gathering points for young Australians. Studies will be conducted to assess the long-term impact on mental health and whether the restrictions support or complicate parents’ efforts to regulate screen time.
Experts say it may take several years to determine whether the ban delivers meaningful improvements to children’s wellbeing.
Tech
OECD Warns of Sharp Rise in Cyberbullying Across Europe
-
Entertainment1 year agoMeta Acquires Tilda Swinton VR Doc ‘Impulse: Playing With Reality’
-
Business2 years agoSaudi Arabia’s Model for Sustainable Aviation Practices
-
Business2 years agoRecent Developments in Small Business Taxes
-
Home Improvement1 year agoEffective Drain Cleaning: A Key to a Healthy Plumbing System
-
Politics2 years agoWho was Ebrahim Raisi and his status in Iranian Politics?
-
Business1 year agoCarrectly: Revolutionizing Car Care in Chicago
-
Sports1 year agoKeely Hodgkinson Wins Britain’s First Athletics Gold at Paris Olympics in 800m
-
Business1 year agoSaudi Arabia: Foreign Direct Investment Rises by 5.6% in Q1
