Apple And Google Caught Steering Users Towards Controversial 'Nudify' Apps

  • By Nico
  • April 17, 2026, 7 a.m.

Tech Giants in Hot Water: Apple and Google Under Scrutiny

In a surprising twist, Apple and Google – the titans of tech – have found themselves under the microscope for inadvertently guiding users toward apps capable of creating AI-generated nudes. A recent investigation by the Tech Transparency Project (TTP) has shed light on the app stores' role in promoting these controversial 'nudify' apps, contrary to their stated guidelines against such content.

TTP's detailed probe revealed that searches for terms like "nudify" and "deepnude" on the app stores unveiled numerous apps designed to digitally undress women in photos. The platforms not only displayed these apps prominently but also ran ads and suggested related search terms, amplifying their visibility. This inadvertent endorsement has led to these apps accumulating over 483 million downloads and generating more than $122 million in revenue.

“These findings highlight a stark contradiction between the app stores’ policies and their actual practices,” a TTP representative stated, emphasizing the urgency for tighter regulation.

Age Ratings and Policy Violations

Alarmingly, the investigation identified 31 'nudify' apps rated suitable for minors, raising concerns amidst increasing school-related deepfake scandals. Despite Apple and Google's existing policies against sexually suggestive content, these platforms seem to be lagging in enforcement.

Apple reportedly declined to comment on the issue, while Google spokesperson Dan Jackson mentioned ongoing enforcement actions. Nonetheless, the sheer scale of the problem suggests a significant oversight in their app review processes.

Testing the Apps: A Closer Look

As part of their methodology, TTP tested popular apps from the search results, finding that a significant number could undress women or swap their faces onto nude bodies. The apps often utilized AI technology to generate these explicit images, sometimes even requiring payment for full access.

Despite some developers claiming ignorance or adjusting features post-investigation, the apps continue to push the boundaries of app store policies. Notably, apps like Best Body AI and AI Replace & Remove featured prominently in the investigation, showcasing the ease with which users can exploit these tools for unethical purposes.

Ads and Search Suggestions: Aiding App Discoverability

TTP's research also highlighted how Apple's and Google's ad systems inadvertently promoted these 'nudify' apps. Apple's ad placements – identifiable by their light blue backgrounds – often topped search results, while Google's keyword-based ads enhanced app discoverability.

Autocomplete suggestions further compounded the issue by directing users to even more 'nudify' apps. Experts are now calling for more rigorous scrutiny and adjustments to both companies' search and advertisement algorithms to curb the spread of such apps.

Implications and Next Steps

The TTP's findings underscore a critical need for Apple and Google to reevaluate their app discovery and content moderation strategies. With public concern mounting over privacy and the potential for misuse, particularly among minors, tech giants must enhance their oversight and align actions with their policies.

As the digital landscape continues to evolve, industry leaders face mounting pressure to ensure their platforms do not become conduits for unethical technological exploitation.

Nico
Author: Nico
Nico

Nico

Nico tracks the pulse of SoCal creator culture - from WeHo nights to TikTok mornings. He chases viral moments, fan deals, collabs, and live events with fast, human coverage. Expect Q&As, “Hot Now” briefs, and field notes that tell you what’s popping and why it matters. If it’s trending by noon, Nico had it at breakfast.