Authorities found more than 30,000 images and videos of AI-generated child sexual abuse material across nine devices.
Nudification apps use AI to create naked or sexualised images of people from everyday, fully clothed photos that anyone might upload to social media.