(August 14 12:45) NDTV News
Apple has further detailed that its child safety mechanism will require at least 30 photos matching with Child Sexual Abuse Material (CSAM) flagged by organisations in at least two countries before an...
You can find the original article
here