Sorry, you need to enable JavaScript to visit this website.
Skip to main content

The presence of content depicting child sexual abuse is a serious problem that has found its way into the digital sphere. The huge reach and resources of the large tech companies mean that they play a decisive role in combating the spread of this content.

Artificial intelligence and machine learning

Big technology companies such as Google, Meta, Microsoft and Amazon use advanced artificial intelligence and machine learning algorithms to detect and remove content depicting child sexual abuse. These technologies scan images, videos and text to identify potentially abusive content. For instance, Google’s Content Safety API uses machine learning to identify such content in video files, while Facebook’s photo- and video- matching technology can flag previously identified content.

Apart from these, other artificial intelligence tools are used for proactive identification. Microsoft’s PhotoDNA is an image identification and content filtering technology which generates a unique digital signature on an image, thereby enabling copies of known child sexual abuse material to be detected and removed. This technology is widely used on platforms, including Google, Facebook and Dropbox, to ensure swift identification and removal.

Cooperation and information sharing

Technology companies cooperate in sharing knowledge and resources through industry partnerships. These coalitions, formed by companies such as Google, Meta, Microsoft, Amazon and Twitter, concentrate on developing new technologies to prevent online exploitation and share best practices throughout the sector.

Many companies are taking part in initiatives where, by sharing and cross-referencing data, enterprises can quickly identify and remove known child sexual abuse material that appears on different platforms. This cooperation is crucial in ensuring that harmful content does not resurface on other platforms after it has already been removed from one platform.

User reporting mechanisms

The tech giants offer simple and accessible mechanisms for users to report child sexual abuse material. Platforms such as Instagram, YouTube, TikTok and Snapchat have transparent reporting functions that enable users to flag inappropriate content. These reports are then reviewed by human moderators, frequently with the help of artificial intelligence tools, to ensure swift action.

Additionally, clear and strict community guidelines also play a vital role. Companies regularly update their policies to address new threats and to ensure that a robust framework is in place to deal with child sexual abuse material. Infringements of these policies result in immediate action, including the removal of content and banning users.

Human moderation, expert teams

Although artificial intelligence and machine learning are effective tools, human supervision and review is still indispensable in the fight against child sexual abuse materisl. Companies employ dedicated teams of trained experts to scrutinise flagged content. These experts are able to deal with the nuanced and sensitive nature of the content, ensuring that no harmful, inappropriate content slips through the cracks.

Legal and political initiatives

Technology companies comply with the legal requirements for reporting to the authorities. Legislation and regulations require companies to report content depicting child sexual abuse, thereby ensuring that law enforcement agencies can take appropriate action. For example, Apple has strict guidelines to ensure that such legislation is observed and works closely with law enforcement agencies.

The large technology companies are also involved in initiatives aiming to protect interests and educate. They work together with organisations and governments to heighten awareness of child sexual abuse material and to draw the public’s attention to the protection of children online.

Technological innovations

Whereas end-to-end encryption is essential from the aspect of protecting the data of users, it poses challenges in detecting content depicting child sexual abuse. Tech giants, such as Facebook, are exploring ways to balance data protection with security, and are developing methods of scanning encrypted data without compromising the privacy of users. Among the innovations in this area are client-side scanning, which detects content depicting child sexual abuse before the data are encrypted and sent.

Constant battle and development

Combating child sexual abuse material is an ongoing and multi-faceted effort. The large technology companies are at the forefront of this battle, using advanced technologies, encouraging cooperation in the sector and enforcing rules in order to protect children online. Although considerable headway has been made, the constantly changing nature of digital threats requires continued innovation and vigilance. Through their continued commitment to these efforts, the tech giants are playing a vital role in creating an online environment that is safer for everyone.

27. 06. 2024