In an ongoing effort to tackle the issue of deepfake child abuse material and “pro-terror content,” Australia’s internet watchdog, the eSafety Commissioner, is calling on technology giants to implement stricter regulations. The eSafety Commissioner recently released a set of standards that would require companies like Meta, Apple, and Google to take more proactive measures in addressing seriously harmful content, including synthetic child sexual abuse material created using artificial intelligence.
After giving the technology industry a two-year window to develop their own codes, the eSafety Commissioner concluded that the existing efforts had “failed to provide sufficient safeguards.” According to the regulatory body, the industry codes lacked a strong commitment to identify and remove known child sexual abuse material. As a result, the eSafety Commissioner has taken matters into its own hands and introduced these new industry-wide standards, which are now open for consultation and will require parliamentary approval.
If approved, the new standards would have a significant impact on technology giants such as Meta, Apple, and Google. The regulations would apply to websites, photo storage services, and messaging apps, holding these companies accountable for the worst-of-the-worst online content, including child sexual abuse material and pro-terror content. The eSafety Commissioner aims to ensure that the industry takes meaningful steps in preventing the proliferation of seriously harmful material, particularly content that exploits children.
Australia’s attempts to hold tech giants accountable for user-generated content have encountered obstacles in the past. In 2021, the country pioneered the “Online Safety Act,” which aimed to establish global standards for tech company responsibility. However, enforcing these sweeping powers has proven to be a challenge. Just recently, the eSafety Commissioner fined Elon Musk’s X platform Aus$610,500 (US$388,000) for its failure to demonstrate adequate measures in removing child sexual abuse content. Despite the fine, X has not paid it and has initiated legal action to contest it.
The eSafety Commissioner’s actions reflect Australia’s commitment to creating a safer online environment. By implementing stricter regulations and holding tech giants accountable, the country hopes to address the urgent and growing concern of deepfake child abuse material and pro-terror content. This move is especially significant given the rise of artificial intelligence, which has made it easier to manipulate digital content. The eSafety Commissioner’s standards serve as a crucial step in curbing the spread and impact of harmful content that exploits vulnerable individuals, particularly children.
While the onus falls on technology companies to implement safeguards and take proactive measures, it is also imperative that individuals, communities, and governments play their part in combatting harmful content. Raising awareness about the consequences of sharing or engaging with such material is crucial, as is fostering a culture of reporting and supporting victims. By working together, we can create an online space that prioritizes safety, respects privacy, and protects the most vulnerable members of society.
Australia’s eSafety Commissioner’s push for accountability marks an important milestone in combatting deepfake child abuse material and pro-terror content. By introducing stricter regulations that require tech giants to actively address harmful online content, the country aims to protect individuals, especially children, from exploitation and harm. While challenges in enforcement persist, it is crucial that both industry players and individuals alike recognize their collective responsibility in creating a safer online environment. Only through collective efforts can we ensure the well-being and security of all users in the digital realm.