Skip to main content
Blog

Navigating the Global Regulatory Landscape: A Comprehensive Analysis of Social Media Legislation

By February 27, 2024No Comments

This week, most eyes are on the Supreme Court in the US as they begin hearing arguments and discuss the level of content moderation that should be required by platforms. Cath Virginia of The Verge has written a great article that gives you a complete overview of what is expected this week. 

Looking at the overall legal landscape, we wanted to kick off a series of blog posts around the legislation that is out there regarding content moderation and the reporting requirements required as part of the law. Many of these laws are new, and therefore 2024 will see them being interpreted for the first time, which will be very interesting.  

We are all aware that in recent years the digital landscape has undergone a seismic shift, with social media platforms becoming integral to our daily lives. As the influence of these platforms has grown, so too has the concern for the safety and wellbeing of their users. Governments around the world are responding to these challenges by proposing and enacting legislation aimed at regulating social media to ensure a safer online environment. This article provides an in-depth review of major legislations, exploring the Digital Services Act in the EU, the Online Safety Bill in the UK, and relevant legislation in the US and Australia. 

Digital Services Act in the EU 

The European Union’s Digital Services Act is aimed at creating a safer digital space for users within the EU. It introduces a dual-tier system, categorizing platforms as either very large online platforms (VLOPs) or online intermediaries. VLOPs are subject to more stringent obligations, including increased content moderation measures and transparency reporting. Non-compliance with the Digital Services Act can result in fines of up to 6% of a company’s annual global turnover. This highlights the EU’s commitment to ensuring robust enforcement of online safety regulations. 

Key Provisions 

  • Dual-Tier System – Platforms are categorized based on size, with very large online platforms (45 million users in the EU) facing more stringent obligations. This includes measures to combat the spread of illegal content and ensure user safety. 
  • Content Moderation Standards – The act outlines standards for content moderation, requiring platforms to employ measures to prevent the dissemination of specific types of content, including terrorist content, child sexual abuse material, and hate speech. 
  • Transparency and Accountability – Platforms are mandated to provide transparency reports detailing their content moderation actions. Authorities are empowered to assess the effectiveness of content moderation measures.  
  • Based on EU and national laws – Illegal content will be based on the national laws in the country and the overarching EU law. For example, removing photos of Nazi symbols will be required in Germany, but would not be required in Denmark. 

Impact on Content Moderation 

  • Stricter Requirements for Larger Platforms – Very large platforms face more comprehensive obligations, necessitating advanced content moderation technologies to meet the prescribed standards. 
  • Collaboration with Authorities – Platforms are required to cooperate with national authorities to address illegal content. This collaboration enhances the effectiveness of content moderation efforts by leveraging governmental support. 
  • Financial Penalties – Fines of up to 6% of global turnover serve as a substantial deterrent for non-compliance. The financial impact underscores the EU’s commitment to robust enforcement. 
  • Operating Ban – In exceptional cases of serious harm, the legislation allows for a temporary ban on operations, where platforms could be shut down. The Commission will have enforcement powers similar to those it has under anti-trust proceedings. An EU-wide cooperation mechanism will be established between national regulators and the Commission. 

Already, we have seen the EU open formal probe of TikTok under the DSA, citing child safety, risk management and other concerns.  

Online Safety Bill in the UK 

The Online Safety Bill in the UK represents a comprehensive effort to tackle online harm and protect users. It requires social media companies to take responsibility for the content on their platforms, particularly content that may cause harm. The bill proposes the establishment of a regulatory framework with the power to impose hefty fines on platforms that fail to adhere to prescribed standards. Companies can face fines of up to 10% of their global turnover for non-compliance, providing a significant financial incentive for adherence to the regulations. 

Key Provisions 

  • Duty of Care – The bill establishes a legal duty of care on social media companies to protect users from harmful content. Platforms are required to take proactive measures to ensure user safety. 
  • Regulatory Oversight – The legislation proposes the creation of an independent regulatory body with the authority to enforce compliance. This body will set codes of practice outlining the expectations for online safety. 
  • Harmful Content Definition – The bill defines harmful content broadly, encompassing not only illegal content but also content that may be legal but harmful, such as cyberbullying, hate speech, and misinformation. 

Impact on Content Moderation: 

  • Increased Accountability – Social media companies are held accountable for the content on their platforms. The duty of care places the onus on platforms to implement robust content moderation policies to detect and remove harmful content promptly. 
  • Transparency Reporting – Platforms are required to publish transparency reports outlining the actions taken to address harmful content. This fosters transparency and allows users to gauge the effectiveness of content moderation efforts. 
  • Fines for Non-Compliance – Hefty fines, up to 10% of global turnover, act as a strong deterrent for non-compliance. This financial penalty incentivizes companies to invest in advanced content moderation technologies and practices. Senior management could also face jail time if the non-compliance is egregious enough.  

US Legislation 

In the United States, the regulatory landscape for social media is more nuanced, with various state-level initiatives and ongoing discussions at the federal level. The proposed SAFE TECH Act, for instance, seeks to reform Section 230 of the Communications Decency Act, making platforms more accountable for harmful content. Additionally, individual states such as California have introduced the Silenced No More Act, allowing users to sue platforms for content moderation decisions. While federal legislation is still pending, the trend towards increased scrutiny of social media platforms is evident. 

Key Provisions 

  • SAFE TECH Act – This proposed legislation seeks to reform Section 230, making platforms more accountable for content moderation decisions. It removes immunity in cases involving illegal content and ensures platforms act in good faith when moderating content. 
  • Silenced No More Act (California) – Allows users to sue platforms for content moderation decisions, promoting transparency and accountability. 
  • Regional Nuances – Since 2021, 38 states have introduced over 250 bills to regulate content across digital services’ platforms. Many of which are unconstitutional, conflict with federal law and would place major barriers on platform’s abilities to restrict dangerous content.  
  • New York – SB 9465 establishes a task force on social media and violent extremism, and AB 7865/SB 4511 requires social media networks to provide and maintain mechanisms for reporting hateful conduct on their platform(s). 
  • Minnesota – SF 3933/HF 3724 regulate algorithms that target user-generated content for users under 18. 
  • Ohio – HB441 is aimed at prohibiting platforms from censoring a user, their expression, or a user’s ability to receive the expression of another user.  

Impact on Content Moderation: 

  • Section 230 Reform – If enacted, the SAFE TECH Act would prompt platforms to reevaluate their content moderation practices aligning with the proposed changes to Section 230, potentially leading to more cautious moderation policies. 
  • Increased Accountability in California – The Silenced No More Act introduces a legal avenue for users to challenge content moderation decisions, fostering increased accountability and transparency. 

Australian Legislation 

Australia has taken steps to address online safety concerns through the Online Safety Act 2021. This legislation grants the eSafety Commissioner the authority to issue removal notices for harmful online content and imposes fines for non-compliance. Companies failing to remove specified content within the designated timeframe may face financial penalties. The legislation emphasizes the need for swift and effective content moderation to protect Australian users. 

Key Provisions: 

  • Online Safety Act 2021 – Empowers the eSafety Commissioner to issue removal notices for harmful content. Failure to comply with removal notices can result in fines. 
  • Definitions of Harmful Content – The legislation specifies types of harmful content, including cyberbullying, image-based abuse, and other online harms. 

Impact on Content Moderation: 

  • Swift Removal Requirements – The legislation mandates the prompt removal of specified harmful content, necessitating robust and efficient content moderation processes. 
  • Financial Penalties – Fines of up to AUD 555,000 per day for individuals and AUD 2.77 million per day for corporations provide a significant financial incentive for platforms to prioritize effective content moderation. 

More to Come 

The global surge in legislation targeting social media platforms underscores the growing recognition of the impact these platforms have on society. The diverse approaches taken by the UK, the EU, the US, and Australian authorities reflect the unique challenges each region faces.  

The imposition of fines and prison time for non-compliance is a common thread, signaling a collective determination to ensure that social media companies prioritize the safety and wellbeing of their users. As these legislative frameworks continue to evolve, the impact on companies – both in terms of financial penalties and potential legal consequences – serves as a powerful incentive for the industry to proactively address the challenges posed by harmful online content. 

These legislative acts collectively emphasize the importance of user safety and place substantial responsibility on social media platforms to implement effective content moderation measures. The combination of duty of care, regulatory oversight, transparency reporting, and financial penalties underscores a global commitment to creating a safer online environment.  

The Impact for Content Moderation Wellbeing 

With regulators focusing on the safety and wellbeing of users and providing distinct requirements around corporate social accountability and legal liability of the platforms, we believe that they must also consider the impact that these regulations will have on Trust & Safety teams and Content Moderator wellbeing.  

As a next step after launching and enforcing new legislation, regulators will need to ensure platforms are adequately resourced to support their Trust & Safety teams to adhere and comply with their regulatory requirements, while balancing the need to protect the psychological health and wellbeing of their employees.  

At Zevo, we are working with our customers and regulators to design and build the absolute best clinical solution for Content Moderators and Trust & Safety teams, to ensure their protection, and the reduction of risk for the organization.  

If you are interested in supporting your Content Moderators’ wellbeing, get in touch with our experts today.