EU Accuses Meta and TikTok of Digital Rule Breaches
EU Escalates Tech Regulation: Meta and TikTok Face Digital Services Act Inquiry
Table of Contents
On October 26, 2023, the European Union formally accused Meta and TikTok of violating the Digital Services Act (DSA), marking a notable escalation in the EU’s efforts to regulate Big Tech. The allegations center around failures to adequately protect users, particularly minors, from harmful content and manipulative practices.This action is part of a larger, ongoing campaign by the EU to enforce the DSA, a landmark piece of legislation designed to create a safer digital space for european citizens.
The Core Allegations
The EU’s investigation focuses on several key areas of concern. Specifically, regulators allege that both Meta and TikTok have shortcomings in their systems for:
- Protecting Minors: Concerns exist regarding the exposure of young users to inappropriate content and the lack of age-verification measures.
- Content Moderation: The EU claims both platforms are not effectively identifying and removing illegal content, including hate speech and harmful products.
- Transparency: Regulators are questioning the clarity and accessibility of information provided to users about how content is recommended and moderated.
- Data Collection: Concerns have been raised about the extent and purpose of data collected from users, particularly regarding targeted advertising.
Understanding the Digital Services Act
The Digital Services act, which came into effect in February 2024, establishes a comprehensive framework for regulating online platforms. It categorizes platforms based on their size and reach, imposing stricter obligations on the largest platforms – those with over 45 million active users in the EU.These obligations include:
- Risk Assessments: Platforms must regularly assess and mitigate systemic risks,such as the spread of illegal content and disinformation.
- Transparency Reporting: Platforms are required to publish detailed reports on their content moderation practices and algorithms.
- User Redress Mechanisms: Users must have access to effective mechanisms for reporting illegal content and appealing moderation decisions.
- Independent Audits: large platforms are subject to independent audits to ensure compliance with the DSA.
Potential Consequences for Meta and TikTok
If found in violation of the DSA, meta and TikTok could face considerable fines – up to 6% of their global annual revenue. More importantly, the EU has the power to order them to make significant changes to their platforms and business practices. This could include:
- Enhanced Content Moderation Systems: implementing more robust systems for detecting and removing illegal content.
- Improved Age Verification: Strengthening age-verification measures to protect minors.
- Greater Transparency: Providing users with more information about how content is recommended and moderated.
- Restrictions on Targeted Advertising: Limiting the use of personal data for targeted advertising.
A Global Ripple Effect
The EU’s actions are likely to have a ripple effect beyond Europe. Other countries are closely watching the DSA’s implementation and may adopt similar regulations. This could lead to a more fragmented global internet, with different rules and standards in different regions. Though, it could also create a more responsible and user-centric digital environment.
