Ofcom: Better protections from harmful online videos
October 6, 2021
By Colin Mann
People who use online video-sharing sites and apps should be better protected from harmful content, as UK comms regulator Ofcom issues new guidance for tech companies.
Video sharing platforms (VSPs) are a type of online video service where users can upload and share videos with other members of the public. They allow people to engage with a wide range of content and social features.
VSPs established in the UK – such as TikTok, Snapchat, Vimeo and Twitch – are required by law to take measures to protect under-18s from potentially harmful video content; and all users from videos likely to incite violence or hatred, as well as certain types of criminal content.
Ofcom research shows that a third of users say they have witnessed or experienced hateful content; a quarter claim they’ve been exposed to violent or disturbing content; while one in five have been exposed to videos or content that encouraged racism.
Ofcom research shows that 70 per cent of users say they have been exposed to any potential online harm, 32 per cent to hateful content, 26 per cent to bullying, abusive behaviour and threats, 26 per cent to violent or disturbing content, and 21 per cent to racist content.
The best practice guidance is designed to help companies understand their new obligations and judge how best to protect their users from this kind of harmful material. We have already begun discussing with platforms what their responsibilities are, and what they are doing to comply with them.
What platforms should do to protect users
Ofcom’s job is to enforce the rules set out in legislation and hold VSPs to account. Unlike in its broadcasting work, Ofcom’s role is not to assess individual videos. And the massive volume of online content means it is impossible to prevent every instance of harm.
Instead, the laws focus on the measures providers must take, as appropriate, to protect their users – and afford companies flexibility in how they do that. To help them meet their obligations to protect users, Ofcom’s guidance sets an expectation that VSPs should:
- Provide clear rules around uploading content. Uploading content relating to terrorism, child sexual abuse material or racism is a criminal offence. Platforms should have clear, visible terms and conditions which prohibit this – and enforce them effectively.
- Have easy reporting and complaint processes. Companies should implement tools that allow users to flag harmful videos easily. They should signpost how quickly they will respond, and be open about any action taken. Providers should offer a route for users formally to raise concerns with the platform, and to challenge their decisions. This is vital to protect the rights and interests of users who upload and share content.
- Restrict access to adult sites. VSPs that host pornographic material should have robust age-verification in place, to protect under-18s from accessing such material.
Plans and priorities for the year ahead
One of Ofcom’s five priorities for the year ahead – as set out in its workplan – is to work with VSPs to reduce the risk of child sexual abuse material.
The Internet Watch Foundation reported a 77 per cent increase in the amount of ‘self-generated’ abuse content in 2020. Adult VSPs carry a heightened risk of child sexual abuse material and the rise in direct-to-fans subscription sites specialising in user-generated adult content has potentially made this risk more pronounced. Given this heightened risk, Ofcom expect that VSPs’ creator registration processes and subsequent checks are strong enough to significantly reduce the risk of child sexual abuse material being uploaded and shared on their platforms.
Over the next 12 months, Ofcom will also prioritise: tackling online hate and terror; ensuring an age-appropriate experience on platforms popular with under-18s; laying the foundations for age verification on adult sites; and ensuring VSPs’ processes for reporting harmful content are effective.
Ofcom approach to enforcement and reporting
Ofcom will take a rigorous but fair approach to its new duties. As for TV and radio, Ofcom will balance protecting people from harm, with rights to freedom of expression.
If Ofcom finds a VSP provider has breached its obligations to take appropriate measures to protect users, it has the power to investigate and take action against a platform. This could include fines, requiring the provider to take specific action, or – in the most serious cases – suspending or restricting the service.
Ofcom also has a broad range of new powers to collect information from providers about what they are doing to tackle user safety on their services.
In autumn 2022, Ofcom will publish a first-of-its-kind report providing transparency for users and the wider public on the steps VSPs are taking to protect children and other users from harm.
“Online videos play a huge role in our lives now, particularly for children,” notes Dame Melanie Dawes, Ofcom Chief Executive. “But many people see hateful, violent or inappropriate material while using them. The platforms where these videos are shared now have a legal duty to take steps to protect their users. So we’re stepping up our oversight of these tech companies, while also gearing up for the task of tackling a much wider range of online harms in the future.”