Safer social media means putting control in the hands of users
The news that the watchdog Ofcom will have increased powers over social media platforms to act over harmful content has been met with a mixture of excitement and concern in the UK. Users, advertisers and social media firms themselves are asking how these powers will impact the platforms, the content shared on them, and how we interact with them every day. Right now we have no answer to that question.
However, we can begin to think hard about what the impact of Ofcom’s regulatory powers over social media might be – be that on users, advertisers, or public safety.
Sharing the responsibility
Over the last few years, we have seen social media firms take significant steps to eliminate bad actors, manipulative political content, and hate speech from their platforms. From Twitter experimenting with new solutions to remove toxicity from the platform to Instagram and Facebook’s experiments to hide likes, the platforms have shown how serious they are about making their environments healthier and happier for their users.
Safer social media means putting control in the hands of users / image via Unsplash
Recently, Facebook stated that it has an army of ‘digital police’, made up of algorithms and AI, working alongside humans to create a safer and more transparent online environment. While this all sounds like a lot, it’s clearly not enough. That’s why Ofcom taking a more powerful role in keeping people safe online – determining what content is harmful and how it should be handled – can only be a positive step as part of a shared responsibility model.
Education over censorship
Many politicians and business owners have been asking how big the role of regulators should be in determining what content is harmful and what is merely controversial. While it’s hard to define clear cut lines when it comes to how harmful a piece of content is, it is possible to educate users and the public as a whole about behaviours on social media. While technology and regulation can help, it’s only by teaching people to use social media responsibly that we stand a chance of limiting harmful content in the online world for good.
The latest marketing news and insights straight to your inbox.
Get the best of The Drum by choosing from a series of great email briefings, whether that’s daily news, weekly recaps or deep dives into media or creativity.Sign up
The platforms themselves could be the place to start. Gently educating users, particularly younger users, about how to behave on social media is a step in the right direction. Putting their money where their mouth is and launching a global campaign on this could be a big push that the industry badly needs. But the onus isn’t on the platforms alone. Education initiatives do exist, for example, Safer Internet Day. A European initiative celebrated annually, Safer Internet Day aims to teach users about topics from cyberbullying to social media. Education systems, if they aren’t already, should also be including social media behaviour in their curriculum – and governments should be encouraging this. In short, collaboration between the platforms, the educators and the governments is the right way forward.
Don’t forget the past. Learn from it
Unlike with media such as radio, TV and print, the attempts made to regulate the early days of the internet didn’t meet with much success. Even today innocent searches on the internet can expose users to content that makes them feel upset and confused, or has even worse effects.
Given the scale and open nature of the internet, cracking down on harmful content is an uphill battle for regulators and for users. However, where we have seen some success is in the implementation of greater user controls. By giving users control over the content they see, whether, through ad blockers, parental controls or URL filtering, the internet has become a safer, healthier place for users.
The same could easily be true for social media. If users were given more control over the content they and their children can see, the social world would undoubtedly feel like a safe place to inhabit. This shouldn’t negatively impact brands – rather, it would encourage brands to be more careful and inclusive with the content they share. Harmful stereotyping could be one example of creative advertising that this level of user control could stamp out. This can only be a good thing.
So what’s next?
Yih-Choung Teh, group director of strategy and research at Ofcom, said that for most people the risks of social media "are still outweighed by the huge benefits of the internet. And while most internet users favour tighter rules in some areas, particularly social media, people also recognise the importance of protecting free speech – which is one of the internet’s great strengths."
The move to give Ofcom greater power has the potential to see governments and platforms working hand-in-hand to remove harmful content and toxicity from the social sphere. We work with many of the largest brands in the world, and we know they get value from social media for reaching and engaging with their audiences. It has a positive impact on their business in countless ways, but no brand wants this to come at a cost – of their brand reputation, customer loyalty, or worse. Brands want to be sure that they are investing their ad budgets into safe and trustworthy platforms, which are free from harm and toxicity.
While we shouldn’t expect anything to change overnight, this move by Ofcom is a step in the right direction for both users and advertisers. Anything that makes social media platforms safer and more engaging is a win-win, both for the people using them and for businesses advertising on them.
Yuval Ben-Itzhak is chief executive officer of Socialbakers.