The biggest social media platforms will be required to protect children online by keeping them off suggested friend lists to stop them being contacted by groomers, Ofcom has said.
The new online safety regulator has published its first draft codes of practice under the Online Safety Act, which was signed into law last week.
The first codes focus on illegal material online – such as child sexual abuse material, grooming content and fraud.
Under the code, the largest platforms will be required, by default, to ensure that children on their sites are not presented with lists of suggested friends, do not appear in other users’ lists, that their location information is not visible to other users and that people outside their agreed connections cannot direct message them.
Ofcom is set to publish further codes in the coming months on other areas of online safety, such as guidance for adult sites on keeping children away and on protecting children from harmful content promoting things such as suicide or self-harm.
Each of the draft codes will have a consultation period before requiring final approval from Parliament.
The regulator’s own timetable says it hopes to begin enforcing its first codes of practice by the end of 2024.
The illegal content code also encourages larger sites to use hash matching technology to identify illegal images of sexual abuse and use automated tools to detect websites that have been identified as hosting abuse material.
On fighting fraud and terrorism, Ofcom says services should use automatic detection systems to find and remove posts linked to the sale of stolen financial information and block all accounts run by proscribed terrorist organisations.
The codes of practice also propose that tech firms nominate an accountable person who reports to senior management on compliance around illegal content, reporting and complaints duties, ensure their content moderation teams are well resourced and trained, offer easy reporting and blocking tools to use, and carry out safety tests on recommendation algorithms.
Dame Melanie said: “Regulation is here, and we’re wasting no time in setting out how we expect tech firms to protect people from illegal harm online, while upholding freedom of expression.
“Children have told us about the dangers they face, and we’re determined to create a safer life online for young people in particular.”
Writing in the Daily Telegraph, she said Ofcom “cannot waste a moment” in putting its powers to use, adding: “Children are our first priority, and the risk they face is real.
Technology Secretary Michelle Donelan said the publication of the first codes marked a “crucial” step in making the Online Safety Act a reality by “cleaning up the wild west of social media and making the UK the safest place in the world to be online”.
“Before the Bill became law, we worked with Ofcom to make sure they could act swiftly to tackle the most harmful illegal content first,” she said.
“By working with companies to set out how they can comply with these duties, the first of their kind anywhere in the world, the process of implementation starts today.”
Ofcom said it had been and would continue working with social media and other in scope platforms over the coming months to help ensure they were in compliance with the proposed codes when they come into full force.
Campaign groups have backed the first proposals from the regulator.
Susie Hargreaves, chief executive of the Internet Watch Foundation, said: “We stand ready to work with Ofcom, and with companies looking to do the right thing to comply with the new laws.
“It’s right that protecting children and ensuring the spread of child sexual abuse imagery is stopped is top of the agenda.
“It’s vital companies are proactive in assessing and understanding the potential risks on their platforms, and taking steps to make sure safety is designed in.
“Making the internet safer does not end with this Bill becoming an Act. The scale of child sexual abuse, and the harms children are exposed to online, have escalated in the years this legislation has been going through Parliament.
“Companies in scope of the regulations now have a huge opportunity to be part of a real step forward in terms of child safety.”
Why are you making commenting on The National only available to subscribers?
We know there are thousands of National readers who want to debate, argue and go back and forth in the comments section of our stories. We’ve got the most informed readers in Scotland, asking each other the big questions about the future of our country.
Unfortunately, though, these important debates are being spoiled by a vocal minority of trolls who aren’t really interested in the issues, try to derail the conversations, register under fake names, and post vile abuse.
So that’s why we’ve decided to make the ability to comment only available to our paying subscribers. That way, all the trolls who post abuse on our website will have to pay if they want to join the debate – and risk a permanent ban from the account that they subscribe with.
The conversation will go back to what it should be about – people who care passionately about the issues, but disagree constructively on what we should do about them. Let’s get that debate started!
Callum Baird, Editor of The National
Comments: Our rules
We want our comments to be a lively and valuable part of our community - a place where readers can debate and engage with the most important local issues. The ability to comment on our stories is a privilege, not a right, however, and that privilege may be withdrawn if it is abused or misused.
Please report any comments that break our rules.
Read the rules hereLast Updated:
Report this comment Cancel