CHILDREN as young as two are being exposed to explicit and disturbing content online, a new report from the ANU has found.
Experts are calling for a clear duty of care after research conducted with some of society’s most vulnerable children and youth found more than 70 per cent have seen content online that they found concerning, including violent and explicit content.
“Adults and the law are always 10 steps behind. Kids are telling us that they aren’t alright online, yet companies lack transparency and accountability,” said Dr Faith Gordon, a researcher from the ANU College of Law and the author of the report.
“Alarmingly, we found children as young as two or three years old have been exposed to really violent and sexually explicit content.
“Children also talked about experiencing unwanted contact, often from adults posing as children or being bombarded with scams.
“Often, young people have to navigate distressing pushed content.”
The report found only 40 per cent of children who experienced online harm reported it to the platforms they were using. Young people who complain typically report feeling “re-victimised” due to no response or automated replies and inaction from online platforms.
“Young people questioned why companies do not appear to be held to account. They want tech companies and the government to urgently address these issues and they want their voices and suggestions to be heard,” Dr Gordon said.
“There needs to be a clear duty of care and companies need to be much more transparent. This needs to be coupled with a legislative framework, which upholds and promotes the rights of children.”
The qualitative research, conducted in the UK during COVID-19 lockdowns, calls for children’s and young people’s voices to be at the centre of discussions on all policy reforms in this area.
Dr Gordon said there has never been a more crucial time to regulate our online spaces and educate wider society about acceptable use online, as the pandemic has caused a need for children to be on the internet.
“Children told us they want to be safe online and that’s up to governments, policymakers and the tech companies to genuinely listen and address their concerns in partnership with them,” she said.
“Children’s rights need to be considered in making reforms and that is what is missing in the frameworks that currently exist in Australia.
“Countries, such as Australia and the UK, who signed up to the UN Convention on Rights of the Child in 1990 are obligated to ensure that it is embedded into any frameworks moving forward. The latest UN General Comment on children’s rights in the digital environment needs to be implemented on the ground.”
Dr Gordon said although parts of Australia’s e-safety model are leading the way internationally, there are “grey areas” around legal but harmful content online and how rights-based approaches work in practice.
“In Australia the spotlight is now on new privacy legislation. Proposals will require social media companies, under law, to act in the best interests of children when accessing their data,” said Dr Gordon.
“This is a positive move, but as the study shows we need to see other agencies being held to account on this too, including the education tech sector and gaming platforms.
“We know from this report that children and young people can get around verification processes. There needs to be a more rigorous process for their age group, as well as for adults who should also be verifying their age.”
Who can be trusted?
In a world of spin and confusion, there’s never been a more important time to support independent journalism in Canberra.
If you trust our work online and want to enforce the power of independent voices, I invite you to make a small contribution.
Every dollar of support is invested back into our journalism to help keep citynews.com.au strong and free.
Ian Meikle, editor