Thursday , July 29 2021

Suicide tips hidden on YouTube and YouTube Kids videos alert pediatrician



An American pediatrician has warned parents about videos that show children how to kill YouTube and YouTube children.

Free Hess learned about cold videos in the summer when another mother noticed someone on YouTube Kids.

She said that minutes of the video from a children's video game, a man appeared on the screen – he was instructing how to commit suicide.

"I was shocked," said Ms. Hess, noting that since then, the scene has been united in several more videos from the popular Nintendo "Splatoon" on YouTube and YouTube Kids.

Mrs. Hess from Okala, Florida, publishes blogs about modified videos and works to bring them down against protests by parents and child health experts who say that such visual effects can be harmful to children.

One on YouTube shows a man who is in the frame. “Remember, kids,” he begins, holding what appears to be an imaginary blade to the inside of his arm. “Sideways for attention. Longways for results.”

“I think it's extremely dangerous for our kids,” Ms Hess said about the clips Sunday in a phone interview with the Washington Post,

“I think our kids are facing a whole new world with social media and internet access. It's changing the way they're growing, and it's changing the way they're developing. I think videos like this put them at risk.”

A recent YouTube video appears to include a spliced-in scene showing internet personality Filthy Frank.

It is unclear why he was edited into these clips, but his fans have been known to put him in memes and other videos. There is a similar video on his channel shot in front of a green screen, but the origin and context of the video in question are not clear.

Andrea Faville, a spokeswoman for YouTube, said in a written statement that the company works to ensure that it is “not used to encourage dangerous behaviour and we have strict policies that prohibit videos which promote self-harm.”

“We rely on both user flagging and smart detection technology to flag this content for our reviewers,” Ms Faville added.

“Every quarter we remove millions of videos and channels that violate our policies and we remove the majority of these videos before they have any views. We are always working to improve our systems and to remove violative content more quickly, which is why we report our progress in a quarterly report and give users a dashboard showing the status of videos they've flagged to us.”

Videos come amid growing questions about how YouTube, the world's largest video sharing platform, monitors and removes problematic content.

YouTube has long struggled with how to keep the platform free from such material – removing hateful and violent videos, banning dangerous pranks and cracking down on child sexual exploitation.

Last month, YouTube announced that it was rebuilding its recommendation algorithm to prevent it from prompting videos that include conspiracy theories and other bogus information, though the videos would remain on the site.

Ms Hess said she has been writing about the distressing video clips on her blog, PediMom, to raise awareness and to get them removed from the platform.

Earlier this month, she found a second one – this time on YouTube. She recorded it, wrote it for her and reported the content of the video sharing platform, she said. The video was downloaded.

Another version was reposted on 12 February, receiving more than 1,000 views before it, too, was removed from the site.

Ms Hess said the doctored “Splatoon” videos are not the only ones pushing dark and potentially dangerous content on social media platforms, particularly on YouTube Kids.

In a blog post last week, Ms Hess alerted other parents to numerous concerning videos she said she found on the app – a “Minecraft” video depicting a school shooting, a cartoon centred on human trafficking, one about a child who committed suicide by stabbing and another who attempted to commit suicide by hanging.

Nadine Kaslow, a past president of the American Psychological Association, told the Washington Post that it is a “tragic” situation in which “trolls are targeting kids and encouraging kids to kill themselves.”

Ms Kaslow, who teaches at Emory University's school of medicine, said that some children may ignore the grim video content but that others, particularly those who are more vulnerable, may be drawn to it.

She said that such videos can make children have nightmares, cause bad memories of loved ones who have committed suicide or even encouraged them to try it, although some of them may be too young to understand consequences.

Ms Kaslow said parents should monitor what their children do online and tech companies should ensure such content is removed. Still, she said, that is not enough.

“I don't think you can just take them down,” she said about the videos. “For children who have been exposed, they've been exposed. There needs to be messaging – this is why it's not OK.”

Support free-thinking journalism and subscribe to Independent Minds

Though parents should talk to their children about the videos, Ms Kaslow said, YouTube Kids also should address the issue, explaining to children what the videos were and why children should never harm themselves.

She added that there should be “serious consequences” for those who had a hand in the videos, noting that it was “very worrisome” that they were targeting children.

According to the Centers for Disease Control and Prevention, suicide risk factors may include mental disorders such as clinical depression, previous attempts at suicide, a barrier to access to mental health treatment, physical illness and feelings of hopelessness or isolation.

Washington Post


We will tell you what is true. You can create your own view.

at The Independentno one tells us what to write. Therefore, in the age of political lies and bias of Brexit, more readers turn to an independent source. Subscribe only from 15p per day for additional exclusions, events and eBooks – no ads.

Subscribe now


Source link