YouTube is overhauling its kid-focused video app to give parents the option of letting humans, not computer algorithms, select what shows their children can watch.
The updates rolling out today are a response to complaints that the YouTube Kids app has repeatedly failed to filter out disturbing content.
Google-owned YouTube launched the toddler-oriented app in 2015. It has described it as a "safer" experience than the regular YouTube video-sharing service for finding "Peppa Pig" episodes or watching user-generated videos of people unboxing toys, teaching guitar lessons or experimenting with science.
In order to meet U.S. child privacy rules, Google says it bans kids under 13 from using its core video service. But its official terms of agreement are largely ignored by tens of millions of children and their families who don't bother downloading the under-13 app.
Both the grown-up video service and the YouTube Kids app have been criticized by child advocates for their commercialism and for the failures of a screening system relying on artificial intelligence. The app is engineered to automatically exclude content not appropriate for children and recommend videos based on what children have watched before. It hasn't always worked to parents' liking -- especially when videos with profanity, violence or sexual themes slip through the filters.
The updates allow parents to switch off the automated system and choose a contained selection of children's programming such as "Sesame Street" and "PBS Kids." But the automated system remains the default.
"For parents who like the current version of YouTube Kids and want a wider selection of content, it's still available," said James Beser, the app's product director, in a blog post Wednesday. "While no system is perfect, we continue to fine-tune, rigorously test and improve our filters for this more-open version of our app."
Beser also encouraged parents to block videos and flag them for review if they don't think they should be on the app. But the practice of addressing problem videos after children have already been exposed to them has bothered child advocates who want the more controlled option to be the default.
"Anything that gives parents the ability to select programming that has been vetted in some fashion by people is an improvement, but I also think not every parent is going to do this," said Josh Golin, director of the Boston-based Campaign for a Commercial-Free Childhood.
"Giving parents more control doesn't absolve YouTube of the responsibility of keeping the bad content out of YouTube Kids."
He said Google should aim to build an even cleaner and safer kids' app, then pull all the kid-oriented content off the regular YouTube -- where most kids are going -- and onto the app.
Golin's group recently asked the Federal Trade Commission to investigate whether YouTube's data collection and advertising practices violate federal child privacy rules. He said advocates plan to meet with FTC officials next week.
Connect with the Southeast Missourian Newsroom:
For corrections to this story or other insights for the editor, click here. To submit a letter to the editor, click here. To learn about the Southeast Missourian’s AI Policy, click here.