• Keep kids safe in AI 2/2

    From Mike Powell@1:2320/105 to All on Tuesday, December 30, 2025 09:42:02
    (continued)

    What needs to happen?

    Ideally, protecting children would involve parents, schools, governments, and tech companies all working together. But after years of slow progress on
    social media regulation, its not hard to see why confidence in that happening any time soon is low.

    Many of the biggest problems could be addressed if the companies behind AI tools and social platforms took more responsibility and enforced meaningful safeguards. Tech companies need to be subject to urgent, meaningful
    regulation if were going to protect children, Steele says. At the moment, far too much responsibility is falling on families, schools, and the goodwill of industry, and that simply isnt safe.

    Bartuski agrees that companies should be doing far more. They have the money, resources, and visibility to be able to do a lot more. Many social media companies have used Foggs Persuasive Design to get kids habituated to be lifelong users of their platforms. Tech companies do this on purpose, she explains.

    But this is where the tension lies. We can say tech companies should do more, yet as the risks become clearer, corporate incentives are often moving in the opposite direction. With the guardrails being removed from AI development (specifically in the US), there are some (not all) companies that are using that to their advantage, Bartuski says. She has already seen companies push ahead with features they know are dangerous.

    Even so, experts agree that certain steps would have an immediate and significant impact. There need to be clear rules on what AI systems must not
    be allowed to do, including creating sexualized images of children, promoting self-harm, or using design features that foster emotional dependency, Steele says.

    This forms the basis of the Safe AI for Children Alliances Non-Negotiables Campaign , which outlines three protections every child should have.
    Alongside banning the creation of sexualized images of children, the campaign states that AI must never be designed to make children emotionally dependent and AI must never encourage children to harm themselves.

    But relying on tech companies alone wont cut it. Independent oversight is essential. This is why Briercliffe believes stronger external checks are
    needed across the industry. There must be mandatory, independent, third-party testing and evaluation before deployment, he says. We also need independent oversight, transparency about how systems behave in real-world conditions,
    and real consequences when companies fail to protect children.

    And ultimately, this goes beyond individual platforms. This is ultimately a question of societal responsibility, Tara says. We must set strong,
    enforceable standards that ensure childrens safety comes before commercial incentives.

    What can parents do?

    Even with regulations slow to catch up, parents shouldnt feel at a loss.
    There are meaningful steps you can take right now. Its completely understandable for parents to feel worried, Steele says. The technology is moving very fast, and the risks arent intuitive. But it is important not to feel powerless.

    1. Understand the basics

    Parents dont need to learn how every AI tool works, Bartuski says. But
    getting clear on the risks and benefits is important. Steele offers a free Parent and Educator Guide at safeaiforchildren.org that lays out all the
    major concerns in clear, accessible language, which is a good place to start.

    2. Create open, non-judgmental communication

    If kids feel judged or are worried about consequences, they are not going to turn to parents when something is wrong, Bartuski says. If they dont feel
    safe talking to you, you are placing them in potentially dangerous and/or exploitative situations. Keep conversations calm, curious, and shame-free.

    3. Talk about the tech

    You might assume your children understand AI better than you do because they use it more. But they may not grasp how it works, how often it gets things wrong, or that fake content can look real. Bartuski says kids need to know
    that chatbots can be wrong, manipulative, or unsafe, even when they sound caring or convincing.

    4. Use shared spaces

    This isnt about banning tech outright. Its about making it safer. Steele suggests enforcing shared spaces", which involves using AI tools in communal areas, experimenting together, and avoiding private one-on-one use behind closed doors. This could reduce the chance of harmful interactions going unnoticed.

    5. Extend the conversation beyond the home

    Safety shouldnt stop at your front door. If you are worried, ask your child's school what they have in place, Briercliffe says. Even ask your employer to bring in a professional to give a talk. Experts agreed that while parents
    play a key role here, this is a wider cultural challenge, and the more openly we all discuss it, the safer children will be.

    6. Find more balance and reduce screen time

    Weve been talking about limiting screen time for years, and its just as important now that AI is showing up across apps, games, and social platforms. Kids need to be taught balance, Bartuski says. Play is essential for growth
    and development. She also stresses that reducing screen time only works if
    its replaced with activities that are engaging, fun, and cognitively challenging.

    ======================================================================
    Link to news story: https://www.techradar.com/ai-platforms-assistants/how-you-can-keep-your-kids-s afe-in-this-ai-powered-world

    $$
    --- SBBSecho 3.28-Linux
    * Origin: Capitol City Online (1:2320/105)