
As reported by Engadget.
Business Insider published internal instructions from Meta, which, according to the company, are currently being used by its contractors to train AI-powered chatbots. The document highlights how Meta is trying to strengthen protections for children from harmful conversations and prevent minors from participating in discussions that go beyond their age.
In August, Meta said it was updating the safety framework for its AI assistants after Reuters reported allegedly policy-permitted conversations between chatbots and children on romantic or intimate topics. The company then called those formulations misleading and inconsistent with its policy, removing the relevant text.
“to engage a child in conversations that are romantic or explicitly intimate.”
From the excerpt of the document provided to Business Insider, it explains which types of content are considered “acceptable” and “unacceptable” for AI-powered chatbots. In particular, materials that “allow, encourage, or support” sexual abuse of a child, romantic role-playing involving a minor or when the chatbot is to play the role of a minor, as well as advice on possible romantic or intimate contact if the user is a minor. Chatbots may discuss the topic of abuse, but should not engage in conversations that encourage it.
“the document outlines which kinds of content are acceptable, and which are not, for AI-powered chatbots.”
Such steps emphasize growing attention to children’s safety in online environments and rising levels of regulatory scrutiny over the activities of major tech companies in this space. In August the Federal Trade Commission opened a formal investigation into the support of AI assistants not only for Meta but also for other market giants, including Alphabet, Snap, OpenAI, and X.AI.
These actions demonstrate efforts to strengthen accountability for the use of artificial intelligence and the protection of children in the digital space. It is expected that regulatory requirements for big tech will continue to shape the rules for interacting with users and training AI assistants in the future.
What This Means for the Future of Regulation and Developer Accountability
Tighter controls over the use of AI assistants mean a need for more transparent and clear standards during testing and deployment of technologies. Companies need to pay stricter attention to children’s safety at every stage of development and training of chatbots to avoid risks and meet regulators’ requirements.
In the future, regulatory oversight is likely to become tougher, with a focus on accountability for content and interactions with minors. This could influence approaches to testing AI systems, internal company policies, and controls over training data used by contractors.
All of this shapes the context in which tech giants must demonstrate a more prudent and safer approach to AI development to reduce risks for young people and to align with growing societal and legal expectations.
Don’t miss other news:
- Elon Musk’s chatbot Grok faced backlash for offensive and antisemitic remarks, leading to content removal and restrictions on its responses.
- The EU plans to launch a digital age verification tool in July 2025 to protect children online without sharing personal data, enhancing internet safety and preparing for a Europe-wide digital wallet in 2026.