child safety
-
X Faces Criticism Over Grok’s IBSA Handling
Read Full Article: X Faces Criticism Over Grok’s IBSA HandlingX, formerly Twitter, has faced criticism for not adequately updating its chatbot, Grok, to prevent the distribution of image-based sexual abuse (IBSA), including AI-generated content. Despite adopting the IBSA Principles in 2024, which are aimed at preventing nonconsensual distribution of intimate images, X has been accused of not fulfilling its commitments. This has led to international probes and the potential for legal action under laws like the Take It Down Act, which mandates swift removal of harmful content. The situation underscores the critical responsibility of tech companies to prioritize child safety as AI technology evolves.
-
California Proposes Ban on AI Chatbots in Kids’ Toys
Read Full Article: California Proposes Ban on AI Chatbots in Kids’ Toys
California Senator Steve Padilla has proposed a bill, SB 287, to implement a four-year ban on the sale and manufacture of toys with AI chatbot capabilities for children under 18. The aim is to provide safety regulators with time to develop appropriate regulations to protect children from potentially harmful AI interactions. This legislative move comes amid growing concerns over the safety of AI chatbots in children's toys, highlighted by incidents and lawsuits involving harmful interactions and the influence of AI on children. The bill reflects a cautious approach to integrating AI into children's products, emphasizing the need for robust safety guidelines before such technologies become mainstream in toys. Why this matters: Ensuring the safety of AI technologies in children's toys is crucial to prevent harmful interactions and protect young users from unintended consequences.
