Connect with us

Hi, what are you looking for?

AI Regulation

China Proposes New Rules on AI Chat Logs, Mandates User Consent for Data Use

China proposes new regulations mandating explicit user consent for AI chat log data use, enhancing user safety and potentially slowing AI chatbot innovation.

China is poised to tighten regulations on how artificial intelligence (AI) systems learn from user interactions, with a focus on enhancing user safety. The Cyberspace Administration of China announced on Saturday that it has drafted new measures to control how AI platforms can collect and utilize chat logs for training purposes. This initiative aims to ensure that “human-like” AI services, such as chatbots and virtual companions, remain secure for users.

The proposed regulations reflect China’s dual approach of fostering innovation in “human-like” interactive AI while instituting governance to prevent misuse and loss of control over sensitive data. Platforms would be required to notify users when they are interacting with AI and must offer options to access or delete their chat histories. Additionally, employing conversation data for model training or sharing it with third parties would necessitate explicit user consent, as outlined by the agency.

For minors, service providers would need to obtain additional consent from guardians before sharing conversation data. Guardians would also possess the right to request the deletion of a minor’s chat history. The draft measures are currently open for public consultation, with feedback due by late January.

Analysts say China is balancing AI safety with continued development

If finalized, these regulations could slow the advancement of AI chatbots. Lian Jye Su, chief analyst at Omdia, noted that restricting access to chat logs might limit “human-feedback mechanisms in reinforcement learning,” which play a crucial role in developing engaging and accurate conversational AI. Despite these concerns, he emphasized that China’s AI ecosystem remains robust, benefiting from extensive public and proprietary datasets.

Su pointed out that the move aligns with Beijing’s broader goals of national security and public interest. Tightening controls over chat logs indicates that some user conversations are deemed too sensitive to be classified as free training data. Wei Sun, principal analyst for AI at Counterpoint Research, added that these regulations serve more as “directional signals” than restrictions, emphasizing user protection and transparency in data practices rather than curtailing innovation.

Sun explained that the draft encourages AI providers to expand the use of human-like AI into more areas, such as cultural dissemination and companionship for the aging population, once safety and reliability are established. This can be seen as a policy nudge to accelerate the development of human-like AI interactions in a socially responsible and scalable manner.

Concerns surrounding AI chat log safety are increasingly relevant as the dialogue intensifies regarding how companies manage sensitive user conversations. Reports have emerged indicating that contract workers for tech giants like Meta can access user conversations with chatbots for quality assessment purposes. Some contractors disclosed that the material they reviewed contained personal details that could potentially identify individual users, including intimate exchanges resembling therapy sessions or private discussions with friends and partners.

In response to these concerns, a Meta spokesperson stated that the company employs “strict policies” governing who has access to personal data. While they collaborate with contractors to enhance training data quality, they intentionally limit the visibility of personal information and implement procedures instructing contractors on how to handle any sensitive information that may come to light.

A Google engineer specializing in AI security also cautioned against sharing sensitive information with chatbots, underscoring the need for users to safeguard their private data. “AI models use data to generate helpful responses, and we users must protect our private information so that harmful entities, like cybercriminals and data brokers, can’t access it,” he stated.

As China moves forward with these proposed regulations, the implications for innovation in AI technologies will be significant. Balancing user safety with the need for rapid advancements in AI capabilities will pose a critical challenge for both policymakers and industry leaders in the months to come.

See also
Staff
Written By

The AiPressa Staff team brings you comprehensive coverage of the artificial intelligence industry, including breaking news, research developments, business trends, and policy updates. Our mission is to keep you informed about the rapidly evolving world of AI technology.

You May Also Like

AI Cybersecurity

Schools leverage AI to enhance cybersecurity, but experts warn that AI-driven threats like advanced phishing and malware pose new risks.

AI Tools

Only 42% of employees globally are confident in computational thinking, with less than 20% demonstrating AI-ready skills, threatening productivity and innovation.

AI Research

Krites boosts curated response rates by 3.9x for large language models while maintaining latency, revolutionizing AI caching efficiency.

AI Marketing

HCLTech and Cisco unveil the AI-driven Fluid Contact Center, improving customer engagement and efficiency while addressing 96% of agents' complex interaction challenges.

Top Stories

Cohu, Inc. posts Q4 2025 sales rise to $122.23M but widens annual loss to $74.27M, highlighting risks amid semiconductor market volatility.

Top Stories

ValleyNXT Ventures launches the ₹400 crore Bharat Breakthrough Fund to accelerate seed-stage AI and defence startups with a unique VC-plus-accelerator model

AI Regulation

Clarkesworld halts new submissions amid a surge of AI-generated stories, prompting industry-wide adaptations as publishers face unprecedented content challenges.

AI Technology

Donald Thompson of Workplace Options emphasizes the critical role of psychological safety in AI integration, advocating for human-centered leadership to enhance organizational culture.

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.