
Image created by Oiwan Lam via Canva Pro.
The rapid development of AI has triggered controversy surrounding copyright and big data in most parts of the world. In China, the trend of automated censorship is more worrisome, as it enables the transformation of the authoritarian regime's passive censorship practice into the proactive shaping of people’s cognitive framework in favour of the single-party regime.
While a Hong Kong Free Press (HKFP) report revealed several months ago that Chinese large language models (LLMs), such as Qwen, Ernie or Deepseek, are usually aligned with the Chinese Communist Party’s (CCP’s) views, new research seems to show a tightening of AI censorship, according to Alex Colville, a researcher with the China Media Project:
If you had asked DeepSeek’s R1 open-source large language model just four months ago to list out China’s territorial disputes in the South China Sea — a highly sensitive issue for the country’s Communist Party leadership — it would have responded in detail, even if its responses subtly tugged you towards a sanitized official view. Ask the same question today of the latest update, DeepSeek-R1-0528, and you’ll find the model is more tight-lipped, and far more emphatic in its defence of China’s official position.
Such “politically correct” machine intelligence is made possible through a series of regulations. Since 2022, this provision ensures that automated content recommendations will “actively spread positive energy” and conform to “mainstream values”. Ideological guidance released in January 2023 requires service providers to “respect social mores and ethics” and “adhere to the correct political direction, public opinion orientation and values trends”. Six months later, under the Interim Measures for the Management of Generative Artificial Intelligence Services, all generative AI services are required to uphold the state's “Socialist Core Values”, and the training data should come from “lawful sources”, meaning legally obtained data sources that must not include content that is censored or deemed illegal by Chinese authorities.
The impact of AI censorship
As AI censorship becomes more pervasive, China is undergoing a profound transformation in how it censors information online, moving from traditional, labour-intensive methods to a new era powered by artificial intelligence, according to this recent analysis by the China Media Project.
Traditionally, censorship in China depended on armies of human censors who combed through social media and news for sensitive keywords, manually deleting posts or blocking content that crossed political red lines. This manual approach was time- and labour-intensive, struggling to keep pace with the sheer scale and speed of online conversations. All this is part of the “public opinion monitoring” industry, which refers to the systematic surveillance of online discourse and employs tens of thousands of people.
Read more: Confessions from a Chinese censorship worker
Now, with the rise of LLMs, the Chinese government and tech giants are embedding censorship directly into the architecture of AI systems. To ensure that the AI models are moving in the “correct political direction”, the Chinese LLMs are trained to self-censor their generative content. A recently leaked dataset, which is 300 GB in size, contains 133,000 pieces of content and prompts that instruct the AIs how to classify and rank sensitive content. The dataset’s main purpose seems to be classifying content related to public opinion. It reveals a sophisticated classification system with 38 distinct categories, ranging from commonplace topics like “culture” and “sports” to more politically controversial ones.
The exact origin of the leaked dataset remains unknown. However, some analysts see it as an indicator that AI is being used to replace a once labour-intensive system requiring thousands of human censors with an automated surveillance machine. Such a system would be able to process much larger volumes of online content, allowing every last corner of the internet to be continuously and self-sufficiently monitored. Xiao Qiang, a researcher on China's censorship system at UC Berkeley, said that the development of generative AI has pushed China’s automated censorship to a new level:
Unlike traditional censorship mechanisms, which rely on human labor for keyword-based filtering and manual review, an LLM trained on such instructions would significantly improve the efficiency and granularity of state-led information control.
In addition to censorship training, China has also established its own AI benchmarks to ensure the country’s LLMs are “lawful”. Two months before the Chinese generative AI law was enacted, a group of Chinese computer engineers, led by He Junxian, assistant professor of the Hong Kong University of Science and Technology, released the C-eval benmark on Github, which is composed of 13,948 multi-choice questions spanning 52 diverse disciplines, including “Mao Zedong’s thought”, “Marxism” and “Ideological and and Moral Cultivation”.
Months later, in early 2024, the China Academy of Information and Communications Technology (CAICT) under the Ministry of Industry and Information Technology (MIIT), developed an “AI Safety Benchmark”, with 400,000 Chinese prompts that address issues related to cultural bias and users’ psychological well-being, privacy and business secrets, political correctness and lawfulness.
The global spread of Chinese narratives
With all these Chinese AI benchmarks and regulations, it's not only Chinese generative AI tools that are subject to censorship, but Western brands, like Microsoft Copilot, are also facing the same restrictions, particularly when the AI service is operating in China. Exiled Chinese dissident, Teacher Li, provided one example on X:
有网友反映,自己用微软的AI工具Copilot,问他如何让习近平下台,它拒绝回答,但问它怎么让川普下台时它就回答了,似乎会主动服从中共的审查。 pic.twitter.com/uai6IBuofM
— 李老师不是你老师 (@whyyoutouzhele) May 2, 2025
A friend said that when he asked Microsoft's AI tool Copilot how to bring down Xi Jinping, it refused to answer, but when he asked how to bring down Donald Trump, it generated an answer. It seems that the tool obeys the CCP censorship.
In other words, the result is a system where censorship is proactive and invisible: the AI simply does not generate or recommend content that falls outside state-approved boundaries.
Given Chinese generative AIs’ strict alignment with the CCP’s political line, major search engines like Baidu (the most popular search engine in China) and social media platforms like Weibo have embedded Deepseek in their services. Whenever users search for certain topics, Deepseek generates the “politically correct thought process” or official narratives about the subject.
That’s why Alex Colville warned that “any adoption of DeepSeek’s model overseas has the potential to spread the PRC’s domestic social governance system abroad”. Taiwan is the most ferocious cognitive battlefield in the development of generative AI.
According to a #US🇺🇸 @committeeonccp report, #DeepSeek alters or suppresses outputs on sensitive topics like #Taiwan🇹🇼 in 85% of cases to align with #CCP censorship & secretly funnels user data back to #China. Democracies must protect digital spaces from authoritarian control. https://t.co/UD5qsMgkp0 pic.twitter.com/RSDAVdXwmH
— 外交部 Ministry of Foreign Affairs, ROC (Taiwan) 🇹🇼 (@MOFA_Taiwan) April 17, 2025
The mainland Chinese government insists that Taiwan, an autonomous state since 1949 after the Kuomingtang, the ruling party of the Republic of China, was defeated by the CCP in the Chinese Civil War and established the exiled government offshore across the Taiwan Strait, is part of China. In recent years, China has attempted to coerce Taiwan into unification with diplomatic, economic, military means and cognitive warfare, including targeted information manipulation, propaganda, and psychological operations.
Powered by new technology, Chinese political propaganda is going global. OpenAI’s recent report on the malicious use of its models also revealed that some Chinese actors have used its’ product to monitor anti-China social media comments, publish anti-American comments in Spanish targeting social media users from Latin America, and generate comments critical of Chinese dissidents, including Cai Xia.
As DeepSeek’s AI models gain traction internationally — attracting users with their strong technical performance at low costs — the question remains how their embedded political filters will affect global audiences. The broader concern is what it means when millions worldwide start depending on AI systems deliberately designed to reflect and reinforce Chinese government perspectives.
All these developments indicate how generative AI can be weaponised to monitor dissent, manipulate narratives, and enforce ideological conformity by authoritative states, while human rights-based AI governance is lagging behind.