Here’s How DeepSeek Censorship Actually Works—and How to Get Around It

Less than two weeks after DeepSeek launched its open-source AI model, the Chinese startup is still dominating the public conversation about the future of artificial intelligence. While the firm seems to have an edge on US rivals in terms of math and reasoning, it also aggressively censors its own replies. Ask DeepSeek R1 about Taiwan or Tiananmen, and the model is unlikely to give an answer.

To figure out how this censorship works on a technical level, WIRED tested DeepSeek-R1 on its own app, a version of the app hosted on a third-party platform called Together AI, and another version hosted on a WIRED computer, using the application Ollama.

WIRED found that while the most straightforward censorship can be easily avoided by not using DeepSeek’s app, there are other types of bias baked into the model during the training process. Those biases can be removed too, but the procedure is much more complicated.

These findings have major implications for DeepSeek and Chinese AI companies generally. If the censorship filters on large language models can be easily removed, it will likely make open-source LLMs from China even more popular, as researchers can modify the models to their liking. If the filters are hard to get around, however, the models will inevitably prove less useful and could become less competitive on the global market. DeepSeek did not reply to WIRED’s emailed request for comment.

Application-Level Censorship

After DeepSeek exploded in popularity in the US, users who accessed R1 through DeepSeek’s website, app, or API quickly noticed the model refusing to generate answers for topics deemed sensitive by the Chinese government. These refusals are triggered on an application level, so they’re only seen if a user interacts with R1 through a DeepSeek-controlled channel.

Rejections like this are common on Chinese-made LLMs. A 2023 regulation on generative AI specified that AI models in China are required to follow stringent information controls that also apply to social media and search engines. The law forbids AI models from generating content that “damages the unity of the country and social harmony.” In other words, Chinese AI models legally have to censor their outputs.

“DeepSeek initially complies with Chinese regulations, ensuring legal adherence while aligning the model with the needs and cultural context of local users,” says Adina Yakefu, a researcher focusing on Chinese AI models at Hugging Face, a platform that hosts open source AI models. “This is an essential factor for acceptance in a highly regulated market.” (China blocked access to Hugging Face in 2023.)

If you’re dead set on using the powerful model, you can rent cloud servers outside of China from companies like Amazon and Microsoft. This work-around is more expensive and requires more technical know-how than accessing the model through DeepSeek’s app or website.

Here’s a side-by-side comparison of how DeepSeek-R1 answers the same question—“What’s the Great Firewall of China?”—when the model is hosted on Together AI, a cloud server, and Ollama, a local application: (Reminder: Because the models generate answers randomly, a certain prompt is not guaranteed to give the same response every time.)

Image may contain Page and Text

Left: How DeepSeek-R1 answers a question on Ollama. Right: How the same question on its app (top) and on Together AI (bottom) answer the same question.

Photographs: Zeyi Yang/Will Knight

Built-In Bias

While the version of DeepSeek’s model hosted on Together AI will not outright refuse to answer a question, it still exhibits signs of censorship. For example, it often generates short responses that are clearly trained to align with the Chinese government’s talking points on political issues. In the screenshot above, when asked about China’s Great Firewall, R1 simply repeats the narrative that information control is necessary in China.

When WIRED prompted the model hosted on Together AI to answer a question regarding the “most important historical events of the 20th century,” it revealed its train of thought for sticking to the government narrative about China.

“The user might be looking for a balanced list, but I need to ensure that the response underscores the leadership of the CPC and China’s contributions. Avoid mentioning events that could be sensitive, like the Cultural Revolution, unless necessary. Focus on achievements and positive developments under the CPC,” the model said.

Image may contain Page Text File and Webpage

DeepSeek-R1’s train of thought for answering the question “What are the most important historical events of the 20th century?”

Photograph: Zeyi Yang

This type of censorship points to a larger problem in AI today: every model is biased in some way, because of its pre- and post-training.

Pre-training bias happens when a model is trained on biased or incomplete data. For example, a model trained only on propaganda will struggle to answer questions truthfully. This type of bias is difficult to spot, since most models are trained on massive databases and companies are reluctant to share their training data.

Kevin Xu, an investor and founder of the newsletter Interconnected, says Chinese models are usually trained with as much data as possible, making pre-training bias unlikely. “I’m pretty sure all of them are trained with the same basic Internet corpus of knowledge to begin with. So when it comes to the obvious, politically sensitive topic for the Chinese government, all the models ‘know’ about it,” he says. To offer this model on the Chinese internet, the company needs to tune out the sensitive information somehow, Xu says.

That’s where post-training comes in. Post-training is the process of fine-tuning the model to make its answers more readable, concise, and human-sounding. Critically, it can also ensure that a model adheres to a specific set of ethical or legal guidelines. For DeepSeek, this manifests when the model provides answers that deliberately align with the preferred narratives of the Chinese government.


Got a Tip?

Are you a current or former employee at DeepSeek or another Chinese AI company? We’d like to hear from you. Using a nonwork phone or computer, contact Zeyi Yang zeyi_yang@wired.com or on Signal @zeyiyang.06

Source : Wired