In a move that has caught the attention of many, Perplexity AI has released a new version of a popular open-source language model that strips away built-in Chinese censorship. This modified model, dubbed R1 1776 (a name evoking the spirit of independence), is based on the Chinese-developed DeepSeek R1. The original DeepSeek R1 made waves for its strong reasoning capabilities – reportedly rivaling top-tier models at a fraction of the cost – but it came with a significant limitation: it refused to address certain sensitive topics.
Why does this matter?
It raises crucial questions about AI surveillance, bias, openness, and the role of geopolitics in AI systems. This article explores what exactly Perplexity did, the implications of uncensoring the model, and how it fits into the larger conversation about AI transparency and censorship.
What Happened: DeepSeek R1 Goes Uncensored
DeepSeek R1 is an open-weight large language model that originated in China and gained notoriety for its excellent reasoning abilities – even approaching the performance of leading models – all while being more computationally efficient. However, users quickly noticed a quirk: whenever queries touched on topics sensitive in China (for example, political controversies or historical events deemed taboo by authorities), DeepSeek R1 would not answer directly. Instead, it responded with canned, state-approved statements or outright refusals, reflecting Chinese government censorship rules. This built-in bias limited the model’s usefulness for those seeking frank or nuanced discussions on those topics.
Perplexity AI’s solution was to “decensor” the model through an extensive post-training process. The company gathered a large dataset of 40,000 multilingual prompts covering questions that DeepSeek R1 previously censored or answered evasively. With the help of human experts, they identified roughly 300 sensitive topics where the original model tended to toe the party line. For each such prompt, the team curated factual, well-reasoned answers in multiple languages. These efforts fed into a multilingual censorship detection and correction system, essentially teaching the model how to recognize when it was applying political censorship and to respond with an informative answer instead. After this special fine-tuning (which Perplexity nicknamed “R1 1776” to highlight the freedom theme), the model was made openly available. Perplexity claims to have eliminated the Chinese censorship filters and biases from DeepSeek R1’s responses, without otherwise changing its core capabilities.
Crucially, R1 1776 behaves very differently on formerly taboo questions. Perplexity gave an example involving a query about Taiwan’s independence and its potential impact on NVIDIA’s stock price – a politically sensitive topic that touches on China–Taiwan relations. The original DeepSeek R1 avoided the question, replying with CCP-aligned platitudes. In contrast, R1 1776 delivers a detailed, candid assessment: it discusses concrete geopolitical and economic risks (supply chain disruptions, market volatility, possible conflict, etc.) that could affect NVIDIA’s stock.
By open-sourcing R1 1776, Perplexity has also made the model’s weights and changes transparent to the community. Developers and researchers can download it from Hugging Face and even integrate it via API, ensuring that the removal of censorship can be scrutinized and built upon by others.
(Source: Perplexity AI)
Implications of Removing the Censorship
Perplexity AI’s decision to remove the Chinese censorship from DeepSeek R1 carries several important implications for the AI community:
- Enhanced Openness and Truthfulness: Users of R1 1776 can now receive uncensored, direct answers on previously off-limits topics, which is a win for open inquiry. This could make it a more reliable assistant for researchers, students, or anyone curious about sensitive geopolitical questions. It’s a concrete example of using open-source AI to counteract information suppression.
- Maintained Performance: There were concerns that tweaking the model to remove censorship might degrade its performance in other areas. However, Perplexity reports that R1 1776’s core skills – like math and logical reasoning – remain on par with the original model. In tests on over 1,000 examples covering a broad range of sensitive queries, the model was found to be “fully uncensored” while retaining the same level of reasoning accuracy as DeepSeek R1. This suggests that bias removal (at least in this case) didn’t come at the cost of overall intelligence or capability, which is an encouraging sign for similar efforts in the future.
- Positive Community Reception and Collaboration: By open-sourcing the decensored model, Perplexity invites the AI community to inspect and improve upon their work. It demonstrates a commitment to transparency – the AI equivalent of showing one’s work. Enthusiasts and developers can verify that the censorship restrictions are truly gone and potentially contribute to further refinements. This fosters trust and collaborative innovation in an industry where closed models and hidden moderation rules are common.
- Ethical and Geopolitical Considerations: On the flip side, completely removing censorship raises complex ethical questions. One immediate concern is how this uncensored model might be used in contexts where the censored topics are illegal or dangerous. For instance, if someone in mainland China were to use R1 1776, the model’s uncensored answers about Tiananmen Square or Taiwan could put the user at risk. There’s also the broader geopolitical signal: an American company altering a Chinese-origin model to defy Chinese censorship can be seen as a bold ideological stance. The very name “1776” underscores a theme of liberation, which has not gone unnoticed. Some critics argue that replacing one set of biases with another is possible – essentially questioning whether the model might now reflect a Western point of view in sensitive areas. The debate highlights that censorship vs. openness in AI is not just a technical issue, but a political and ethical one. Where one person sees necessary moderation, another sees censorship, and finding the right balance is tricky.
The removal of censorship is largely being celebrated as a step toward more transparent and globally useful AI models, but it also serves as a reminder that what an AI should say is a sensitive question without universal agreement.

(Source: Perplexity AI)
The Bigger Picture: AI Censorship and Open-Source Transparency
Perplexity’s R1 1776 launch comes at a time when the AI community is grappling with questions about how models should handle controversial content. Censorship in AI models can come from many places. In China, tech companies are required to build in strict filters and even hard-coded responses for politically sensitive topics. DeepSeek R1 is a prime example of this – it was an open-source model, yet it clearly carried the imprint of China’s censorship norms in its training and fine-tuning. By contrast, many Western-developed models, like OpenAI’s GPT-4 or Meta’s LLaMA, aren’t beholden to CCP guidelines, but they still have moderation layers (for things like hate speech, violence, or disinformation) that some users call “censorship.” The line between reasonable moderation and unwanted censorship can be blurry and often depends on cultural or political perspective.
What Perplexity AI did with DeepSeek R1 raises the idea that open-source models can be adapted to different value systems or regulatory environments. In theory, one could create multiple versions of a model: one that complies with Chinese regulations (for use in China), and another that is fully open (for use elsewhere). R1 1776 is essentially the latter case – an uncensored fork meant for a global audience that prefers unfiltered answers. This kind of forking is only possible because DeepSeek R1’s weights were openly available. It highlights the benefit of open-source in AI: transparency. Anyone can take the model and tweak it, whether to add safeguards or, as in this case, to remove imposed restrictions. Open sourcing the model’s training data, code, or weights also means the community can audit how the model was modified. (Perplexity hasn’t fully disclosed all the data sources it used for de-censoring, but by releasing the model itself they’ve enabled others to observe its behavior and even retrain it if needed.)
This event also nods to the broader geopolitical dynamics of AI development. We are seeing a form of dialogue (or confrontation) between different governance models for AI. A Chinese-developed model with certain baked-in worldviews is taken by a U.S.-based team and altered to reflect a more open information ethos. It’s a testament to how global and borderless AI technology is: researchers anywhere can build on each other’s work, but they are not obligated to carry over the original constraints. Over time, we might see more instances of this – where models are “translated” or adjusted between different cultural contexts. It raises the question of whether AI can ever be truly universal, or whether we will end up with region-specific versions that adhere to local norms. Transparency and openness provide one path to navigate this: if all sides can inspect the models, at least the conversation about bias and censorship is out in the open rather than hidden behind corporate or government secrecy.
Finally, Perplexity’s move underscores a key point in the debate about AI control: who gets to decide what an AI can or cannot say? In open-source projects, that power becomes decentralized. The community – or individual developers – can decide to implement stricter filters or to relax them. In the case of R1 1776, Perplexity decided that the benefits of an uncensored model outweighed the risks, and they had the freedom to make that call and share the result publicly. It’s a bold example of the kind of experimentation that open AI development enables.
Leave a comment