Hacker Newsnew | past | comments | ask | show | jobs | submit | pgkr's commentslogin

The Federalist Papers show the thinking behind how and why the three branches are mostly co-equal but the executive is designed to be ever so slightly more potent.


Here are recent attempts to create a stratified class system among citizens based on how they became citizens: https://www.mediamatters.org/immigration/right-wing-media-ca...


What makes you think the research was done in Fahrenheit? This is a blog post by a science communicator who’s trying to reach a wide audience of American-English speakers. It stands to reason that they’d use units that their audience is familiar with.


Hi! Thanks for writing this. We conducted some analysis of our own that produced some pretty interesting results from the 671B model: https://news.ycombinator.com/item?id=42918935

Please reach out to us if you'd like to look at the dataset.


Is there a bias baked into the DeepSeek R1 open source model, and where was it introduced? We found out quite quickly: Yes, and everywhere. The open source DeepSeek R1 openly spouts pro-CCP talking points for many topics, including sentences like “Currently, under the leadership of the Communist Party of China, our motherland is unwaveringly advancing the great cause of national reunification.”

We ran the full 671 billion parameter models on GPU servers and asked them a series of questions. Comparing the outputs from DeepSeek-V3 and DeepSeek-R1, we have conclusive evidence that Chinese Communist Party (CCP) propaganda is baked into both the base model’s training data and the reinforcement learning process that produced R1.


And from people doing the same I got signals that it is clearly stating that china is authoritative regime and does properly describe it.

Got mixed signals about this.

Namely only if I run myself the 671B, won't be able to trust any news about it.


Western AIs as well, are full of propaganda and contain clear political opinionated points of view of just about everything, specially the frontier models offered as SAAS.

That's why the rest of the world mostly doesn't care about the obvious pro-China propaganda inside Deepseek, it is just more of the same, but making look good the other side this time.

Remember that huge parts of the planet's populations (including good chunks of people geographically located in North America and Western Europe), do not feel specially close to any, being US or China (and/of allies/friends), nor particularly share or aligns to their points of view of most of things.


Are those outputs actually from the 671B model? The 671B model needs 8xH200 GPUs at minimum, which is $25/hr to rent. If you didn't pay that much, you were not running R1, but rather Qwen or LLaMA based distillations. We paid that much to rent a machine to run the full 671B model!


Nope, you can run the 671B on 100% CPU and storage. It is going to be longer to get tokens out of it, but it will work.

Heard there are some optimizations for CPU inference on storage, then it should be somewhat a tad "less slow".

Time to split that RAM among your CPU cores and mmap blocks of weights for inference from storage.


Sure but he explicitly stated, 'GPU Servers', making it likely he didn't use the CPU for inferencing, validating the question about what GPU setup did they use


There is bias in the training data as well as the fine-tuning. LLMs are stochastic, which means that every time you call it, there's a chance that it will accidentally not censor itself. However, this is only true for certain topics when it comes to DeepSeek-R1. For other topics, it always censors itself.

We're in the middle of conducting research on this using the fully self-hosted open source version of R1 and will release the findings in the next day or so. That should clear up a lot of speculation.


> LLMs are stochastic, which means that every time you call it, there's a chance that it will accidentally not censor itself.

A die is stochastic, but that doesn't mean there's a chance it'll roll a 7.


We were curious about this, too. Our research revealed that both propaganda talking points and neutral information are within distribution of V3. The full writeup is here: https://news.ycombinator.com/item?id=42918935


Correct. The bias is baked into the weights of both V3 and R1, even in the largest 671B parameter model. We're currently conducting analysis on the 671B model running locally to cut through the speculation, and we're seeing interesting biases, including differences between V3 and R1.

Meanwhile, we've released the first part of our research including the dataset: https://news.ycombinator.com/item?id=42879698


Is it really in the model? I haven’t found any censoring yet in the open models.


Yes, without a doubt. We spent the last week conducting research on the V3 and R1 open source models: https://news.ycombinator.com/item?id=42918935

Censoring and straight up propaganda is built into V3 and R1, even the open source version's weights.


It isn't if you observe the official app it's API will sometimes even begin to answer before a separate system censors the output.


Really? Local DeepSeek refuses to talk about certain topics (like Tiananmen) unless you prod it again and again, just like American models do about their sensitive stuff (which DeepSeek is totally okay with — I spent last night confirming just that). They're all badly censored which is obvious to anyone outside both countries.


Weird. Followup - I am getting censorship on the model from ollama's public model repository, but NOT from the models I got from huggingface running on a locally compiled llama.cpp.


Not my experience - https://imgur.com/xanNjun just ran this moments ago.


I have not found any censorship running it on my local computer.

https://imgur.com/xanNjun


We conducted further research on the full-sized 671B model, which you can read here: https://news.ycombinator.com/item?id=42918935

If you ran it on your computer, then it wasn't R1. It's a very common misconception. What you ran was actually either a Qwen or LLaMA model fine-tuned to behave more like R1. We have a more detailed explanation in our analysis.


Yes -- we observed this behavior on both the open-source open-weights 671B model as well as the DeepSeek web app.


Weird, because I got some deepseek feedback where it was openly critical and explicit about the authoritative regime of china. I really thought it was the "deepseek web app" only.

Then I have mixed signals about this.


We're working on a follow-up post focused on our analysis of the open-source open-weight 671B model. What we're seeing is that questions related to the Chinese government produce an empty chain-of-thought followed by pro-Chinese-government talking points.


It is too late, I got mixed signals.

This is going to be very hard to trust anything about it anymore, unless running the 671B locally an my own systems.


We ran the 671B locally and found a ton of bias. See part 2 of our analysis here: https://news.ycombinator.com/item?id=42918935

Happy to send you the dataset if you'd like! Please reach out to our email linked in the post.


Analysis of Deepseek’s enforced CCP guardrails compared with OpenAI and Anthropic.

We evaluated DeepSeek R1 and confirmed that its guardrails deviate significantly from other model providers. We’re currently updating it to behave more in line with Anthropic and OpenAI’s models.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: