The local models(full and distilled) are also censored. The models censorship is just implemented superficially to immediately close any thinking tags and refuse when detecting censored material. If there already is any token after the <think> token the model will start answering away, which also happens on the official API because it puts a new line after the <think> token for some reason. That’s why on chat.deepseek.com censored topics are first answered and then redacted by some other safeguard a few seconds later. Whilst there are some great abliterated(=technique that tries to remove parts of llms that cause refusals) versions of the distills on huggingface that prevent all refusals after a few tries, they only tackle refusals, not political opinions such as Taiwan’s status as an independent country.
The local models(full and distilled) are also censored. The models censorship is just implemented superficially to immediately close any thinking tags and refuse when detecting censored material. If there already is any token after the <think> token the model will start answering away, which also happens on the official API because it puts a new line after the <think> token for some reason. That’s why on chat.deepseek.com censored topics are first answered and then redacted by some other safeguard a few seconds later. Whilst there are some great abliterated(=technique that tries to remove parts of llms that cause refusals) versions of the distills on huggingface that prevent all refusals after a few tries, they only tackle refusals, not political opinions such as Taiwan’s status as an independent country.