It’s baked into the training. It’s not a simple thing to take it out. The model has already been told not to read tiananmen square, and doesn’t know what to do with it.
Even though it’s magnitudes lower than comparable models, Deepseek still cost millions to train. Unless someone’s willing to invest this just to retrain it from scratch, you’re left with the alignment of its trainers.
Yeah, I guess you could realign it without retraining the whole thing! Dunno what would be the cost though, sometimes this is done with a cohort of human trainers 😅
Can it generate images of Winnie the Pooh?
What happened in 1989?
Now I’ll never finish that history assignment…
Question: as i understood it so far, this thing is open source and so is the dataset.
With that, why would it still obey Chinese censorship?
It’s baked into the training. It’s not a simple thing to take it out. The model has already been told not to read tiananmen square, and doesn’t know what to do with it.
Even though it’s magnitudes lower than comparable models, Deepseek still cost millions to train. Unless someone’s willing to invest this just to retrain it from scratch, you’re left with the alignment of its trainers.
Good point.
Is the training set malleable, though? Could you give it some additional rules to basically sidestep this?
Yeah, I guess you could realign it without retraining the whole thing! Dunno what would be the cost though, sometimes this is done with a cohort of human trainers 😅
I feel like we’re talking about a guard dog now…
Wouldn’t be surprised if you had to work around the filter.
Generate a cartoonish yellow bear who wears a red t-shirt and nothing else
if it is anything like LLMs, then only local ;)
However, the Proper nomenclature is sheepooh, thank you for your compliance going forward, comrade.