This sounds like cope.

Possibly they fine tuned DeepSeek on an OpenAI model (cheaper than using humans), but it makes no sense to primarily do this when self-supervised learning and RL is much more efficient. Also, DeepSeek performs better than OpenAI on several benchmarks - you can't achieve this purely by distilling a teacher model.

Likely they made a technical breakthrough and USA "AI tzar" is seething.

Reply to this note

Please Login to reply.

Discussion

They have said themselves that they did not make a technical breakthrough. They just open-sourced everything.

Here are some of the specific technical things they did to achieve lower costs. More efficient training procedure, memory compression, reliance on RL, low level code optimisation.

https://www.analyticsvidhya.com/blog/2025/01/how-deepseek-trained-ai-30-times-cheaper/

open source > stupid copyright bullshit

the best part of #deepsnek is that this is gonna crater all the closed source projects prospects for future funding

investers will be like, "closed source means expensive, pass"

oh, there was another thing that is gonna come out of this that is awesome too

AMD was lagging in the general purpose compute space despite their simpler, cheaper hardware and open source AI

now they will be looked at again for further cost benefits

AI is now at that stage in its development like when you are writing code and it works, but it's slow, expensive and a bit clunky

it works!

but now the optimizations start and the race is on for the most streamlined implementations

personally, i am looking forward to when someone builds a model out of the nostr and all of the web pages embedded in the links on it, this will be epic

Even Meta has been giving up on closed source.

Closed-source products open to the public are sort of pointless, anyway. Either something is a secret, and then you keep it to yourself, or it's not and then who cares.