Global Feed Post Login
Replying to Avatar Dan

Almost 100% agree, but there is no way they did this without Nvidia chips.

They've open sourced a ground breaking LLM scaling paradigm (RL on COT), which is no small thing believe me, but our closed source reasoning models are likely doing something similar (we just can't see it).

This newly open scaling paradigm is a game changer, but you still don't get this performance without massive compute.

They have illegal H100s, I'm nearly certain of it. Nvidia was probably due for a correction anyway. But it'll be funny to see what happens when we all find out they did this with Nvidia chips

Avatar
Guy Swann 11mo ago

Do you have a good write up on the scaling paradigm? I read that in another post but couldn’t confirm it yet and wasn’t sure what that meant.

Any explanation or breakdown link would be appreciatively zapped

Reply to this note

Please Login to reply.

Discussion

Avatar
Dan 11mo ago

this overview is pretty solid

https://www.youtube.com/watch?v=sGUjmyfof4Q

https://mirror-feeling-d80.notion.site/DeepSeek-R1-182808527b17801585dadb84f7c66cd9

Thread collapsed