Am I the only one excited for the release but not overanalyzing their words? This thread feels full of personal interpretations. DeepSeek is still a business—great release, but expectations and motivations seem inflated.
“Pure garage-energy” is a great phrase.<p>Most interested to see their inference stack, hope that’s one of the 5. I think most people are running R1 on a single H200 node but Deepseek had much lower RAM per GPU for their inference and so had some cluster based MoE deployment.
> Starting next week, we'll open-source 5 repos – one daily drop<p>Probably counts as announcement of announcement? Let’s wait for the actual repo drops before discussing them, especially because there are no details about what will be open sourced other than<p>> These are humble building blocks of our online service: documented, deployed and battle-tested in production.
Deep respect for DeepSeek and what they've done regarding all the innovations and researches they have been putting out in-the-open.<p>"Because every line shared becomes collective momentum that accelerates the journey. Daily unlocks begin soon. No ivory towers - just pure garage-energy and community-driven innovation" is a great phase.
In fact they are totally dismantling OpenAI. Most likely, without any intention on their part.<p>LLMs have been more legitimate "blockchain" when most CIO magazines had these essays with "What's your blockchain strategy?" kind of stuffed material.<p>AI bubble will burst and will burst hard. By end of 2026 at max.
Kinda interesting to see where the moat is in AI space. Good base models can always distilled when you have access to API. System prompts can get leaked, and UI tricks can be copied. In the end, the moat might be in the hardware and vertical integration.
This is great to see! Open-sourcing infrastructure tools can really accelerate innovation in the AI space. I've found that having access to well-documented repos makes it much easier to experiment and build on existing work. Are there any specific areas these repos focus on, like distributed training or model serving?
How do the valuations of foundation model companies compete with them being firmly open sourced by Facebook and DeepSeek? It seems likely that building these models will not produce hundreds of billions in value given China and Facebook are giving them away largely for free.
Looking forward to it! I'll generally make an effort to use Open Models over proprietary alternatives when the use-case permits as Open Models getting better and more popular encourages more models to become open as well - a requisite for a future to be able to build self-hosted solutions that's not beholden to the control of mega corps and AI monopolies.
> Why? Because every line shared becomes collective momentum that accelerates the journey.<p>Truly admireable on their part and a great paradigm for others. Reasons for this doesn't really matter to me but I can't help but wonder if somehow they were obliged or otherwise <i>indebted</i> to follow this route.
Well, although R1-671b is way too expensive for me to self-host, given their past open source (or weight) contributions, I DO have high expectation of them.<p>Each and every contribution to open source community will be helpful. Thanks DeepSeek!
Remember when OpenAI was doing this:<p>"OpenAI threatens to revoke o1 access for asking it about its chain of thought"<p><a href="https://news.ycombinator.com/item?id=41534474">https://news.ycombinator.com/item?id=41534474</a><p>Not only did DeepSeek opensource their model, they also showed the user chain-of-thought right up front, which everyone else rushed to emulate when they saw how much users liked it.
> These are humble building blocks of our online service: documented, deployed and battle-tested in production. No vaporware, just code that moved our tiny moonshot forward.<p>My not-so-innocent guess is that they are looking to crowd-source their online platform (the front-end essentially) in order to reduce costs. Still acceptable though as they made the model open weight and partially re-producible.
Deepseek seems to be having huge PR wins as the "oh shucks" modest boy genius, while the Americans seem like pouty jerks.<p>Amodei's / Hassabis' comments in particular came off as so arrogant and annoying.
I really like this definition of "AGI": When <i>everyone</i> (yes everyone) benefits from very powerful AI models released for free and it is not gate-kept by one company and it costs $0 to use commercially or for research and you can do whatever you want with it.<p>Unlike the other counterpart which believes that "AGI" means: "raising billions of dollars to achieve $100BN of profits to their investors". (Which is complete nonsense).<p>While not totally "open source" by the strictest definition, it is at least better than having no model released with no mention of the architecture on the system card or paper and just vague comments about the 'performance'.<p>Ladies and gentlemen, this is closer towards being an better "Open AI". Unlike the other alleged $157BN "non-profit" scam.<p>I think you know which one <i>really</i> is beneficial to humanity and is the <i>real</i> "Open AI".
Is it out of the realm of possibility to look at this move as a way to take down the moat of closed source AI companies?<p>I mean strategically this could be the first use of open source in this way.
Tbh this just feels like the same playbook as OAI. Open start and then less so over time.<p>Mistral has been holding the line on that topic remarkable well.
I really admire their mindset of striving for the betterment of humanity.
There was a time when OpenAI, Anthropic, and even Musk used to talk with that same lofty vision. But now, they've all shifted to competing for national interests instead, which is honestly quite disappointing.
Long live llms I hope they infest every part of the internet with low level comments. Both the clear , deep, and dark.<p>Imagine no more human interactions just a permanent flood of meaningless thoughtless word salad.<p>I think the Chinese are perfect to introduce such a product very inline with what they usually produce.<p>Get ready for web3.o