You are currently viewing SemiWiki as a guest which gives you limited access to the site. To view blog comments and experience other SemiWiki features you must be a registered member. Registration is fast, simple, and absolutely free so please, join our community today!
I think you're trying to be pedantic, about something you're wrong about, but something that is also meaningless to the original discussion. The base model, DeepSeek V3 used distilled data using OpenAI API - it's becoming more and more apparent. DeepSeek R1 was built on top of DeepSeek V3, adding reasoning using reinforcement learning.
I think you're trying to be pedantic, about something you're wrong about, but something that is also meaningless to the original discussion. The base model, DeepSeek V3 used distilled data using OpenAI API - it's becoming more and more apparent. DeepSeek R1 was built on top of DeepSeek V3, adding reasoning using reinforcement learning.
TSMC was $122 before Deepseek. It dropped to $188 on Monday 27th when Deepseek "shock the world". It closed today at $209, still $13 off from $122 but $21 recovered from $188.
Looks like the Deepfake glorious one-day "breakthrough" tour is over?
TSMC was $122 before Deepseek. It dropped to $188 on Monday 27th when Deepseek "shock the world". It closed today at $209, still $13 off from $122 but $21 recovered from $188.
Looks like the Deepfake glorious one-day "breakthrough" tour is over?
Thought this was an interesting view from some experts. Good chance that DeepSeek did distill from OpenAI, but that’s not the important thing. Biggest reason to train/create biggest (most compute intensive) models to be able to distill into smaller targeted solutions models.
Thought this was an interesting view from some experts. Good chance that DeepSeek did distill from OpenAI, but that’s not the important thing. Biggest reason to train/create biggest (most compute intensive) models to be able to distill into smaller targeted solutions models.
Buried in the discussion at the end is that the focus moves from just the models and to the entire AI app solution framework - the models are just building blocks.
Buried in the discussion at the end is that the focus moves from just the models and to the entire AI app solution framework - the models are just building blocks.
I don't think there is a strong argument of solution stack. I think the most important aspects are the capability of a model and then cost. Once you have those, there are frameworks such as LangChain that people can leverage. For serious development, unique data set and evaluation approaches unique to an application, are important but they are not shared and hence I don't think they are part of any stack.
- A common disease in some Silicon Valley circles: a misplaced superiority complex. - Symptom of advanced stage: thinking your small tribe has a monopoly on… | 323 comments on LinkedIn
SoftBank inked a $3 billion deal with OpenAI in a joint venture to market OpenAI tech in Japan with its newly-minuted "Cristal Intelligence" suite of tools.
SoftBank inked a $3 billion deal with OpenAI in a joint venture to market OpenAI tech in Japan with its newly-minuted "Cristal Intelligence" suite of tools.
He also funded WeWork. At the moment, even with the current rates charged by OpenAI, they are not profitable. I believe that due to Deepseek, OpenAI is rushing to launch new services. How can they make money with increasing competition (significantly lower costs)?