Array ( [content] => [params] => Array ( [0] => /forum/threads/will-the-chinese-deepseek-ai-upset-the-ai-ml-race.21768/page-4 ) [addOns] => Array ( [DL6/MLTP] => 13 [Hampel/TimeZoneDebug] => 1000070 [SV/ChangePostDate] => 2010200 [SemiWiki/Newsletter] => 1000010 [SemiWiki/WPMenu] => 1000010 [SemiWiki/XPressExtend] => 1000010 [ThemeHouse/XLink] => 1000970 [ThemeHouse/XPress] => 1010570 [XF] => 2021770 [XFI] => 1050270 ) [wordpress] => /var/www/html )
I tend to try things myself instead of relying on randos on YouTube or TikTok. Here's simple output from OpenAI with reasoning path:
I think you're trying to be pedantic, about something you're wrong about, but something that is also meaningless to the original discussion. The base model, DeepSeek V3 used distilled data using OpenAI API - it's becoming more and more apparent. DeepSeek R1 was built on top of DeepSeek V3, adding reasoning using reinforcement learning.I think there is a misunderstanding about "reasoning process". In reasoning LLM,it is referred as "chain of thoughts"
What you have posted the result from ChatGPT,is mere final answer of the question,without showing user chain of thoughts.
Here is the output from the exact same question from deepseek,you can see the detailed chain of thoughts before final answer
I think you're trying to be pedantic, about something you're wrong about, but something that is also meaningless to the original discussion. The base model, DeepSeek V3 used distilled data using OpenAI API - it's becoming more and more apparent. DeepSeek R1 was built on top of DeepSeek V3, adding reasoning using reinforcement learning.
TSMC was $122 before Deepseek. It dropped to $188 on Monday 27th when Deepseek "shock the world". It closed today at $209, still $13 off from $122 but $21 recovered from $188.
Looks like the Deepfake glorious one-day "breakthrough" tour is over?
Thought this was an interesting view from some experts. Good chance that DeepSeek did distill from OpenAI, but that’s not the important thing. Biggest reason to train/create biggest (most compute intensive) models to be able to distill into smaller targeted solutions models.AI experts already debunked OpenAI's claim that deepseek R1 has anything to do with OpenAI's product
Then you can run those models on low cost devices. I believe IBM is a customer of Gaudi 3.Thought this was an interesting view from some experts. Good chance that DeepSeek did distill from OpenAI, but that’s not the important thing. Biggest reason to train/create biggest (most compute intensive) models to be able to distill into smaller targeted solutions models.
Then you can run those models on low cost devices. I believe IBM is a customer of Gaudi 3.
Buried in the discussion at the end is that the focus moves from just the models and to the entire AI app solution framework - the models are just building blocks.Also distillation requires much less GPU training time.
I don't think there is a strong argument of solution stack. I think the most important aspects are the capability of a model and then cost. Once you have those, there are frameworks such as LangChain that people can leverage. For serious development, unique data set and evaluation approaches unique to an application, are important but they are not shared and hence I don't think they are part of any stack.Buried in the discussion at the end is that the focus moves from just the models and to the entire AI app solution framework - the models are just building blocks.
Congrats to DeepSeek on producing an o1-level reasoning model! Their research paper demonstrates that they’ve independently found some of the core ideas that we did on our way to o1.
— Mark Chen (@markchen90) January 28, 2025
Revaluing Nvidia, I leave market share and margin unchanged, but shrink the end market for AI chips. My value drops to $78, but with the price at $123, it is over valued by 59%, driving my decision to sell half of my remaining Nvidia shares, but it should not drive your… pic.twitter.com/wWzBISgEJ5
— Aswath Damodaran (@AswathDamodaran) January 31, 2025
— Marc Andreessen 🇺🇸 (@pmarca) February 1, 2025
These meme stock dynamics will chase a lot of funds away despite the attractive fundamentals. I have already heard from many large buy-side funds they are considering moving off Nvidia to other more stable names. https://t.co/UhBQjHaOoV
— Ben Bajarin (@BenBajarin) February 1, 2025
He also funded WeWork. At the moment, even with the current rates charged by OpenAI, they are not profitable. I believe that due to Deepseek, OpenAI is rushing to launch new services. How can they make money with increasing competition (significantly lower costs)?And yet SoftBank will pay $3B / year to OpenAI so they can offer AI solutions in Japan... Real money flow is more probably important than benchmarks.
![]()
SoftBank joins with OpenAI in yearly $3B venture to expand AI in Japan
SoftBank inked a $3 billion deal with OpenAI in a joint venture to market OpenAI tech in Japan with its newly-minuted "Cristal Intelligence" suite of tools.ca.finance.yahoo.com
Rumor: A U.S. securities firm has adjusted the shipment forecasts for GB200 and GB300.
— Jukanlosreve (@Jukanlosreve) February 4, 2025
- 2382 Quanta: U.S. Investment Bank Lowers Target Price
A major U.S. investment bank has lowered its EPS forecasts for 2025-2026 to $17.1/$19.4 and adjusted its target price downward, based… https://t.co/bV8tzkdqSG