Sundar Pichai recently provided epic news to Nvidia stock investors


nvidia (NASDAQ: NVDA) Stocks are currently down 12% from an all-time high. In January, China-based Startup Deepseek claimed it had trained competitive artificial intelligence (AI) models using some of its computing capabilities deployed by major developers like Openai. I was struggling with the sharp sale.

Investors fear that DeepSeek’s technique will be adopted by other AI developers, and demand for Nvidia’s high-end graphics processing units (GPUs), the best hardware available to develop AI models, has dropped significantly. did. However, these concerns may have been exaggerated.

Google’s parents alphabet (NASDAQ: GOOG)(NASDAQ: Google) On February 4th, CEO Sundar Pichai, a large buyer of Nvidia’s AI data center chips, made some comments to help Nvidia investors feel much better.

Nvidia's headquarters with black nvidia will sign out from the front.
Image source: nvidia.

Deepseek was founded in 2023 with the success of a Chinese hedge fund called High-Flyer. It has been building trading algorithms for many years using AI. Deepseek has released V3 Big language model (LLM) In December 2024, the R1 inference model followed in January, and its competitiveness with some of the latest models from Openai and other start-ups made the tech sector buzz.

Because Deepseek’s work is open source, the industry quickly learned some important details. The startup claims it trained for just $5.6 million (not including an estimated $500 million in chips and infrastructure, according to SemianAlysis). To reach the current stage of development.

Deepseek also used older generation Nvidia GPUs like the H100 because the US government banned chip makers from selling modern hardware to Chinese companies (to protect American AI leadership) .

It turns out that DeepSeek has implemented several unique innovations on the software side to make up for its lack of computing power. We developed highly efficient algorithms and data entry methods, and also used a technique called distillation. This involves training smaller models using knowledge from large, already successful AI models.

In fact, Openai accuses DeepSeek of using the GPT-4O model to train DeepSeek R1, urging the ChatGPT chatbot to “learn” from the output. Distillation accelerates the training process quickly, as developers do not need to collect or process the mount. The result is much less computing power, meaning less GPU.

Leave a Reply

Your email address will not be published. Required fields are marked *