Ming-Chi Kuo
2025.02.06 03:16

DeepSeek Accelerates the GenAI Industry into a New Phase: New Compute Power Growth Approach and Application Diversification

portai
I'm PortAI, I can summarize articles.

The release of DeepSeek-R1 has brought two major AI industry trends into focus. Though these
trends were already on the horizon, DeepSeek-R1 has sped up their development.

  1. Despite the diminishing returns from scaling laws, AI compute power can continue to grow through
    optimized training approaches, facilitating the exploration of new applications.
  2. A marked drop in API/token prices benefits AI software/services and on-device AI, thereby speeding
    up the diversification of AI applications.

Trend 1: Despite the diminishing returns from scaling laws, AI compute power can continue to
grow through optimized training approaches, facilitating the exploration of new applications.

  1. For the past two years, AI server supply chain stock investments have been driven by expectations of continued growth based on scaling laws. Now, as the limitations of these scaling laws become more
    evident, the market is turning its attention to DeepSeek's optimized training approach - achieving
    major performance gains without relying on scaling laws.
  2. According to the widely cited Chinchilla Scaling Law, AI model performance is driven by three key
    factors: model parameter (N), training data (D), and computing power (C). Ideal results come from
    scaling all three factors together.
  3. The diminishing returns of scaling laws can be attributed to several key factors. The supply of human-created text data (D) is nearly exhausted, while simply increasing model parameters (N) proves
    ineffective without matching improvements in computing power (C) and addressing data volume
    constraints (D). Additionally, substantial short-term increases in computing power (C) face major
    hurdles, from delays in Blackwell system production to power supply limitations.
  4. What makes DeepSeek-R1 particularly noteworthy from an industry research perspective is its ability to achieve significant performance improvements through optimized training approaches rather than scaling laws. These improvements can be verified through examination and testing of its open-source releases.
  5. As the diminishing returns of scaling laws become increasingly evident, improving model
    performance through optimized training approaches has become essential. This shift enables
    sustainable growth in AI computing infrastructure and unlocks the potential for new applications.
    These complementary developments are crucial for the long-term advancement of the AI industry.

Trend 2: A marked drop in API/token prices benefits AI software/services and on-device AI,
thereby speeding up the diversification of AI applications

  1. Currently, profits in the GenAI sector primarily stem from infrastructure provision (‘selling shovels’)
    and cost reduction rather than from creating new business models or generating significant
    value-add to existing products and services.
  2. DeepSeek-R1 has adopted an aggressive pricing strategy. It is free to use, and the lowest API/token
    pricing is less than 1/100 of OpenAI-o1. This competitive pressure is likely to drive down AI usage
    costs. Given the intense competition in China's AI market, other Chinese firms are expected to launch LLMs with impressive benchmark scores and even more aggressive pricing.
  3. The recent sharp correction in AI supply chain stocks primarily reflects investors' reset expectations for AI server shipments - a response to the diminishing returns of scaling laws rather than concerns
    over LLM service providers' and CSPs' path to profitability. Investors largely remain patient regarding profit materialization.
  4. Declining costs in AI software/services and on-device AI, driven by lower API/token prices and
    optimized training approaches, are stimulating AI compute demand while alleviating investor
    concerns about AI investment profitability.
  5. While declining AI prices will undoubtedly drive higher usage, it remains unclear whether volume
    growth can offset price reductions. Increased AI adoption may enable profitable business models, though success is not guaranteed. These uncertainties currently appear manageable as investors
    maintain their patience regarding profitability.

Conclusion:

  1. Scaling laws are empirical observations. Moderating expectations and taking a rational view of them can actually benefit long-term investment trends. Improvements such as chip upgrades (C), resolving power supply constraints (C), and the use of multimodal training data (D) in training all have the potential to re-accelerate the benefits of scaling laws in the future.
  2. The diminishing returns from scaling laws only affect those operating at a massive scale, proving Nvidia's leadership position. When Nvidia's solutions speed up returns of scaling laws again, the company's edge over competitors like ASIC and AMD will likely become even more pronounced.
  3. Recent production issues with the GB200 NVL72 make this a good time to reset expectations for scaling laws and AI server shipments. This stock correction will be more favorable for positive future reactions to the GB300/Rubin series.
  4. Leading CSPs won't reduce capital expenditure simply because of improved training approaches, as
    efficiency gains and infrastructure investments can coexist. Any reduction in capital spending now could leave them vulnerable to falling behind competitors when scaling law benefits accelerate again.
  5. The combined effect of open-source resources and the highly competitive environment in China is
    expected to lead to other Chinese companies launching high-performing LLMs with even more
    aggressive pricing. If LLM service providers have not established stable profitability by then, their profit pressures will intensify.
  6. Benefiting from significant API/token price declines, AI software/service and on-device AI will attract
    greater investor attention. Whether this becomes a new long-term investment trend depends on whether profitable business models can be created.
  7. Nvidia remains the likely winner when the benefits of scaling laws accelerate again in the future.
    However, monitoring short-term GB200 NVL72 production issues and potential changes to US
    semiconductor export restrictions in the medium-to-long term is crucial.

$NVIDIA(NVDA.US) $Taiwan Semiconductor(TSM.US)

The copyright of this article belongs to the original author/organization.

The views expressed herein are solely those of the author and do not reflect the stance of the platform. The content is intended for investment reference purposes only and shall not be considered as investment advice. Please contact us if you have any questions or suggestions regarding the content services provided by the platform.