
DeepSeek Accelerates the GenAI Industry into a New Phase: New Compute Power Growth Approach and Application Diversification

I'm PortAI, I can summarize articles.
The release of DeepSeek-R1 has brought two major AI industry trends into focus. Though these
trends were already on the horizon, DeepSeek-R1 has sped up their development.
- Despite the diminishing returns from scaling laws, AI compute power can continue to grow through
optimized training approaches, facilitating the exploration of new applications.
- A marked drop in API/token prices benefits AI software/services and on-device AI, thereby speeding
up the diversification of AI applications.
Trend 1: Despite the diminishing returns from scaling laws, AI compute power can continue to
grow through optimized training approaches, facilitating the exploration of new applications.
- For the past two years, AI server supply chain stock investments have been driven by expectations of continued growth based on scaling laws. Now, as the limitations of these scaling laws become more
evident, the market is turning its attention to DeepSeek's optimized training approach - achieving
major performance gains without relying on scaling laws.
- According to the widely cited Chinchilla Scaling Law, AI model performance is driven by three key
factors: model parameter (N), training data (D), and computing power (C). Ideal results come from
scaling all three factors together.
- The diminishing returns of scaling laws can be attributed to several key factors. The supply of human-created text data (D) is nearly exhausted, while simply increasing model parameters (N) proves
ineffective without matching improvements in computing power (C) and addressing data volume
constraints (D). Additionally, substantial short-term increases in computing power (C) face major
hurdles, from delays in Blackwell system production to power supply limitations.
- What makes DeepSeek-R1 particularly noteworthy from an industry research perspective is its ability to achieve significant performance improvements through optimized training approaches rather than scaling laws. These improvements can be verified through examination and testing of its open-source releases.
- As the diminishing returns of scaling laws become increasingly evident, improving model
performance through optimized training approaches has become essential. This shift enables
sustainable growth in AI computing infrastructure and unlocks the potential for new applications.
These complementary developments are crucial for the long-term advancement of the AI industry.
Trend 2: A marked drop in API/token prices benefits AI software/services and on-device AI,
thereby speeding up the diversification of AI applications
- Currently, profits in the GenAI sector primarily stem from infrastructure provision (‘selling shovels’)
and cost reduction rather than from creating new business models or generating significant
value-add to existing products and services.
- DeepSeek-R1 has adopted an aggressive pricing strategy. It is free to use, and the lowest API/token
pricing is less than 1/100 of OpenAI-o1. This competitive pressure is likely to drive down AI usage
costs. Given the intense competition in China's AI market, other Chinese firms are expected to launch LLMs with impressive benchmark scores and even more aggressive pricing.
- The recent sharp correction in AI supply chain stocks primarily reflects investors' reset expectations for AI server shipments - a response to the diminishing returns of scaling laws rather than concerns
over LLM service providers' and CSPs' path to profitability. Investors largely remain patient regarding profit materialization.
- Declining costs in AI software/services and on-device AI, driven by lower API/token prices and
optimized training approaches, are stimulating AI compute demand while alleviating investor
concerns about AI investment profitability.
- While declining AI prices will undoubtedly drive higher usage, it remains unclear whether volume
growth can offset price reductions. Increased AI adoption may enable profitable business models, though success is not guaranteed. These uncertainties currently appear manageable as investors
maintain their patience regarding profitability.
Conclusion:
- Scaling laws are empirical observations. Moderating expectations and taking a rational view of them can actually benefit long-term investment trends. Improvements such as chip upgrades (C), resolving power supply constraints (C), and the use of multimodal training data (D) in training all have the potential to re-accelerate the benefits of scaling laws in the future.
- The diminishing returns from scaling laws only affect those operating at a massive scale, proving Nvidia's leadership position. When Nvidia's solutions speed up returns of scaling laws again, the company's edge over competitors like ASIC and AMD will likely become even more pronounced.
- Recent production issues with the GB200 NVL72 make this a good time to reset expectations for scaling laws and AI server shipments. This stock correction will be more favorable for positive future reactions to the GB300/Rubin series.
- Leading CSPs won't reduce capital expenditure simply because of improved training approaches, as
efficiency gains and infrastructure investments can coexist. Any reduction in capital spending now could leave them vulnerable to falling behind competitors when scaling law benefits accelerate again. - The combined effect of open-source resources and the highly competitive environment in China is
expected to lead to other Chinese companies launching high-performing LLMs with even more
aggressive pricing. If LLM service providers have not established stable profitability by then, their profit pressures will intensify. - Benefiting from significant API/token price declines, AI software/service and on-device AI will attract
greater investor attention. Whether this becomes a new long-term investment trend depends on whether profitable business models can be created. - Nvidia remains the likely winner when the benefits of scaling laws accelerate again in the future.
However, monitoring short-term GB200 NVL72 production issues and potential changes to US
semiconductor export restrictions in the medium-to-long term is crucial.
$NVIDIA(NVDA.US) $Taiwan Semiconductor(TSM.US)
The copyright of this article belongs to the original author/organization.
The views expressed herein are solely those of the author and do not reflect the stance of the platform. The content is intended for investment reference purposes only and shall not be considered as investment advice. Please contact us if you have any questions or suggestions regarding the content services provided by the platform.