When young, I didn't know how good Su Ma was, mistakenly treating trash as treasure.
At that time, DEEPSEEK was released, and the AI stocks in the US plummeted; AMD even fell to 80. In just a few months, it has reached 300, which really leaves one speechless.
But there's nothing to be done; missing out is all due to lack of understanding. If you don't understand, then learn and practice.
The AI computing power matter has mainly been consumed in "training" over the past three years—OpenAI trained GPT-4, Anthropic trained Claude, Google trained Gemini; these are all training.
The characteristics of training are one-off, concentrated, and peak value.
But every time you ask ChatGPT a question, every time you use Claude to write a piece of code, every time you use Midjourney to generate an image—you are not consuming training computing power, but rather inference computing power.
The characteristics of inference are continuous, distributed, and long tail.
Once the training is completed, the model goes live. Once online, it responds to hundreds of millions of user requests 24 hours a day. Three months later, that bit of computing power consumed for training has become invisible on the overall account—all that remains is inference.
This shift in scale prompts me to illustrate with a comparison.
In 2023, inference accounted for about 20% of AI computing power expenditure; in 2024, this ratio climbs to 50%, and by 2026, it stands above 55%, continuing to rise.
Some more aggressive predictions believe that by 2030, inference will account for 70-80%. Note that this is not because training demand is shrinking—the absolute expenditure on training is still rising, but inference is growing much faster than training.
The true leader on this big slope of inference is NVIDIA. NVIDIA's data center revenue for fiscal year 2026 (ending January 2026) is $194 billion, while two years ago, this figure was less than $50 billion. Such growth has never been seen in semiconductor history. With a CUDA ecosystem of five million developers, built up over twenty years, both training and inference sides are being embraced simultaneously—this is true monopoly.
Ranking first is NVIDIA, second is AMD, and third are Google TPU, Amazon Trainium, Meta MTIA, and these self-developed ASICs—this is today’s table structure.
What position does AMD hold at this table? It is the second chair. This chair is very important—without the second chair, the first chair has no bargaining pressure. But the second chair is not the first chair.
So the real question becomes two sub-questions:
First, can AMD firmly hold this second chair for ten years?
Second, how much is this chair worth once it is secured?
Additionally, AMD has a severely underestimated angle: the true story behind Meta's 170,000 MI300X units.
“AMD Research Report: Looking Back Over 10 Years, Is $300 Expensive?”
https://mp.weixin.qq.com/s/jOLAESOTfEdm3a4Xcxc-oA
免责声明:本文章仅代表作者个人观点,不代表本平台的立场和观点。本文章仅供信息分享,不构成对任何人的任何投资建议。用户与作者之间的任何争议,与本平台无关。如网页中刊载的文章或图片涉及侵权,请提供相关的权利证明和身份证明发送邮件到support@aicoin.com,本平台相关工作人员将会进行核查。