Building on my post from yesterday....
in this post I would like to write more about the transition from AI infrastructure expansion to the inference market, taking into account my personal assessment and using some sources. The stocks mentioned do not constitute investment advice. As always, the lines are intended to provide impulses to understand the potential winners of tomorrow.
![attachment](https://static.getquin.com/thumbnails/ceeb16ba5188b0f8f61832379ee2cb82.jpeg)
In recent years and currently, a large proportion of investment has gone into building the infrastructure for artificial intelligence.
Tech giants and cloud providers have built huge data centers to enable the training phase of AI models such as GPT-4 or Gemini [1].
However, a paradigm shift is now emerging: the focus is shifting from expensive training to the inference market, i.e. the practical use and monetization of AI models, as general AI models will eventually become a "commodity".
When we say that AI is becoming a commodity, this means that artificial intelligence is increasingly becoming a standard technology that is readily available and accessible to many companies and industries, similar to electricity, the internet or cloud computing today.
Standardization and availability
- AI solutions are being standardized to such an extent that companies can use them without in-depth specialist knowledge.
- Instead of creating individually developed AI systems, companies are turning to ready-made solutions or platforms (e.g. AI models, APIs or services from companies such as OpenAI, Google or Microsoft).
Lower costs
- Economies of scale and technological advances are making AI systems cheaper and therefore affordable for more companies.
- Cloud providers such as AWS, Azure or Google Cloud offer AI tools as pay-as-you-go services, which lowers the barriers to entry for smaller companies.
Wide range of applications
- AI is used in all kinds of industries, from logistics and finance to healthcare and agriculture.
- Technologies such as image and speech processing, automated decision-making and data analysis are becoming the basis for innovation.
Loss of competitive advantages
- When AI is equally available to all companies, those that used to be leaders will lose their unique competitive advantage.
- The difference will no longer lie in the use of AI itself, but in the efficient and creative application of the technology.
Long-term significance
- AI as a commodity will lead to a democratization of technology where every company has access to the same basic tools.
- Competition will shift to who uses AI better, rather than who owns or develops it.
- At the same time, new differentiators could emerge, such as industry-specific customization or unique data.
In practice, this means that AI is no longer the exclusive domain of tech giants, but a tool that anyone can use to drive innovation.
📈 Why the shift to the inference market is inevitable
From training to inference: What's the difference?
Training:
AI models are trained with enormous amounts of data, a computationally intensive process that requires specialized hardware such as GPUs and TPUs. Companies such as Nvidia, AMD and Broadcom are benefiting greatly from the boom in data centers at this stage.
Inference:
Once training is complete, the models must be used in the real world, whether through chatbots, voice assistants or image processing. Efficiency is key here, as AI has to respond to millions of requests in real time.
Capital shift: the money follows the monetization
Infrastructure development completed:
Building training infrastructure is costly, but after a certain phase, the focus shifts to optimization and more efficient use over frugal training or specialized chips.
Energy efficiency and cost optimization:
Companies are looking for solutions that are not only powerful but also energy-efficient and cost-effective, a crucial factor in maximizing margins.
🔮 The growth market of the future
The inference market is expanding rapidly as AI applications penetrate more and more areas of life and work:
Automated customer service:
Chatbots and virtual assistants are replacing traditional call centers and offer round-the-clock support (e.g. IBM watsonx Assistant) [2].
Medical diagnostics:
AI-supported image analyses improve diagnostics and enable personalized therapies [3].
Language models & generative AI:
Applications such as ChatGPT or Google Bard are revolutionizing the field of generative AI.
Industry 4.0 & automation:
Smart factories and predictive maintenance increase efficiency in production [4].
🏆 The profiteers of the inference market - how are the Big Techs positioned?
📌 AMD
$AMD (+0,11 %) Nvidia's challenger in the inference sector
Technology:
- AMD's MI300X is considered a direct competitor to Nvidia's H100 and offers advantages in terms of energy efficiency and costs [5].
Market outlook:
- With a strong positioning in the cloud and data center sector, AMD could gain market share in the growing inferencing market
Nvidia $NVDA (-1,24 %) From training king to inference king?
Dominance in the training market:
- Nvidia has dominated the training sector to date with its A100 and H100 GPUs [6].
Switch to inference mode:
- With new software optimizations such as TensorRT and specialized hardware solutions such as the Grace Hopper superchips, Nvidia is preparing to be a leader in the inference sector as well.
Google
$GOOGL (-0,96 %) TPUs as the inference solution of the future
Own AI chips:
- Google relies on its own Tensor Processing Units (TPUs) to reduce inference costs in the cloud [7].
Range of applications:
- Optimized inference chips play an essential role in Google Cloud as well as YouTube AI recommendations.
Amazon
$AMZN (-0,61 %) The cloud giant with its own AI hardware
Inferentia chips:
- Amazon's specially developed Inferentia chips are designed for the cost-effective operation of AI inference solutions [8].
Market penetration:
- With a steadily growing number of AWS customers using customized AI inference solutions, Amazon remains a key player in the AI sector.
Microsoft
$MSFT (-0,71 %) Profiteer through OpenAI & Azure AI
Partnership with OpenAI:
- The close cooperation with OpenAI positions Microsoft as one of the largest providers of inference solutions.
Azure AI:
- With Azure AI, Microsoft is building a powerful cloud infrastructure for generative AI.
Future developments:
- In-house AI chips could reduce dependence on external providers such as Nvidia in the long term.
Broadcom
$AVGO (-0,04 %) The "hidden champion" for inference networks
Network technology:
- Broadcom's high-speed networking technology enables the fast data processing that is critical for AI inference.
Market Leadership:
- As a leading provider of networking chips in hyperscale data centers, Broadcom benefits from the growing demand for AI-optimized solutions.
Qualcomm
$QCOM (-0,12 %) Inference on the edge market
Edge Computing:
- In addition to the cloud, inference is also needed on mobile devices and IoT systems. Qualcomm offers powerful, energy-efficient AI chips for this.
Smartphone applications:
- In particular, the use of Snapdragon AI processors in smartphones makes Qualcomm an important player in the edge sector.
🎯 Conclusion: the inference market could be the new AI gold rush
The AI boom is far from over, it's just shifting. While the expansion of infrastructure laid the foundation, the focus is now on continuous revenue from inference solutions.
- Companies that specialize in energy-efficient and high-performance inference hardware have long-term growth potential.
- Providers with their own AI solutions can benefit massively by reducing costs and improving margins.
- Companies like Broadcom provide the technological foundation for fast and reliable AI inference.
The inference market has long been a reality and is not a short-term phenomenon.
❓Which companies do you see benefiting alongside the big players?
In yesterday's post, I briefly outlined my personal portfolio breakdown on the topic of the AI revolution.
Thanks for reading! 🤝
__________
Sources:
[1]
https://datacentremagazine.com/it/gartner-why-global-it-spending-will-hit-us-5-61tn-in-2025
[2]
[3]
https://www.mckinsey.com/featured-insights/themes/whats-next-in-ai-and-healthcare
[4] https://www.deloitte.com/de/de/issues/innovation-ai/industrie-40.html
[5] https://www.amd.com/de/products/accelerators/instinct/mi300/mi300x.html
[6]
https://www.nvidia.com/en-us/data-center/
[7]
[8]