Leo
Welcome everyone to this week's episode! I'm Leo, your host, and today we have a very special guest, Eugene Cheah, joining us. We're diving into the fascinating world of GPUs and exploring how the market has evolved, especially focusing on the H100s. It's been quite a ride, hasn't it, Eugene?
Eugene Cheah
Absolutely, Leo! The changes we've seen in the GPU market, particularly with the H100s, are really remarkable. Just a year ago, they were priced at an astonishing $8 an hour, and now you can find them for less than $2. It’s like a rollercoaster ride, and it raises a lot of questions about what’s driving these shifts.
Leo
Right! The initial demand spike was driven by the hype around AI and the push for more powerful models. Companies wanted to keep up with the competition, and the H100s promised a significant performance boost over previous generations. It seemed like a gold rush for investors, and suddenly everyone was pouring money into GPU-rich startups.
Eugene Cheah
Exactly, Leo. It was a classic case of supply not keeping up with demand. When the H100s hit the market, they offered incredible capabilities, and the rush to acquire them led to inflated rental prices. It felt like a no-brainer for startups trying to build the next big AI model. But as we’ve seen, that kind of rapid growth can lead to instability.
Leo
And then came the correction phase. As more players entered the market and supply began to exceed demand, we saw those prices plummet. It's fascinating how quickly the market can shift from scarcity to oversupply. What do you think are the primary factors that contributed to this dramatic price drop?
Eugene Cheah
A few key factors come to mind, Leo. First, we had a significant increase in compute resales and the emergence of open-source models that allowed companies to fine-tune existing models instead of training from scratch. This approach requires far less compute power, which directly impacts the demand for high-end GPUs like the H100s.
Leo
That's a great point. The shift towards open-source models is a double-edged sword. While it democratizes access to powerful AI tools, it also reduces the need for expensive hardware. Do you think this trend is here to stay?
Eugene Cheah
I believe so, Leo. The community around open-source models is growing rapidly, and more organizations are realizing the benefits of fine-tuning existing models rather than starting from scratch. Plus, as these models become more sophisticated, the gap between open-source and proprietary models is narrowing, which further supports this trend.
Leo
It sounds like the rental market is changing as well. With H100 prices dropping to below $2, what does this mean for companies considering investing in GPU infrastructure?
Eugene Cheah
For many companies, it now makes more financial sense to rent GPUs rather than purchase them outright. Given the rapid depreciation of these assets, it's likely that companies could find better returns by renting as needed instead of committing to a large capital expenditure upfront. This flexibility allows them to scale up or down based on current needs.
Leo
Ultimately, it seems like we are witnessing a significant transformation in the AI landscape which is going to affect how businesses plan their AI strategies moving forward. What are your thoughts on how this might evolve in the next few years?
Eugene Cheah
As we look ahead, I think we'll see an even greater reliance on open-source models and managed services for AI inference. This shift could lead to a more competitive landscape where companies prioritize innovation and application over raw compute power. The focus will likely be on how to leverage these models effectively rather than just having the latest hardware.
Leo
Host
Eugene Cheah
Industry Expert