Futures
Access hundreds of perpetual contracts
TradFi
Gold
One platform for global traditional assets
Options
Hot
Trade European-style vanilla options
Unified Account
Maximize your capital efficiency
Demo Trading
Introduction to Futures Trading
Learn the basics of futures trading
Futures Events
Join events to earn rewards
Demo Trading
Use virtual funds to practice risk-free trading
Launch
CandyDrop
Collect candies to earn airdrops
Launchpool
Quick staking, earn potential new tokens
HODLer Airdrop
Hold GT and get massive airdrops for free
Pre-IPOs
Unlock full access to global stock IPOs
Alpha Points
Trade on-chain assets and earn airdrops
Futures Points
Earn futures points and claim airdrop rewards
Just caught something significant from Jensen Huang during Nvidia's latest earnings call that most people might be glossing over. The CEO basically dropped a bombshell on what AI infrastructure spending actually looks like over the next few years, and it reframes the entire growth narrative.
Here's what jumped out: Huang pointed out that the world has historically spent around $400 billion annually on classical computing infrastructure. But when you factor in what AI workloads actually require, he's saying we're talking about a thousand times more capacity. That's not incremental growth - that's a complete paradigm shift in how much computing power needs to get deployed.
The timing is interesting because Nvidia is literally about to ship its next-generation Vera Rubin platform starting in the second half of this year. This isn't just another GPU refresh. The Rubin architecture is so efficient that AI models can be trained with 75% fewer GPUs compared to the current Blackwell generation, and inference token costs drop by 90%. When you're running AI services at scale, that kind of cost reduction changes the entire economics of the business.
What makes this relevant right now is the sheer magnitude of the opportunity. Huang previously estimated AI data center spending could hit $4 trillion annually by 2030. That sounded aggressive back then, but if you actually do the math on the capacity requirements he's describing, it starts looking more realistic. Especially if token costs keep falling and usage accelerates.
On the valuation side, Nvidia's trading at a forward P/E of 21.5 based on fiscal 2027 earnings estimates, which is actually cheaper than the S&P 500's current multiple of 24.7. The stock's P/E of 36.1 sits 41% below its 10-year average of 61.6. Wall Street's consensus sees earnings growing to $8.23 next year from the $4.77 the company just posted.
I'm not making any predictions here, but if those earnings estimates prove accurate and the stock doesn't move, it would need to jump 186% just to get back in line with historical valuations. Even partial upside from that scenario would be substantial, which is probably why institutional investors keep finding reasons to add positions on any weakness.