Jensen Huang projects $1 trillion in Nvidia hardware revenue through 2027 at GTC

    Nvidia CEO Jensen Huang told investors at the company's GPU Technology Conference that Nvidia's hardware revenue could total $1 trillion through 2027, a projection that lands roughly $50 billion to $70 billion above current Wall Street consensus estimates for the same period. The number is large enough that it requires some unpacking. Huang is not projecting a single year of $1 trillion in revenue. He is projecting cumulative hardware revenue across the full 2025 through 2027 window, driven by what he described as accelerating demand for AI inference workloads.

    Inference is the part of AI that actually runs in production. Training a large language model requires enormous compute for a finite period, but inference, running that model millions or billions of times daily to serve queries, requires continuous hardware investment at scale. Huang's argument is that the inference demand from deployed AI applications is growing faster than most financial models assumed, and that Nvidia's chips are the primary hardware capturing that spending.

    The gap between Huang's projection and Wall Street estimates

    Current consensus estimates compiled by Bloomberg put Nvidia's hardware revenue at approximately $230 billion for fiscal year 2026 and $260 billion for fiscal year 2027, totalling roughly $490 billion across those two years. Adding fiscal year 2025 revenue of approximately $130 billion brings the three-year cumulative total to around $620 billion under current consensus. Huang's $1 trillion projection implies that actual revenues will run about $380 billion higher than what analysts are currently modelling, spread across the same period.

    That is a significant divergence. Nvidia has beaten consensus revenue estimates by meaningful margins in recent quarters, including a 22% beat in its fiscal Q3 2025 earnings report, so the market has learned not to dismiss Huang's projections out of hand. Whether this specific number holds depends heavily on how quickly enterprise AI inference spending scales and whether competitors can take meaningful market share before 2027.

    Nvidia GPU and AI chip technology
    Nvidia GPU and AI chip technology

    What Huang announced about new hardware at GTC

    Huang unveiled the Nvidia Groq 3 Language Processing Unit at the conference. The LPU is specifically designed for inference workloads rather than training, with a chip architecture optimised for the sequential token generation that large language models require at runtime. Groq's original LPU design had demonstrated significantly faster inference throughput than GPU-based alternatives in benchmark tests, and Nvidia's version is its direct answer to that competition entering the inference hardware segment.

    Huang also confirmed that the Vera Rubin chip platform is expected to debut later in 2026. Vera Rubin is the successor to the Blackwell architecture currently shipping to hyperscale customers. Based on Nvidia's previous product cycle timelines, Vera Rubin will represent roughly a 2x to 3x performance improvement per chip for training workloads and a similar gain for inference throughput. The platform is named after the astronomer Vera Rubin, continuing Nvidia's practice of naming major chip architectures after scientists.

    Why inference demand is the central argument

    The AI hardware market is in the middle of a structural shift. Through 2023 and most of 2024, the primary driver of GPU purchases was training: building the large foundation models that now power most commercial AI products. That spending is not going away, but it is maturing. The hyperscalers, Microsoft, Google, Amazon, and Meta, have built out substantial training capacity. The next wave of spending is inference, which scales with user adoption of AI products rather than with model development cycles.

    OpenAI's ChatGPT processes an estimated 10 million queries per day. Google processes more than 8.5 billion searches daily, a growing fraction of which now involve AI-assisted responses. Every one of those queries requires compute. As the fraction of digital interactions that involve AI inference increases, the total compute demand grows continuously rather than in the concentrated bursts that training cycles create. That is the market dynamic Huang is projecting Nvidia to capture.

    Competitive risks to the projection

    Huang's $1 trillion projection assumes Nvidia maintains something close to its current market position. That is not guaranteed. AMD's MI300X accelerator has gained meaningful traction with enterprise customers, particularly those running open-source models like Meta's Llama series that do not require Nvidia's proprietary CUDA software stack. AMD's data centre revenue grew 94% year over year in Q4 2025, reaching $4.7 billion in a single quarter.

    Custom silicon from the hyperscalers is a separate constraint. Google's TPU v5, Amazon's Trainium 2, and Microsoft's Maia 100 are all designed to run specific inference workloads internally, reducing those companies' dependence on external GPU purchases. If the hyperscalers shift 20% of their inference workloads to internal chips by 2027, the impact on Nvidia's revenue at the scale Huang is projecting would be in the range of $60 billion to $80 billion. The $1 trillion number assumes that internal chip adoption does not accelerate beyond current trajectories.

    Nvidia shares rose 4.2% on the day of Huang's GTC keynote, adding approximately $110 billion in market capitalisation. The next earnings report, covering fiscal Q1 2027, is scheduled for May 28, 2026, and will be the first opportunity for analysts to test whether actual revenue is tracking toward Huang's projection.

    Love this story? Explore more trending news on nvidia

    Share this story

    Frequently Asked Questions

    Q: Is Jensen Huang projecting $1 trillion in a single year or cumulatively?

    Huang is projecting cumulative hardware revenue across the full 2025 through 2027 window, not a single-year figure. The projection implies Nvidia will generate roughly $380 billion more than current Wall Street consensus estimates across that three-year period.

    Q: What is the Nvidia Groq 3 LPU and how is it different from a GPU?

    The Groq 3 Language Processing Unit is designed specifically for AI inference workloads rather than training. Its architecture is optimised for the sequential token generation that large language models require when responding to user queries, which can deliver faster throughput for that specific task compared to a general-purpose GPU.

    Q: When is the Vera Rubin chip platform expected to launch?

    Huang confirmed at GTC that Vera Rubin is expected to debut later in 2026. It succeeds the Blackwell architecture currently shipping to hyperscale customers and is expected to deliver roughly 2x to 3x performance improvement per chip compared to its predecessor.

    Q: How much has AMD grown in data centre chips and does that threaten Nvidia's projection?

    AMD's data centre revenue grew 94% year over year in Q4 2025, reaching $4.7 billion in a single quarter. AMD's MI300X has gained traction with customers running open-source models that do not require Nvidia's CUDA software stack, which is the primary competitive pressure on Nvidia's market share heading into 2027.

    Q: Why does inference demand matter more than training for Nvidia's long-term revenue?

    Training spending happens in concentrated bursts as companies build foundation models, and the major hyperscalers have largely completed their initial training buildouts. Inference spending scales continuously with user adoption of AI products and does not plateau the same way. As more digital interactions involve AI-generated responses, inference compute demand grows every quarter regardless of whether new models are being trained.

    Read More