NVDA — NASDAQ
$177.39
Last close • April 4, 2026
Market Cap
$4.33T
52-Week Range
$86.62 – $212.19

On Christmas Eve 2025, nvidia-stock/”>Nvidia quietly signed a $20 billion licensing deal with AI inference startup Groq that most of Wall Street initially shrugged off as another routine tech partnership. Three months later, the deal has become the most scrutinized transaction in semiconductor history, drawing a Senate investigation, FTC attention, and a fierce debate about whether the arrangement is a genuine licensing agreement or a cleverly disguised acquisition built to dodge antitrust review.

Here is what actually happened, why it matters for NVDA shareholders, and what the regulatory fallout could mean for the stock in 2026 and beyond.

What Nvidia Actually Bought (and What It Didn’t)

The deal announced on December 24, 2025, is structured as a non-exclusive licensing agreement, not a traditional acquisition. Under the terms, Nvidia paid $20 billion to license Groq’s inference chip architecture, specifically the Language Processing Unit (LPU) technology that had made Groq the fastest inference provider in the industry.

But here is where it gets complicated. Nvidia also hired Groq’s founder and CEO Jonathan Ross, president Sunny Madra, and the majority of Groq’s engineering talent. Groq technically remains an independent company under new CEO Simon Edwards, retains its intellectual property, and continues operating GroqCloud, its inference-as-a-service platform.

Critics argue this structure is an acqui-hire dressed up as a licensing deal. Supporters counter that Groq remains independent with its IP intact. The truth, as usual, sits somewhere in the middle, and that ambiguity is exactly what has regulators concerned.

Deal Structure at a Glance

ComponentWhat Nvidia GetsWhat Groq Retains
IP LicenseNon-exclusive rights to LPU architectureFull IP ownership; can license to others
Key PersonnelCEO Jonathan Ross, President Sunny Madra, core engineering teamNew CEO Simon Edwards; remaining ops staff
Physical AssetsManufacturing rights, hardware designsGroqCloud infrastructure and Middle East contracts
Valuation$20B total (2.9x Groq’s prior $6.9B valuation)Independent entity status
Antitrust FilingNo Hart-Scott-Rodino filing was made (licensing deals are exempt)

Why Groq’s LPU Technology Changes the Inference Game

To understand why Nvidia paid $20 billion for a licensing deal, you need to understand what Groq’s LPU actually does differently.

Traditional GPUs, including Nvidia’s own H100 and Blackwell chips, use High Bandwidth Memory (HBM) stored off-chip. Every time the processor needs data, it has to reach out to external memory, creating a bottleneck. Groq’s LPU flips this model entirely by using on-chip SRAM as its primary working storage.

The numbers tell the story: Groq’s on-chip SRAM delivers roughly 150 TB/s of memory bandwidth per chip, compared to about 22 TB/s for a Vera Rubin GPU. That is nearly seven times faster. The result is inference speeds of 500 to 750 tokens per second versus roughly 100 tokens per second on comparable GPU setups. Energy consumption drops to 1 to 3 joules per token compared to 10 to 30 joules per token on GPU-based inference.

The tradeoff is capacity. Each LPU holds only 500 MB of SRAM versus the multiple gigabytes of HBM on a GPU. That means LPUs cannot handle the computationally intensive prefill and attention phases of large language model inference on their own. They excel at the decode phase, where low-latency, high-throughput token generation is the bottleneck.

Nvidia recognized this complementarity. Rather than competing against Groq, it made more sense to bolt LPU technology onto its existing GPU platform and let each architecture handle what it does best.

LPU vs. GPU: The Inference Performance Gap

MetricGroq LPU (SRAM)Traditional GPU (HBM)Advantage
On-Chip Memory Bandwidth150 TB/s per chip~22 TB/s per chipLPU ~7x faster
Token Generation Speed500–750 tok/s~100 tok/sLPU 5–7x faster
Energy per Token1–3 joules10–30 joulesLPU ~10x efficient
On-Chip Memory Capacity500 MB SRAMMultiple GB HBMGPU far larger
Best Use CaseDecode phase, low-latency servingPrefill, attention, trainingComplementary

Groq 3 LPU: The First Product from the Deal

Nvidia wasted no time. At GTC 2026 in March, Jensen Huang unveiled the Groq 3 LPU, the first chip to emerge from the partnership. Manufactured by Samsung on a 4nm process, the Groq 3 slots into the Vera Rubin platform as a dedicated decode-phase co-processor.

The system architecture uses what Nvidia calls Attention-FFN Disaggregation (AFD). The Vera Rubin GPUs handle prefill, KV cache creation, and full-context attention operations. The Groq 3 LPUs take over for the latency-sensitive feed-forward networks, MoE expert execution, and pointwise operations where SRAM bandwidth dominance matters most.

Groq 3 LPX Rack Specifications

SpecificationLPX Rack (256 LPUs)Per Tray (8 LPUs)
FP8 Compute315 PFLOPS9.6 PFLOPS
Total SRAM128 GB4 GB
SRAM Bandwidth40 PB/s1.2 PB/s
Scale-Up Bandwidth640 TB/s20 TB/s
Chip-to-Chip Links96 links per device at 112 Gbps (2.5 TB/s per chip)
Configuration32 liquid-cooled 1U trays8 Groq 3 LPU accelerators
DRAM ExpansionUp to 384 GB per tray (256 GB fabric + 128 GB host)

The headline performance claim from Nvidia’s GTC 2026 keynote: the LPX rack paired with a Vera Rubin NVL72 delivers 35 times higher inference throughput per megawatt than Blackwell NVL72 alone for trillion-parameter models, at a target of $45 per million tokens.

Samsung has already begun mass production on its 4nm process, with shipments expected in Q3 2026. This is a remarkably fast timeline, just nine months from deal announcement to silicon shipping.

The Senate Investigation: What Warren and Blumenthal Want

On March 20, 2026, Senators Elizabeth Warren (D-MA) and Richard Blumenthal (D-CT) sent a letter to Jensen Huang raising pointed questions about whether the Groq deal was deliberately structured to evade antitrust review.

Their core argument: Nvidia controls roughly 90% of the GPU market. Groq was one of the few credible competitors in AI inference, with an architecture that was genuinely faster and more energy-efficient for certain workloads. By licensing Groq’s technology and hiring its leadership team while technically leaving the company independent, Nvidia may have found a loophole to consolidate its dominance without triggering the Hart-Scott-Rodino premerger notification requirements.

The senators set an April 3, 2026, deadline for Nvidia to respond to their questions. They also urged the DOJ and FTC to open formal investigations.

Key Regulatory Questions

QuestionWhy It Matters
Was the deal structured to avoid HSR filing?Licensing deals are exempt from premerger notification; acquisitions are not
Does hiring the CEO + core team constitute de facto control?If yes, the deal could be reclassified as an acquisition retroactively
Can Groq realistically compete after losing its founder and top engineers?If Groq is effectively hollowed out, the “independent company” argument collapses
Is the non-exclusive license meaningfully non-exclusive?If no other company licenses LPU tech, exclusivity exists in practice
Does this fit the “reverse acqui-hire” pattern?FTC is already investigating Microsoft-Inflection for the same structure

This investigation sits alongside broader regulatory scrutiny of Big Tech acqui-hires. In January 2026, FTC Chair Andrew Ferguson announced the agency would investigate these types of deals across the tech industry. The Microsoft-Inflection AI deal is already under active FTC investigation for being what regulators call a “merger in disguise.”

What This Means for NVDA Stock

Nvidia shares closed at $177.39 on April 4, with a market cap of $4.33 trillion. The 38 analysts covering the stock maintain a consensus Strong Buy rating with an average price target of $266, implying roughly 50% upside from current levels. The highest target sits at $360, while the most bearish estimate is $100.

The Groq deal creates both significant upside and measurable risk for shareholders. Investors evaluating the best AI stocks should weigh this deal’s implications carefully.

The Bull Case

Inference is where the money goes next. Training dominated the first wave of AI infrastructure spending, but as models mature, inference workloads are expected to account for 60 to 80 percent of total AI compute spending by 2028. Nvidia buying its way into purpose-built inference hardware positions it to capture both sides of the market.

The 35x efficiency claim is a competitive moat. If the Groq 3 LPX delivers anywhere close to 35 times the throughput per megawatt of Blackwell, hyperscalers building out inference infrastructure will have a hard time justifying AMD or custom ASIC alternatives. The $45 per million tokens target is aggressive enough to pressure every competitor in the inference space.

Samsung manufacturing diversifies supply chain risk. Nvidia’s GPU production is concentrated at TSMC. Having Samsung produce the Groq 3 on 4nm gives Nvidia a second manufacturing partner and reduces its dependence on a single foundry, a concern that has grown louder amid Taiwan Strait tensions.

Full-stack dominance. At GTC 2026, Nvidia introduced three new systems simultaneously: the Groq LPX inference rack, the Vera ETL256 CPU rack, and the STX storage reference architecture. Combined with existing GPU leadership, Nvidia now offers the complete AI data center stack from training to inference to storage. No competitor matches this breadth.

The Bear Case

Regulatory overhang is real. If the FTC reclassifies the deal as an acquisition, Nvidia could face an unwinding order, forced divestiture of the license, or substantial fines. The Microsoft-Inflection precedent suggests regulators are willing to look past deal labels and focus on economic substance.

$20 billion is a lot for a licensing deal. Groq’s last private valuation was $6.9 billion (per CNBC). Paying 2.9 times that for a non-exclusive license raises questions about capital allocation, especially when Nvidia could have invested that money in developing its own SRAM-based inference architecture internally.

The acqui-hire label invites ongoing scrutiny. Even if the FTC does not force an unwind, the regulatory cloud creates uncertainty for institutional investors. ESG-focused funds and governance-conscious allocators may trim NVDA positions until the legal picture clarifies.

AMD and custom silicon are not standing still. AMD’s MI450 accelerators have secured major commitments from OpenAI. Google continues developing its TPU inference stack. Amazon’s Inferentia chips are gaining traction in AWS workloads. The inference market is not winner-take-all the way training has been.

NVDA Analyst Snapshot

MetricValue
Consensus RatingStrong Buy (38 analysts)
Average Price Target$266
Highest Target$360
Lowest Target$100
Implied Upside (from $177.39)~50%
Current Market Cap$4.33 Trillion

Competitive Landscape: Who Loses if Groq 3 Delivers

The Groq deal does not exist in a vacuum. Nvidia’s move into purpose-built inference hardware reshuffles the competitive dynamics across the entire AI chip industry.

CompetitorInference ProductThreat Level from Groq 3Why
AMD (AMD)MI450 InstinctHighMI450 is GPU-based inference; 35x efficiency gap would be devastating if validated
Google (GOOGL)TPU v6 TrilliumMediumTPUs are captive to Google Cloud; less exposed to merchant silicon market
Amazon (AMZN)Inferentia3MediumInferentia serves AWS workloads; Groq 3 competes in colo and on-prem
Cerebras (Private)CS-3 WSEMedium-HighWafer-scale approach targets similar large-model inference workloads
Intel (INTC)Gaudi 3LowIntel is focused on training and cost-sensitive inference; different market segment

The most vulnerable player is AMD. Its MI450, while a capable GPU-based inference accelerator, faces an architectural disadvantage against SRAM-based designs for latency-sensitive decode workloads. AMD secured a major OpenAI partnership, but if hyperscalers start adopting Groq 3 LPX racks for their inference fleets, AMD’s inference market share could erode faster than the market currently prices in.

Timeline: What Happens Next

DateEventImpact
April 3, 2026Nvidia’s deadline to respond to Warren/Blumenthal letterResponse tone will signal legal strategy
Q2 2026FTC expected to announce whether formal investigation proceedsBinary event for NVDA sentiment
Q3 2026Groq 3 LPU shipments begin (Samsung 4nm)First real-world performance validation
Q3 2026Nvidia FY2027 Q2 earningsFirst quarter with Groq-related revenue guidance
H2 2026Hyperscaler inference deployment decisionsDetermines whether 35x claim translates to purchase orders
2027+Potential Senate hearings on acqui-hire regulationCould reshape M&A rules across tech sector

The Verdict: A Brilliant Move Wrapped in Legal Risk

The Nvidia-Groq deal is one of the most strategically significant transactions in semiconductor history. By licensing LPU technology and hiring the team that built it, Nvidia positioned itself to dominate both training and inference, the two pillars of AI compute, with purpose-built hardware for each.

The Groq 3 LPU’s specifications are genuinely impressive. 315 PFLOPS of FP8 compute, 40 PB/s of SRAM bandwidth, and a 35x efficiency improvement over Blackwell for trillion-parameter inference are the kind of numbers that force competitors to rethink their roadmaps. Samsung manufacturing on 4nm with Q3 2026 shipments means this is not vaporware; it is shipping silicon.

But the regulatory risk is not trivial. The Warren-Blumenthal investigation has legs. The FTC’s broader crackdown on acqui-hires, the Microsoft-Inflection precedent, and Nvidia’s existing 90% GPU market share all create a legal environment where this deal faces genuine scrutiny. An adverse ruling could force Nvidia to unwind the license, divest the technology, or face penalties that temporarily pressure the stock.

For investors, the calculus comes down to time horizon. If you believe the Groq 3 delivers on its performance promises and the regulatory investigation ultimately resolves favorably, this deal transforms Nvidia from a GPU company into the undisputed AI infrastructure platform company. If regulators force a restructuring, the $20 billion investment becomes a write-down risk.

At $177.39 with a $266 consensus price target, NVDA trades at a roughly 50% discount to Wall Street’s expectations. The Groq deal is a significant reason analysts remain bullish. Whether the market agrees will depend on what comes out of Q3 2026: working silicon and regulatory clarity, or Congressional subpoenas and FTC complaints.

Disclosure: This article is for informational purposes only and does not constitute investment advice. The author and TECHi may hold positions in securities mentioned. Always conduct your own research or consult a licensed financial advisor before making investment decisions. Stock prices and analyst targets are as of the date published and subject to change.