Alphabet's Google is in talks with Marvell Technology to co-develop two new chips aimed at running AI models more efficiently, The Information reported on Sunday, , citing two people with knowledge of the discussions. According to the report, one chip is a memory processing unit designed to work alongside Google's existing tensor processing units, and the second is a next-generation TPU built specifically for running AI models. The companies aim to finalize the memory processing unit's design as soon as next year before handing it off for test production, the sources told The Information. Google and Marvell did not respond to requests for comment.

The talks matter for three reasons at once. They extend Google's multi-year effort to build a credible alternative to Nvidia's dominant GPU stack. They hand Marvell a marquee design win in a sector where every new AI accelerator program looks like a valuation catalyst. And they sharpen a specific architectural bet that most hyperscalers are quietly running in parallel: that the next performance leap in AI systems comes from purpose-built memory components, not just faster compute.

What Google Is Actually Trying to Build

The memory processing unit (MPU) in the reported design is the more interesting chip of the two. AI workloads, particularly inference at production scale, are increasingly memory-bound rather than compute-bound. Moving model weights and attention states between high-bandwidth memory and compute cores costs energy, time, and silicon area. A dedicated memory processor that sits between Google's TPUs and the memory subsystem could offload data movement, compression, and caching operations that currently burn a meaningful share of the power budget.

The second chip is a new generation of Google's tensor processing unit. Google has been shipping TPUs since 2016 and has used them internally to train and serve Search, Gmail spam filtering, Translate, and the Gemini model family. What the new TPU design adds, per The Information's framing, is purpose-built optimization for running models rather than training them. Inference is where Google's cloud margin case lives, because serving billions of queries per day against a frozen model requires different tradeoffs than the high-flops training pipelines that dominated the first TPU generations.

Google TPU generation snapshot
GenerationYearPrimary workload
TPU v12016Inference
TPU v2 / v32017-2018Training + inference
TPU v42020Large-scale training
TPU v5e / v5p2023-2024Efficient inference / training
Trillium (v6)2025Gemini-scale training
New TPU w/ Marvell2026-2027 (projected)Model serving
Approximate Google TPU evolution with the reported Marvell partnership projected at the next inflection point.

A Marvell partnership is a departure from Google's recent custom-chip playbook. Broadcom has been Google's primary TPU co-design partner since the TPU v4 generation, and the two companies have collaborated on packaging, interconnects, and networking silicon. Bringing Marvell into the equation suggests either a deliberate diversification of supply-chain risk, a specific Marvell capability Broadcom cannot match on timeline, or both.

Why Marvell Shares Responded

Marvell shares gained roughly 4.6% in trading on , the session after the report broke. The move is not just about a single design win, though that in itself is material given the revenue profile of hyperscaler AI chips. The broader story for Marvell is that it has spent the past two years repositioning itself as a data-center silicon specialist after divesting legacy consumer and networking businesses. AWS, Microsoft Azure, and Meta have all been reported as Marvell partners on custom AI silicon programs.

"Google has been pushing to make its TPUs a viable alternative to Nvidia's dominant GPUs. TPU sales have become a key driver of growth in Google's cloud revenue as it aims to show investors that its AI investments are generating returns."

Reuters, April 19, 2026

The Google-Marvell tie-up would add a fourth hyperscaler to that list, if confirmed. For analysts covering Marvell, the value of a Google program is roughly proportional to the scale of TPU deployment in Google Cloud plus Google's internal consumption. Both numbers have been accelerating. TPU sales have become one of the cleanest growth drivers in Google Cloud's recent earnings commentary, with Alphabet executives citing AI-related revenue as the primary reason the segment has been narrowing its margin gap with AWS and Azure.

Technology infographic showing Google TPU generations evolution plus Marvell partnership projection for AI memory processing unit and inference chip
Google TPU evolution with Marvell partnership context

What This Actually Threatens for Nvidia

The short version: not much in the next 18 months. Nvidia's GPU dominance is anchored by the CUDA software stack and by a 10-year head start on datacenter-grade GPU architectures. A new TPU plus MPU pairing, co-designed with Marvell, would ship in test quantities in 2027 at earliest, production volumes likely 2028. By then Nvidia will have shipped at least two more generations of Blackwell Ultra successors.

The longer-term version is more interesting. Google's explicit strategy, visible across the last three TPU generations, is to make TPUs the default for Google Cloud AI workloads while keeping Nvidia GPU availability as an option for customers with existing CUDA investments. The higher the share of Google Cloud AI revenue that runs on TPUs, the less Google has to pay Nvidia per AI-workload dollar. Marvell joining the supply chain adds leverage to that negotiation, both operationally and commercially.

The Google-Meta collaboration reported in Reuters in December 2025, which centered on building open-source software tooling to reduce CUDA lock-in, is the software counterpart to the Marvell hardware partnership. Neither move is a frontal assault on Nvidia. Both are moves to make the long-term Google stack self-sufficient in a way that preserves optionality.

The Industry Context

The Marvell talks are landing in the middle of what has become the most active US semiconductor corporate calendar in a decade. Cerebras filed to go public on . SpaceX, Anthropic, and OpenAI are preparing their own public listings, according to The New York Times. Bloomberg reported on that China's cheap AI models are creating new winners in Chinese tech stocks, including chip designers less constrained by US export controls.

The AI chip market is no longer a binary Nvidia-versus-everyone conversation. It is becoming a stratified landscape with hyperscaler in-house programs (Google TPU, AWS Trainium, Meta MTIA, Microsoft Maia), independent challengers (Cerebras, Groq, SambaNova, Tenstorrent), Nvidia's continued market leadership, AMD's catch-up efforts, and China's domestic substitution programs. The Google-Marvell partnership, if it ships on roughly the reported timeline, fits inside that broader multi-vendor reality.

Technology data visualization showing AI accelerator market landscape Nvidia dominance hyperscaler custom chips and independent challengers in 2026
AI accelerator landscape entering 2026 design cycle

What Still Has to Happen

Three things need to go right for the reported partnership to matter in the market. First, Marvell and Google need to finalize the MPU design and hand it to test production on the reported timeline. Silicon design cycles slip. Every hyperscaler AI program that has shipped on time is the exception, not the rule. Second, Google's TPU software stack needs to keep compiling models as efficiently as it does on Trillium. Any performance regression in the new TPU generation undermines the architectural argument for it. Third, the MPU needs to deliver a measurable energy-per-inference improvement at realistic batch sizes. The memory-processing approach is theoretically compelling. The empirical evidence will come from benchmarks on production workloads.

None of those conditions are trivial. All of them are achievable on the reported schedule if the engineering teams execute. The industry will learn which direction this tilts over the next 18 months as first silicon comes back from test production.

Frequently Asked Questions

What are Google and Marvell building together?

Per The Information's April 19 report, they are co-developing two chips: a memory processing unit designed to work with Google's existing tensor processing units, and a new-generation TPU built specifically for running AI models. The design work on the memory processing unit is targeted to finalize as soon as next year.

Why is Google partnering with Marvell on AI chips?

Google has been pushing to make its TPUs a viable alternative to Nvidia's GPUs, with TPU sales becoming a key growth driver for Google Cloud revenue. A Marvell partnership adds a second major silicon design partner alongside Broadcom and diversifies the supply chain.

Does this threaten Nvidia's GPU dominance?

Not immediately. The reported chips would not reach volume production until 2028 at earliest. Nvidia's CUDA software stack remains the dominant AI development ecosystem. The longer-term question is how much of Google Cloud's AI workload shifts from Nvidia GPUs to custom silicon as the new chips ship.

How did Marvell stock react?

Marvell shares gained roughly 4.6% on April 20 following the report. Analysts interpret the reported partnership as a positive incremental addition to Marvell's custom data-center silicon business, which already includes work with AWS, Microsoft, and Meta.

What is a memory processing unit?

A memory processing unit is a chip designed to handle data movement, compression, and caching operations between compute cores and memory subsystems. In AI workloads, memory bandwidth and latency are often the bottleneck rather than raw compute, making dedicated memory processors a meaningful architectural lever for energy and performance efficiency.

What to Watch

The next confirmation point is whether Google officially acknowledges the partnership on its next earnings call, or whether either company surfaces technical details at a hardware conference in the coming months. Design specifics on the memory processing unit, in particular, will be the most information-rich indicator of how aggressive the architectural departure is. For Marvell, the pattern to watch is how many hyperscaler design wins it can add over 2026 as the industry's custom silicon programs keep multiplying.


Sources