'The CPU is the systems executive layer': Intel joins SambaNova as both face existential threat from Nvidias Groq-powered inference
Date:
Sun, 12 Apr 2026 18:10:00 +0000
Description:
Intel and SambaNova introduce a hardware system combining GPUs, RDUs, and
CPUs to handle inference workloads across execution, decoding, and orchestration tasks.
FULL STORY ======================================================================Copy link Facebook X Whatsapp Reddit Pinterest Flipboard Threads Email Share this article 0 Join the conversation Follow us Add us as a preferred source on Google Newsletter Tech Radar Pro Are you a pro? Subscribe to our newsletter Sign up to the TechRadar Pro newsletter to get all the top news, opinion, features and guidance your business needs to succeed! Become a Member in Seconds Unlock instant access to exclusive member features. Contact me with news and offers from other Future brands Receive email from us on behalf of our trusted partners or sponsors By submitting your information you agree to the Terms & Conditions and Privacy Policy and are aged 16 or over. You are
now subscribed Your newsletter sign-up was successful Join the club Get full access to premium articles, exclusive features and a growing list of member rewards. Explore An account already exists for this email address, please log in. Subscribe to our newsletter GPUs handle prefill operations by converting prompts into key-value caches SambaNova RDUs generate tokens at high throughput and low latency Intel Xeon 6 processors manage workload distribution and execute compiled code Intel and SambaNova Systems have introduced a joint hardware blueprint combining GPUs, SambaNova RDUs, and Intel Xeon 6 processors for large-scale inference workloads.
The system assigns GPUs to prefill operations, RDUs to decoding, and Xeon
CPUs to execution and orchestration tasks across agent-driven environments. Agentic AI is moving into production and the winning pattern were seeing is GPUs to start the job, Intel Xeon 6 to run it, and SambaNova RDUs to finish
it fast, said Rodrigo Liang, CEO and co-founder of SambaNova Systems. Article continues below You may like The AI data centers of 2036 wont be filled with GPUs: FuriosaAIs CEO on the future of silicon Arm introduces its first in-house AGI CPU with backing from Meta and OpenAI AI in networks isnt CPU
vs. GPU: Intel unveils 18A-based Clearwater Forest Xeon 6+ for edge AI and early 6G infrastructure CPU is the execution and control layer This design is scheduled to be available in the second half of 2026 for enterprises, cloud providers, and sovereign deployments.
The architecture places Intel Xeon 6 processors at the center of system control, where they manage workload distribution, execute code, and
coordinate tool interactions.
It includes handling compilation, validating outputs, and maintaining communication between simultaneous processes.
When thousands of simultaneous coding agents are generating tool calls, retrieval requests, code builds, and encrypted inter-agent messages, the CPU is not a background component it is the systems executive and action layer, said Harry Ault, CRO of SambaNova. Are you a pro? Subscribe to our newsletter Sign up to the TechRadar Pro newsletter to get all the top news, opinion, features and guidance your business needs to succeed! Contact me with news
and offers from other Future brands Receive email from us on behalf of our trusted partners or sponsors By submitting your information you agree to the Terms & Conditions and Privacy Policy and are aged 16 or over.
The statement defines the CPU as the primary layer responsible for system behavior rather than a supporting component.
According to SambaNova, Xeon 6 delivers more than 50% faster LLVM compilation times compared with Arm -based server CPUs .
It also delivers up to 70% faster vector database performance compared with other x86-based systems. What to read next Samsung-backed Rebellions launches AI inference racks to compete with Nvidia No Nvidia, No AMD, No Intel, No
ARM: Meta plans inference-led RISC-y future without friends as 1700w
superchip emerges with 30 PFLOPs performance and half Terabyte (yes 512GB)
HBM The new Nvidia age has begun first Vera Rubin AI chips are rolling out
to customers, now let's see what they can do with it
These figures relate to execution speed within coding and retrieval
workflows, and in this configuration, GPUs process the prefill stage by converting prompts into key-value caches.
SambaNova RDUs operate as the decoding layer, generating tokens at high throughput and low latency.
Xeon 6 processors function as both host CPUs and execution engines, managing system-level operations and running compiled workloads.
Production inference is moving toward heterogeneous hardware no single chip type is optimal for every stage of an agentic workflow, said Banghua Zhu, co-founder and CTO at RadixArk.
He added that combining RDUs with Xeon CPUs allows systems to maintain compatibility with existing software environments.
The system is designed to run inside existing air-cooled data centers without requiring new builds.
According to the companies, this allows scaling of inference workloads
without additional strain on water and energy resources.
As Nvidia and Groq continue to focus on improving inference throughput and latency, this announcement adds a layer of competition.
It offers an alternative approach that distributes workloads across multiple hardware layers rather than relying on a single processing model. Follow TechRadar on Google News and add us as a preferred source to get our expert news, reviews, and opinion in your feeds. Make sure to click the Follow button!
And of course you can also follow TechRadar on TikTok for news, reviews, unboxings in video form, and get regular updates from us on WhatsApp too.
======================================================================
Link to news story:
https://www.techradar.com/pro/the-cpu-is-the-systems-executive-layer-intel-joi ns-sambanova-as-both-face-existential-threat-from-nvidias-groq-powered-inferen ce
--- Mystic BBS v1.12 A49 (Linux/64)
* Origin: tqwNet Technology News (1337:1/100)