Intel and SambaNova team up on heterogenous AI inference platform — different hardware performs different workloads
Intel and SambaNova have collaborated on a new heterogeneous inference platform that utilizes different hardware components for various AI workloads. The platform leverages AI accelerators or GPUs for prefill, SambaNova's SN50 RDU for decoding, and Xeon 6 processors for agent-related operations and workload distribution. This architecture aims to compete with Nvidia by offering a scalable solution for enterprises and cloud operators, set to be available in the second half of 2026. The collaboration emphasizes the performance benefits of Xeon 6 processors and their compatibility with existing data center infrastructures.