AI Inference, Unleashed: Next-Gen Xe3P Graphics with 160GB LPDDR5X

Intel is introducing a new AI inference GPU for data centers, codenamed Crescent Island, built on the next-generation Xe3P graphics architecture. Designed for efficiency, scalability, and cost-effectiveness, this solution targets air-cooled enterprise servers and is tuned for high-throughput inference workloads. Customer sampling is slated for the second half of 2026.

Crescent Island advances Intel’s strategy by prioritizing performance-per-watt and large memory capacity—two critical factors for modern AI inference at scale. The Xe3P architecture is engineered to be more scalable than its predecessors, stretching from client iGPUs to full-fledged data center AI accelerators. On the client side, Xe3P will also underpin a next-gen Arc family, the Arc C-Series, showcasing the architecture’s breadth.

One of Crescent Island’s standout choices is memory. Instead of HBM, Intel is equipping the GPU with 160GB of LPDDR5X. While top-tier rivals are pushing HBM3E today and discussing HBM4 for future platforms, HBM supply has become tighter and more expensive due to surging demand. By opting for LPDDR5X, Intel aims to offer a compelling cost/performance balance, better availability, and attractive power characteristics—particularly beneficial for large-scale inference deployments.

Key highlights of Crescent Island:
– Xe3P microarchitecture optimized for performance-per-watt
– 160GB of LPDDR5X memory for high-capacity inference workloads
– Designed for air-cooled enterprise servers to simplify deployment and reduce costs
– Broad data type support, ideal for tokens-as-a-service providers and general inference use cases
– Backed by Intel’s open, unified software stack for heterogeneous AI systems, with early optimizations underway on Arc Pro B-Series GPUs

The software story is central to Crescent Island’s appeal. Intel is already developing and testing a unified AI stack across its ecosystem, enabling developers to prepare for the new architecture ahead of time. This should translate into smoother adoption and faster time to value when Crescent Island reaches customers.

What this means for data centers is straightforward: a power- and cost-optimized AI inference GPU with substantial memory capacity, designed to slot into standard air-cooled racks while supporting a wide range of model precisions and data types. For providers running large language model inference, recommendation engines, or tokens-as-a-service platforms, Crescent Island seeks to deliver scalable throughput without the premium and supply constraints typically associated with HBM-based accelerators.

FAQs

What is Crescent Island?
Crescent Island is Intel’s upcoming data center AI inference GPU based on the Xe3P architecture, optimized for power efficiency, cost, and memory capacity.

Why LPDDR5X instead of HBM?
HBM delivers exceptional bandwidth but faces supply constraints and higher costs. LPDDR5X helps Intel balance performance, availability, and total cost of ownership—key factors for large-scale inference.

Who is it for?
Enterprises and AI service providers running inference-heavy workloads, including tokens-as-a-service, large language model serving, and other high-throughput inference applications.

When will it be available?
Customer sampling is targeted for the second half of 2026.

What software support will it have?
Intel is building an open, unified software stack for heterogeneous AI systems and is already testing optimizations on Arc Pro B-Series GPUs to smooth the path to Crescent Island.