AMD is pushing the boundaries of edge AI with its latest advancements, bringing robust AI capabilities to consumer PCs. The company has rolled out a driver update that supports large language models, revolutionizing the possibilities for PC users.
AMD’s innovative Strix Halo platform is at the forefront of this transformation, delivering remarkable AI power through its XDNA engines. The Strix Point and Strix Halo APU lineups are setting high standards in AI performance, and now, with the latest Adrenalin Edition 25.8.1 driver, these platforms support up to 128-billion parameter language models. This update permits consumer devices to run advanced AI models like Meta’s Llama 4 Scout, a significant milestone in consumer AI.
With AMD’s Variable Graphics Memory (VGM), users can allocate up to 96 GB of graphics memory to the integrated GPU, enabling the capability to run large-scale models directly on the device. Although the Llama model is MoE-based and operates with 17 billion parameters at runtime, it delivers impressive throughput, functioning as a capable AI assistant for users.
AMD’s strides extend to model context size, surpassing the industry norm of 4096 tokens. The new context length of 256,000 tokens offers users extensive control and ensures seamless performance without bottlenecks. These advancements highlight AMD’s role in expanding AI computing into consumer PCs.
Currently, the Strix Halo platform is available on select high-end devices, often priced over $2,000. While acquiring these cutting-edge machines might require a substantial investment, AMD’s progress signals a positive shift, making AI power more accessible to the general public.






