Disaggregated approaches to AI inference are creating a new category of GPUs optimized for the compute-intensive process of loading context data into LLMs. New products at NVIDIA, Intel, and Huawei shed light both on the vendors' target markets and the nature of cutting-edge AI workloads

A subscription is required to view this content.

Already subscribed? Continue Continue