Qualcomm has announced some details of its tilt at the AI datacenter market by revealing a pair of accelerators and rack scale systems to house them, all focused on inferencing workloads.

The company offered scant technical details about its new AI200 and AI250 “chip-based accelerator cards”, saying only that the AI 200 supports 768 GB of LPDDR memory per card, and the AI250 will offer “innovative memory architecture based on near-memory computing” and represent “a generational leap in efficiency and performance for AI inference workloads by delivering greater than 10x higher effective memory bandwidth and much lower power consumption.”

Qualcomm will ship the cards in pre-configured racks that will use “direct liquid cooling for thermal efficiency, PCIe for scale up, Ethernet for scale o

See Full Page