A GPU-based chip used to train and execute AI systems. The training side (deep learning) requires the most processing, and large language models cause quadrillions of calculations per second to be executed for days, weeks or months.
The execution side, called "inference," also requires high-performance chips. When people type a prompt into a chatbot, they expect results in a few seconds, and GPU chips are used for inference processing as well. NVIDIA is the world leader in AI chips (see
Blackwell). See
Tensor core,
neural processing unit,
deep learning and
Cerebras AI computer.
The Xilinx Versal System-on-Chip
Today's chips often include AI processing. This Versal SoC contains more than 30 billion transistors and provides the parallel processing required for AI (green). It also contains programmable hardware, a rarity on any SoC (red) (see
SoC and
FPGA). See
Versal.
(Image courtesy of Xilinx.)