Term of the Moment

digital vacation


Look Up Another Term


Definition: AI chip


A GPU-based chip used to train and execute AI systems. The training side (deep learning) requires the most processing, and large language models cause quadrillions of calculations per second to be executed for days, weeks or months.

The execution side, called "inference," also requires high-performance chips. When people type a prompt into a chatbot, they expect results in a few seconds, and GPU chips are used for inference processing as well. NVIDIA is the world leader in AI chips (see Blackwell). See Tensor core, neural processing unit, deep learning and Cerebras AI computer.




The Xilinx Versal System-on-Chip
Today's chips often include AI processing. This Versal SoC contains more than 30 billion transistors and provides the parallel processing required for AI (green). It also contains programmable hardware, a rarity on any SoC (red) (see SoC and FPGA). See Versal. (Image courtesy of Xilinx.)