VeriSilicon announced on June 9 that the company's ultra-low energy consumption and high performance neural network processor IP now supports large language model inference on mobile devices. AI computing power can be expanded to more than 40 TOPS, which can meet the growing demand for generative AI on mobile platforms and provide computing power support for terminal devices such as AI PCs.

Zhitongcaijing · 06/09 10:57
VeriSilicon announced on June 9 that the company's ultra-low energy consumption and high performance neural network processor IP now supports large language model inference on mobile devices. AI computing power can be expanded to more than 40 TOPS, which can meet the growing demand for generative AI on mobile platforms and provide computing power support for terminal devices such as AI PCs.