And that would be completely inaccurate. Neural Networks and libraries are employed in multiple real-world applications on phones currently. Your on-device speech to text processing doesn't happen unless the machine is listening to your spoken word and inferring what you're saying and translating it to text. This is done with anything from tensor cores, which do exist in modern smartphone SoCs like Snapdragon chips, to DSP complexes as well that are on-chip. Also, machine vision to improve image capture and processing is used as well, detecting actual objects in a scene and determining what color/contrast, sharpening and other balance options need to be applied. Even silly Tik-Tok and Instagram image/video manipulation uses AI on phones these days. So don't call BS unless you really understand what you're speaking of. The benchmarks are relative to the capabilities of a given handset to perform tasks just like these. Obviously workstation and server AI is on another level, and often where AI training is done, rather than just inference.