Qualcomm AI Hub Released at MWC 2024; Run AI Models on Your Device
At MWC 2024, Qualcomm has finally launched its AI Hub to deliver on-device AI experiences on several devices. The American chip maker says its AI engine is powerful enough to offer intelligent computing on all platforms, from smartphones to PCs.
Qualcomm AI Hub has a library of fully optimized AI models you can download and run locally on your device. It also supports many popular AI frameworks like Google’s TensorFlow and Meta AI’s PyTorch, as part of Qualcomm AI Stack.
Qualcomm AI Hub Has Optimized Over 75 AI Models
Qualcomm says its AI Hub hasover 75 optimized AI modelsand supports AI inferencing up to 4x than the competition. You can use these AI models for various tasks including image segmentation, image generation, text generation, low light enhancement, super-resolution, speech recognition, and more. In tandem, Qualcomm AI Hub can power multimodal AI models that can see, hear, and speak locally.
The chip giant mentioned that it canrun multimodal AI modelswith 7B+ parameters on Android devices locally. It can run models like LLaVa with text, speech, and image output support. The whole software stack has been optimized to deliver the best AI performance while consuming minimum power. Since everything is done locally, yourprivacy is preserved, and it also unlocks personalized AI experiences.
Even on Windows laptops running the upcomingSnapdragon X Elitechipset, you can input your voice, and it can reason with the speech and generate output in no time. Qualcomm has also employed a technique calledLoRA (Low-Rank Adaptation) to generate AI imagesusing Stable Diffusion locally on the device. It can run the Stable Diffusion model with 1B+ parameters.
Qualcomm vs Intel: NPU Performance
To further prove Snapdragon X Elite’s powerful on-device AI capability, Qualcomm did an image generation test on GIMP using the Stable Diffusion plugin. The Snapdragon X Elite-powered laptop generated an AI image injust 7.25 secondswhereas the latest x86-powered Intel Core UItra processor took 22.26 seconds.
It’s important to note that Intel launched theMeteor Lake architecture(powering Core Ultra CPUs) with various compute tiles, in which the NPU tile (built on Intel 4 node) is designed for on-device AI tasks. Despite that, the ARM-powered Snapdragon X Elite delivers3x better AI performance, largely because its NPU can do 45 TOPS alone. And when combined with CPU, GPU, and NPU, it results in compute performance of up to 75 TOPS.
Finally, Qualcomm is also working to bringHybrid AI experienceswhere AI processing tasks are distributed across local compute and cloud resources to offer better performance. You can check out Qualcomm AI Hub right here (visit).
So what do you think about Qualcomm AI Hub? Will it spur novel on-device generative AI experiences? Let us know in the comment section below.
Arjun Sha
Passionate about Windows, ChromeOS, Android, security and privacy issues. Have a penchant to solve everyday computing problems.
Add new comment
Name
Email ID
Δ
01
02
03
04
05