Artificial Intelligence

AI is changing everything. Combined with powerful, energy efficient processors and ubiquitous connectivity to the wireless edge, intelligence is moving to more devices, changing industries, and inventing new experiences.

On-device AI allows for real-time responsiveness, improved privacy, and enhanced reliability along with better overall performance and with or without a network connection. Our Qualcomm Artificial Intelligence (AI) Engine along with our AI Software and Hardware tools (including our Qualcomm® Neural Processing SDK for AI) as outlined below, are designed to accelerate your on-device AI-enabled applications and experiences.

The Qualcomm Artificial Intelligence (AI) Engine is available on supported Snapdragon® 888, 865, 855, 845, 835, 821, 820 and 660 mobile platforms, and with cutting-edge on-device AI processing found in the Snapdragon 888.


Snapdragon and Qualcomm Neural Processing are products of Qualcomm Technologies, Inc. and/or its subsidiaries.

Developers strive to balance power efficiency and AI predictive performance at the edge by optimizing their machine learning models. In this blog, we take a closer look at ML model optimization...
Image processing tasks generally perform better when working with high-quality image data. But what if you don’t have high-quality images? In this blog we explore how technology from NexOptic helps...
What is a smart sensor? Learn how you can use smart sensors in your development.
Overview on distributed Artificial Intelligence webinar

Download the newest version of our Neural Processing SDK, or access any of our archived versions of the SDK.
This SDK is engineered to supply computer vision algorithms for Snapdragon platforms.
This SDK for AI helps optimize performance of trained neural networks on devices with Snapdragon.
This SDK library includes vision processing functions for use across multiple mobile devices.

Single-board computer (SBC) built on the Snapdragon 888 mobile platform.
Designed for mobile device development on the Snapdragon 865 mobile platform.
High-performance development device designed to support on-device AI application development
A highly integrated and optimized Android development kit designed on the Snapdragon 855 mobile platform

Qualcomm AI Research works to advance AI and make its core capabilities – perception, reasoning, and action – ubiquitous across devices. The goal is to make breakthroughs in fundamental AI research and scale them across industries. One way we contribute innovative and impactful AI research to the rest of the community is through novel papers at academic conferences.

Beyond papers, the Qualcomm Innovation Center (QuIC) actively contributes code based on this breakthrough research to open source projects.

The AI Model Efficiency Toolkit (AIMET) is a library that provides advanced quantization and compression techniques for trained neural network models. QuIC open sourced AIMET on GitHub to collaborate with other leading AI researchers, provide a simple library plugin for AI developers, and help migrate the ecosystem toward integer inference. Read the blog post or watch some informational AIMET videos to learn more.

The AIMET Model Zoo, another GitHub project, provides the recipe for quantizing popular 32-bit floating point (FP32) models to 8-bit integer (INT8) models with little loss in accuracy. Read the blog post to learn more. Check out the Qualcomm Innovation Center YouTube channel for informational videos on our open source projects to help developers get started.

Data is another crucial element for machine learning. If you need the Qualcomm Abstract Syntax Tree (QAST) dataset that was used to support the experiments in our workshop paper at ICLR 2019: Simulating Execution Time of Tensor Programs Using Graph Neural Networks, check out our QAST Project Page. We hope this new dataset will benefit the graph research community and raise interest in Optimizing Compiler research.