Frequently Asked Questions¶
General¶
What is Cloud AI 100 Accelerator?
Cloud AI 100 accelerators enable high performance inference on deep learning models. The accelerators are available in multiple form factors and associated SKUs. Cloud AI SDKs enable end to end workflows - from onboarding a pre-trained model to deployment of the ML inference application in production.
Cloud AI SDK Installation and Platform/OS Support¶
What operating systems and platforms are supported?
Where do I download the SDKs?
Cloud AI SDK consists of a Platform and Apps SDK. Refer to Cloud AI SDKs for more information.
For Platform SDK download, see Platform SDK Download
For Apps SDK download, see Apps SDK Download
What environment variables need to be set to resolve toolchain errors such as libQAic.so?
Set the environment variables as mentioned here.
Deep Learning frameworks and networks¶
Which deep learning frameworks for supported by Cloud AI SDKs?
ONNX, TensorFlow, and PyTorch are supported by the
compiler.
qaic-exec
can dump the operators supported across different
frameworks.
Which deep learning neural networks are supported?
Cloud AI platforms supports many model categories - Computer vision, object detection, semantic segmentation, natural language processing, and Generative AI networks. Performance information can be found at Qualcomm Cloud AI 100. Model recipes can be found in the Cloud AI SDK GitHub.
I have a neural network that I would like to run on Cloud AI platforms. How do I go about it?
There are 3 steps to run an inference on Cloud AI platforms.
Export the model in ONNX format (preferred due to operator support) and prepare the model
Compile the model to generate a QPC binary (Qaic Program Container)
Execute, integrate and deploy into production pipeline
The Quick Start Guide provides a quick overview of the steps involved in running inference using a vision transformer model as an example.
Refer to Inference Workflow for detailed information how to onboard and run inference on Cloud AI platforms.
Users can also refer to the Model Recipes for the best performance for several models across several categories.
Tutorials <https://github.com/quic/cloud-ai-sdk> are another resource that walks through the workflows to onboard models, tune for best performance and profile inferences.
Runtime errors during inference¶
While running inference I encounter ‘IOCTL: Connection timed out ERROR’. What is the fix?
There are 3 timeout settings that we recommend users to increase when this issue is encountererd. If the issue is not fixed, please raise a case through the Qualcomm case support system or Cloud AI SDK GitHub.
sudo sh -c "echo 4000000 > /sys/module/qaic/parameters/wait_exec_default_timeout"
sudo sh -c "echo 4000000 > /sys/module/qaic/parameters/control_resp_timeout"
sudo sh -c "echo 4000000 > /sys/module/qaic/parameters/mhi_timeout"
System management¶
Which utility/tool is used to query health, telemetry and other information for all Cloud AI cards in the server?
Use the qaic-util command-line tool to query all Cloud AI cards in the server.
The Cloud AI device shows Status:Error
. How do i fix it?
Status: Error
could be due to one of the following:
Card boot-up not completed
User is not in the qaic group or has not used sudo prefix. sudo /opt/qti-aic/tools/qaic-util
BIOS secure boot needs to be disabled
Unsupported OS/platforms
Try soc_reset to recover the device.