Skip to content

Introduction

Developers can access Qualcomm Cloud AI hardware through cloud instances or by purchasing servers equipped with Qualcomm Cloud AI hardware.

Cloud Instances

Cloud AI 100 cards are now available at 2 Cloud service providers - Amazon Web Services (AWS) and Cirrascale Cloud Services. The AI 100 accelerator SKUs and instance configurations offered at these providers can vary.

Note

Developers using cloud instances can skip the rest of the installation section. Click here to go to the next section, Inference Workflow

On-Premise Servers

Developers with on-premise servers need to work with system administators to ensure Cloud AI SDKs are installed and verified properly. It is recommended for developers and system admins to go through the installation section in its entirety.

Installation

The Platform SDK (x86-64 and ARM64) and Apps SDK (x86-64 only) are targeted for Linux-based platforms. The SDKs can be installed natively on Linux operating systems. Container and orchestration are also supported through Docker and Kubernetes. Virtual machines, including KVM, ESXi, and Hyper-V, are also supported. This section covers:

  • Installation of the SDKs across multiple Linux distributions
  • Building a docker image with the SDKs and third-party packages for a seamless execution of QAic inference tools/workflow
  • Setting up KVM, ESXi, and Hyper-V, and installation of SDKs

Compilation and Execution modes

Apps and Platform SDKs enable just-in-time(JIT) or ahead-of-time(AOT) compilation and execution on x86-64 platforms while only AOT compilation/execution is supported on ARM64.

In JIT mode, compilation and execution are tightly coupled and require Apps and Platform SDKs to be installed on the same system/VM.

In AOT mode, compilation and execution are decoupled. Networks can be compiled ahead-of-time on x86-64 (with Apps SDK only) and the compiled networks can be deployed on x86-64 or ARM64 with Platform SDK.

Both JIT and AOT are supported on x86-64 when Apps and Platform SDK are installed on the same server/VM.

Hardware Requirements

  • CPU - Server grade x86 or ARM multi-core CPU
  • RAM - minimum 512 GB (recommend 768GB or higher)
    • Required for the compilation of large models and corresponding configurations like batch size, context length etc.
  • Storage - minimum 1TB (recommend 4TB)
    • Large storage is recommended to store multiple models (and associated artifacts), serialized engines etc.

Supported Operating Systems, Hypervisors, and Platforms

The Cloud AI Platform SDK is compatible with the following operating systems (OS) and platforms.

Operating Systems

Operating systems Kernel
Ubuntu 20.04 Default Kernel (GA or HWE)
Ubuntu 22.04 Default Kernel (GA or HWE)
Ubuntu 24.04 Default Kernel (GA or HWE)
Red Hat Enterprise Linux 8.9 Default Kernel
Red Hat Enterprise Linux 9.3 Default Kernel
AWS Linux 2 Amazon 5.10 Kernel
AWS Linux 2023 Default Kernel
Note1: Arm is a trademark of Arm Limited (or its subsidiaries) in the US and/or elsewhere.
Note2: Apps SDK is available only for x86-64 platforms.
Note3: Ultra cards consume 4 DRM/Accel device resources per card. Kernels prior to 6.2 are limited to 64 DRM device resources for the entire system. Using Ubuntu 22.04 with the HWE kernel is recommended as the resource limit is 256 Accel devices. Deployments with large numbers of Ultra cards that do not follow this recommendation may not be able to use all the combined hardware resources of the Ultra cards.
Note4: AWS Linux 2 support available only for x86-64 platforms.

Hypervisors

Cloud AI only supports PCIe passthrough to a virtual machine. This means that the virtual machine completely owns the Cloud AI device. A single Cloud AI device cannot be shared between virtual machines or between a virtual machine and the native host.

Hypervisor x86-64 ARM64
KVM
Hyper-V
ESXi
Xen
Note Cloud AI SDKs are not required to be installed on the hypervisor. All the Cloud AI software/SDKs are installed on the guest VMs.