Deployment¶
This section covers how to deploy Qualcomm Cloud AI workloads across different environments — from local containers to cloud instances, Kubernetes clusters, and virtualized infrastructure.
Docker — run inference workloads using pre-built Qualcomm Cloud AI container images. Covers available images, image selection by use case (LLM, CV, disaggregated serving), and instructions for building and launching custom images.
Kubernetes — deploy containerized Cloud AI workloads on Kubernetes using the QAic device plugin. Covers device plugin configuration, SKU-based and fractional device allocation, and deployment YAML examples.
KServe — deploy vLLM as a Kubernetes-native
InferenceServiceusing KServe. CoversServingRuntimeconfiguration, autoscaling, and inference examples on Minikube and AWS EKS.
AWS — get started on AWS DL2q instances powered by Cloud AI 100 Standard accelerators. Covers AMI setup, instance configuration, and running your first LLM with Qualcomm Efficient-Transformers.
Hypervisors — assign Cloud AI devices to virtual machines via PCIe passthrough. Covers KVM, Hyper-V, ESXi, and Xen hypervisor configurations.