logo
Qualcomm® Cloud AI
Export the Model
Initializing search
    • User Guide
    • API
    • FAQ
    • Blogs
      • Introduction
      • Quick Start Guide
      • Installation
        • Checklist
        • Pre-requisites
        • Cloud AI SDK
        • Hypervisors
        • Containers
          • Docker
          • Kubernetes
        • Triton Inference Server
        • vLLM
        • AWS
        • Text Generation Inference
      • Inference Workflow
        • Export the Model
          • Exporting ONNX Model from Different Frameworks
          • Operator and Datatype support
          • Introduction to the Model Preparator Tool
        • Export the Model
          • Exporting ONNX Model from Different Frameworks
          • Operator and Datatype support
          • Introduction to the Model Preparator Tool
        • Compile the Model
          • Compile the Model
          • Tune Performance
        • Execute the QPC
          • Execute the QPC
          • Inference Profiling
      • Pytorch Workflow
      • Model Architecture Support
        • Large Language Models (LLMs)
      • Features
        • Custom Operations (C++)
        • Model Sharding
      • System Management
        • System Management
        • AIC-manager
      • Architecture
      • Glossary
      • Python API
        • qaic package
        • class Util
        • class InferenceBenchmarkSet
      • CPP API
        • InferenceSet IO Example
        • Features
        • Runtime
      • ONNX Runtime
        • QAIC execution provider
      • Frequently Asked Questions
      • Train anywhere, Infer on Qualcomm® Cloud AI 100
      • Quadruple LLM Decoding Performance with Speculative Decoding (SpD) and Microscaling (MX) Formats
      • Accelerate Inference of Fully Transparent Open-Source LLMs from LLM360 on Qualcomm® Cloud AI 100 DL2q Instances
      • Extremely Low-Cost Text Embeddings on Qualcomm® Cloud AI 100 DL2q Instances
      • Accelerate Large Language Model Inference by ~2x Using Microscaling (Mx) Formats

    Export the Model¶

    Refer to the following sections for details on exporting models:

    • Exporting ONNX Model from Different Frameworks
    • Operator and Datatype support
    • Introduction to the Model Preparator Tool
    Back to top
    Previous Inference Workflow
    Next Exporting ONNX Model from Different Frameworks
    © Copyright 2025 Qualcomm Innovation Center, Inc.
    Created using Sphinx 7.1.2. and Sphinx-Immaterial