Welcome to Efficient-Transformers Documentation!
- Train anywhere, Infer on Qualcomm Cloud AI 100
- How to Quadruple LLM Decoding Performance with Speculative Decoding (SpD) and Microscaling (MX) Formats on Qualcomm® Cloud AI 100
- Power-efficient acceleration for large language models – Qualcomm Cloud AI SDK
- Qualcomm Cloud AI 100 Accelerates Large Language Model Inference by ~2x Using Microscaling (Mx) Formats
- Qualcomm Cloud AI Introduces Efficient Transformers: One API, Infinite Possibilities