Wallaroo.AI (Version 2025.2)
  • 2025.2 (Current Version) 202502-PU
    • 2025.2 (Current Version)
    • 2025.1
    • 2024.4
    • 2024.3
    • 2023.2
  • Home
  • Inference Tutorials
  1. Inference
  2. Inference Tutorials
  3. Portability
  4. Hardware
  5. Qualcomm

Qualcomm

The following tutorials demonstrate deploying different LLMs with Qualcomm QIAC AI acceleration.


Deploy Llama with Continuous Batching Using Native vLLM Framework and QAIC AI Acceleration

Deploy RAG Llama with QAIC

Deploy Llama with Continuous Batching Using Native vLLM Framework with QAIC using OpenAI Inference

Deploy RAG Llama with OpenAI compatibility on QAIC

© 2026 Wallaroo Labs, Inc.