[Performance]: Qwen3-Omni text-only output inference speed is slower than vllm
April 24, 2026 ยท #3093
Python
Difficulty: Medium
Parent Repository
vllm-project/vllm-omni
Python repository
4,473 832
vllm-project/vllm-omni
Python repository
Sign in required
Authenticate to use favourites & bookmarks