-
-
Notifications
You must be signed in to change notification settings - Fork 5.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Model] Add Qwen2-Audio model support #9248
Conversation
👋 Hi! Thank you for contributing to the vLLM project. Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can do one of these:
🚀 |
Thanks for implementing this! Please run |
@DarkLight1337 Already fixed them. |
To ease testing, can you also update Also, it would be great if you could include this model in |
@DarkLight1337 Hi, I‘ve updated |
I'm unable to run the example script because of an error in By the way, you should also add this model to the Supported Models page of the docs. |
use OpenAI Audio API Client send rquest to Qwen2-Audio vllm backend, return error: |
excellent |
9bbcd67
to
5dff6ea
Compare
@faychu 推理时间长的音频 输入的input.tokens长度不对 |
Co-authored-by: DarkLight1337 <tlleungac@connect.ust.hk> Signed-off-by: Shanshan Wang <shanshan.wang@h2o.ai>
Co-authored-by: DarkLight1337 <tlleungac@connect.ust.hk> Signed-off-by: Shanshan Wang <shanshan.wang@h2o.ai>
Co-authored-by: DarkLight1337 <tlleungac@connect.ust.hk> Signed-off-by: qishuai <ferdinandzhong@gmail.com>
Co-authored-by: DarkLight1337 <tlleungac@connect.ust.hk> Signed-off-by: NickLucche <nlucches@redhat.com>
Co-authored-by: DarkLight1337 <tlleungac@connect.ust.hk> Signed-off-by: NickLucche <nlucches@redhat.com>
I just try more audio files, and this error only happens on one audio file, but I didn't find any special for this audio...... |
@faychu can you take a look at this? |
I build the latest vllm, use the #9248 (comment) and get the error: It seems some ops in moe cause the error, anyone can help me ? |
The latest vLLM uses torch 2.5.1, please update your vLLM dependencies by installing from |
@DarkLight1337 Thanks. I will try update torch to 2.5.1 |
Co-authored-by: DarkLight1337 <tlleungac@connect.ust.hk> Signed-off-by: Sumit Dubey <sumit.dubey2@ibm.com>
Co-authored-by: DarkLight1337 <tlleungac@connect.ust.hk>
Co-authored-by: DarkLight1337 <tlleungac@connect.ust.hk> Signed-off-by: Maxime Fournioux <55544262+mfournioux@users.noreply.github.com>
I downloaded the v0.6.4.post1 I’m encountering an error when trying to use vLLM serve with the Qwen/Qwen2-Audio-7B-Instruct model to process audio input. curl https://huxtwsgqgqkueq-5000.proxy.runpod.net/v1/chat/completions \
-X POST \
-H 'Content-Type: application/json' \
-d '{
"model": "Qwen/Qwen2-Audio-7B-Instruct",
"max_tokens": 1024,
"temperature": 0.1,
"messages": [
{
"role": "user",
"content": [
{
"type": "audio_url",
"audio_url": {
"url": "http://modelscope-open.oss-cn-hangzhou.aliyuncs.com/images/weather.wav"
}
},
{
"type": "text",
"text": "Transcribe Text"
}
]
}
]
}' Observe the error output:
|
Refer to #10493 |
Co-authored-by: DarkLight1337 <tlleungac@connect.ust.hk> Signed-off-by: Tyler Michael Smith <tyler@neuralmagic.com>
Thanks for adding this model. I'm running now via docker with 0.6.4.post1 and getting this error:
I'm passing these arguments:
I also ssh'd into the pod and can confirm that transformers 4.46.2 and vllm 0.6.4.post1+cu124 are installed (I'm running on runpod on an A40). One click template is here. |
Can you post the error log? |
Absolutely, should have done before:
|
You need to install |
ah, yes, thanks. Is there a way to get that in the openai docker image? or is the easiest/best thing for me to just build a docker image that wraps the current one and installs librosa? Thanks |
We don't include it inside our core dependencies because of licensing issues. It would be best for you to create your own docker image for this purpose. |
Noted, with thanks |
This PR adding support for Qwen2-Audio model.
FIX #8394
FIX #8461
Requirements
Use
transformers>=4.45.1
, and please install vLLM from source.Example Usage