Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[The current device architecture is not supported by sycl_ext_oneapi_device_architecture] #12809

Open
ammrabet opened this issue Feb 11, 2025 · 2 comments
Assignees

Comments

@ammrabet
Copy link

Hello,

I am using intelanalytics/ipex-llm-serving-xpu docker image to run VLLM on the Battlemage card B580

i got this error :

terminate called after throwing an instance of 'sycl::_V1::exception'
what(): The current device architecture is not supported by sycl_ext_oneapi_device_architecture.
Aborted (core dumped)

command line :

python3 ./benchmark_vllm_throughput.py --backend vllm --dataset ./ShareGPT_V3_unfiltered_cleaned_split.json --model deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B --num-prompts 100 --seed 42 --trust-remote-code --enforce-eager --dtype float16 --device xpu

Thank you in advance
Amine mrabet

@gc-fu
Copy link
Contributor

gc-fu commented Feb 12, 2025

Hi, we will investigate the issue.

@gc-fu
Copy link
Contributor

gc-fu commented Feb 13, 2025

Hi, we have not encountered this problem after trying with our newest image: intelanalytics/ipex-llm-serving-xpu:2.2.0-b13.

Try pull this image and see if this problem persists~

The command:

python3 ./benchmark_vllm_throughput.py --backend vllm --dataset ./ShareGPT_V3_unfiltered_cleaned_split.json --model /llm/models/DeepSeek-R1-Distill-Qwen-1.5B --num-prompts 100 --seed 42 --trust-remote-code --enforce-eager --dtype float16 --device xpu --gpu-memory-utilization 0.95 --max-model-len 3000 --max-num-batched-tokens 3000 

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants