-
Notifications
You must be signed in to change notification settings - Fork 193
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
LLM inference on Lunar Lake (LNL) iGPU is not working #1358
Comments
Could you please share the error messages that you encountered when using GPU plugin? On another note, could you please run the following command as share the output with us? |
I do not get any error or exception. The code exits at the line The output requested for device supported is |
Thanks for providing the information with us. Could you please provide the additional information with me?
|
Output of command systeminfo | findstr /B /C:"OS Name" /B /C:"OS Version"OS Name: Microsoft Windows 11 Pro CPUIntel(R) Core(TM) Ultra 9 288V AVAILABLE_DEVICES : [''] AVAILABLE_DEVICES : [''] iGPUIntel(R) Arc(TM) 140V GPU (16GB) (iGPU) AVAILABLE_DEVICES : ['0'] AVAILABLE_DEVICES : ['0'] NPUIntel(R) AI Boost AVAILABLE_DEVICES : ['4000'] AVAILABLE_DEVICES : ['4000'] |
Thanks for the information. I'll escalate this to relevant team, and we'll update you as soon as possible. |
Ref. 159930 |
@azhuvath please send the error messages for Int4 and int8 models |
I do not get any error or exception. The code exits at the line |
Tried the sample openvino-genai code for iGPU and it not working. It is not working for NPU also. The below code exits without any exception. It works fine with CPU.
I was following the blog https://medium.com/openvino-toolkit/how-to-run-llama-3-2-locally-with-openvino-60a0f3674549 . It uses a dGPU and not iGPU. Not sure if it is not supported for iGPU.
device = 'GPU'
pipe = openvino_genai.LLMPipeline(args.model_dir, device)
I tried the INT4 model on iGPU. It was not working and hence created the INT8 model using the below command.
optimum-cli export openvino --model meta-llama/Llama-3.2-3B-Instruct --task text-generation-with-past --weight-format int8 --group-size 64 --ratio 1.0 --sym --all-layers llama-3.2-3b-instruct-INT8
Package Details
openvino==2025.0.0.dev20241209
openvino-genai==2025.0.0.0.dev20241209
openvino-telemetry==2024.5.0
openvino-tokenizers==2025.0.0.0.dev20241209
optimum==1.23.3
optimum-intel @ git+https://github.com/huggingface/optimum-intel.git@02835ce8833b5e2b67ba1a87bf85b0739335ac4d
The text was updated successfully, but these errors were encountered: