We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Docker version: vault.habana.ai/gaudi-docker/1.19.0/ubuntu24.04/habanalabs/pytorch-installer-2.5.1:1.19.0-561 Optimum-Habana version: 1.14.1
examples
python ../gaudi_spawn.py --use_deepspeed --world_size 8 run_generation.py \ --model_name_or_path meta-llama/Llama-2-70b-hf \ --use_hpu_graphs \ --limit_hpu_graphs \ --use_kv_cache \ --bf16 \ --trim_logits \ --attn_softmax_bf16 \ --bucket_size=128 \ --bucket_internal \ --batch_size 8 \ --max_input_tokens 40960 \ --max_new_tokens 5120 \ --use_flash_attention \ --flash_attention_recompute \ --flash_attention_causal_mask \ --book_source
This doesn't run, but instead throws the following error after loading the models onto the GPU:
/home/jenkins/workspace/cdsoftwarebuilder/create-binaries-from-sw-sources---bp-dt/repos/hcl/src/ibverbs/hcl_ibverbs.cpp::295(create_qp): The condition [ rc == 0 ] failed. hbldv_modify_qp(INIT) failed: 22, nic: 11
I have uploaded the full file below:
habana_error.txt
The example should run to completion.
The text was updated successfully, but these errors were encountered:
No branches or pull requests
System Info
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
This doesn't run, but instead throws the following error after loading the models onto the GPU:
I have uploaded the full file below:
habana_error.txt
Expected behavior
The example should run to completion.
The text was updated successfully, but these errors were encountered: