-
Notifications
You must be signed in to change notification settings - Fork 1.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Can't PromptTuning in Multi-GPU with DeepSpeed and Qwen2.5-14B-Instruct #2266
Comments
Same error here Code
|
@dongshou Thanks for reporting the error.
It is strange that the labels are changed, the last one looks like an overflow error. PEFT does not touch the data at all, so I'm fairly certain that this is not a PEFT issue. Could you please try full fine-tuning with the same setup and check if the same error occurs? If you don't have enough memory for full fine-tuning, please test a smaller Qwen model and/or reduce the batch size, but leave the data the same. Please report back if you still get the same type of error with full fine-tuning. @hiyamgh Do you get the exact same error? Please paste the full error message. Also, are you using FSDP? |
System Info
Name: peft
Version: 0.12.0
Name: transformers
Version: 4.47.0
Name: accelerate
accelerate 0.34.2
Python 3.11.9
cuda
Build cuda_11.8.r11.8/compiler.31833905_0
Who can help?
No response
Information
Tasks
examples
folderReproduction
1 prompt tuning
2. dataset
3. model
4. trainer
5. deepspeed config json
6 debug info
when mv label to another coda, the label value have been changed!
loss code from transformers/loss/loss_utils.py
6.1log and error
Expected behavior
expect prompt-tuning with multi-gpu
The text was updated successfully, but these errors were encountered: