You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am currently using the Megatron framework and want to use Lora for training. I saw that the Megatron format is supported at https://github.com/huggingface/peft/blob/main/src/peft/tuners/lora/tp_layer.py RowParallelLinear and ColumnParallelLinear do the adaptation. But if I use the transformer engine, the corresponding TELayerNormColumnParallelLinear and TERowParallelLinear will not be adapted.
Motivation
This will better support Megatron framework using LoRA.
Your contribution
I don't have a PR.
The text was updated successfully, but these errors were encountered:
Unfortunately, I don't have any experience with this so I can't really give any tips. Do you have a reproducer for this issue? Also a gentle ping @zhangsheng377 just in case.
Ha, actually I have never used TE directly. But we reserved backend when writing distributed LoRA. You can see if you can specify it as your TE. If it is inconvenient, you can consider using an adaptation layer class to modify the model.
Feature request
I am currently using the Megatron framework and want to use Lora for training. I saw that the Megatron format is supported at https://github.com/huggingface/peft/blob/main/src/peft/tuners/lora/tp_layer.py RowParallelLinear and ColumnParallelLinear do the adaptation. But if I use the transformer engine, the corresponding TELayerNormColumnParallelLinear and TERowParallelLinear will not be adapted.
Motivation
This will better support Megatron framework using LoRA.
Your contribution
I don't have a PR.
The text was updated successfully, but these errors were encountered: