We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support for MPS autocasting has recently be added in PyTorch 2.5.0 here and there is an ongoing effort to implement gradient scaling here.
PyTorch Lightning does not currently support mixed-precision on MPS device but it could be added in a near future when gradient scaling is finalized.
Is this feature considered? This would allow reducing memory usage and improving training time for some models.
Currently PyTorch Lightning falls back to FP32 when trying to use mixed-precision and issues a warning mentioning CUDA.
I think that considering adding a path for MPS mixed-precision would be great.
Stick to FP32 training when using MPS device.
thanks for your work!
cc @Borda
The text was updated successfully, but these errors were encountered:
No branches or pull requests
Description & Motivation
Support for MPS autocasting has recently be added in PyTorch 2.5.0 here and there is an ongoing effort to implement gradient scaling here.
PyTorch Lightning does not currently support mixed-precision on MPS device but it could be added in a near future when gradient scaling is finalized.
Is this feature considered? This would allow reducing memory usage and improving training time for some models.
Pitch
Currently PyTorch Lightning falls back to FP32 when trying to use mixed-precision and issues a warning mentioning CUDA.
I think that considering adding a path for MPS mixed-precision would be great.
Alternatives
Stick to FP32 training when using MPS device.
Additional context
thanks for your work!
cc @Borda
The text was updated successfully, but these errors were encountered: