Nice :) Can we add a test in tests/test_trainer.py
? We can set env variables to configure Accelerate properly (ACCELERATE_MIXED_PRECISION="fp8"
will auto-use TE)
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
The required tests are distributed tests. We need to verify FSDP functionality with and without FP8 mixed precision. The appropriate test file might be tests/trainer/test_trainer_fsdp.py
.
Is the test/trainer
folder included in the CI tests? Where can I check the results for test_trainer_fsdp.py
? @muellerzr @SunMarc
@eljandoubi we can't run them on the normal CI since GPU runners are not part of PR's.
Instead when ready I'll pull the PR down and run it myself
@muellerzr Thank you for the information. I have tested the branch in my code on a multi-node, multi-GPU setup using FSDP mode, both with and without FP8 mixed precision, and it worked as expected. Please let me know if you encounter any issues on your end.
@muellerzr Any updates regarding this PR?
Thanks! Can you do pip install -e .[quality]
followed by make fixup
? I'll then pull it locally to test on my 4090 system and we should be set!
@muellerzr I have done make fixup
.
Don't worry we'll merge as is, failing tests are unrelated!
Login to write a write a comment.
What does this PR do?
It passes the model and the optimizer to accelerate.prepare in order to enable fp8 mixed precision, if any.
Fixes #34024
Who can review?
Library:
-->