pytorch
d6f3265e - [FSDP] Do not clean FQNs even for `use_orig_params=True` (#91767)

Commit
2 years ago
[FSDP] Do not clean FQNs even for `use_orig_params=True` (#91767) Cleaning FQN for `FullyShardedDataParallel(use_orig_params=True)` can cause some discrepancies with respect to the FQN compared to manually looping over `named_modules()` and `named_parameters()` together. There is no requirement for the FQNs to be clean when using wrapper FSDP + `use_orig_params=True`. We can leave clean FQNs to `fully_shard`. Pull Request resolved: https://github.com/pytorch/pytorch/pull/91767 Approved by: https://github.com/zhaojuanmao
Author
Andrew Gu
Committer
Parents
Loading