accelerate
1ace241d - MLU devices : Checks if mlu is available via an cndev-based check which won't trigger the drivers and leave mlu (#3187)

Commit
1 year ago
MLU devices : Checks if mlu is available via an cndev-based check which won't trigger the drivers and leave mlu (#3187) * Add Cambricon MLU accelerator support * up mlu support for test * fix mlu device MULTI_MLU * Update src/accelerate/utils/imports.py it's beautiful ! Co-authored-by: Zach Mueller <muellerzr@gmail.com> * up mlu for quality check * fix mlu device longTensor error * fix mlu device tensor dtype check * fix mlu device send_to_device with torch dynamo error * Refactor AcceleratorState * Should be near complete now * Last missing piece * Make my way to the acceleratorstate * Include update to global var * Don't use global * gpu -> cuda * Don't use update for dict, easier to read * Fix tests * stash * Getting closer... * Needed to spawn at the very end after env was setup * Explain set_device before deepspeed * Make docstring more accurate * Early return insteaD * Delineat blocks * Make prepare_backend return state + backend for clarity/less magic * fix mlu longtensor.to() bugs. * fix MLU devices rng state save and load. * Cambricon MLU features, Checks if `mlu` is available via an `cndev-based` check which won't trigger the drivers and leave mlu uninitialized. * MLU devices : Checks if mlu is available via an cndev-based check which won't trigger the drivers and leave mlu * fix code style and quality * fix is_cuda_available error --------- Co-authored-by: Zach Mueller <muellerzr@gmail.com>
Author
Parents
Loading