Go
Home
Pricing
FAQ
Install
Home
Pricing
FAQ
Install
Login
via GitHub
huggingface/accelerate
Pull Requests
Commits
v0.34.0-release
3d-parallelism
argparse
better-err
big_api
check-docs
check-for-nccl
composable-tp
context-parallel
context-parallel-experiments
context-parallel-flex-attn
cp-dataloader
cp-pc
dataloader-log
debug-tests
deepspeed-inference
deepspeed-version
device_map_xla_support
disable-seedale-rs
enable-dash
feat/async-checkpointing
feat-decorator-to-purge-modified-accelerate-env-vars
fix
fix-compile-regions
fix-deepspeed-autobs
fix-dispatch-model-tied-params-memory
fix-fp8
fix-generate
fix-grad-norm
fix-pjrt_device
fix-prod
fix-warnings
fork-tester
fp8-gradient-checkpointing
fp8-stuff
fsdp2-tp
fully-remove-accelerate-config
grad-acc-optimizer-fixes
grad-accum-test
import-util
llama-to-mistral
load-model-across-devices
low-bit-fsdp2
main
make-version-tests-better
mishig25-patch-1
mishig25-patch-2
mixed-precision-experiments
ms-amp
muellerzr-ds-debugging
muellerzr-fix-1.0
muellerzr-fp8-deepspeed-support-v2
muellerzr-msamp-ds-fsdp
muellerzr-nightly-fixings
muellerzr-stateful-dl
new-instance-type
nouamane/context-parallel
parallelism-config
pin-ruff
pip-uv
pippy-duplicates
pippy-integration
reaction-based-runs
release-v0.6.1
release-v0.6.2
revert-3671
revert-fsdp-improv
revert-pr
rm-112
runner
safetensors-default
slack-reporter
speedup-docker
test-data
test-deepspeed-unpin
torch-22
trainer-tests
transformers-nd-parallel
ulysses-sp
unfreeze-4090
use-partialstate
uv-take2
v0.7-release
v0.12-release
v0.13-release
v0.14-release
v0.15-release
v0.16-release
v0.17-release
v0.18-release
v0.19-release
v0.20-release
v0.21-release
v0.22-release
v0.23-release
v0.24-release
v0.25.0-release
v0.26.0-release
v0.26.1-release
v0.27.0-release
v0.28.0-release
v0.29.0-release
v0.30.0-release
v0.31.0-release
v0.32.0-release
v0.33.0-release
v0.34.0-release
v1.0.0-release
v1.1.0-release
v1.2.0-release
v1.3.0-release
v1.4.0-release
v1.5.0-release
v1.6.0-release
v1.7.0-release
v1.8.0-release
v1.9.0-release
v1.10.0-release
v1.11.0-release
v1.12.0-release
wip-from-pretrained
xla-gpu-runners
Release: v0.34.2
muellerzr
committed
1 year ago
c61f41c6
Release: v0.34.1
muellerzr
committed
1 year ago
beb43781
Allow DataLoaderAdapter subclasses to be pickled by implementing `__reduce__` (#3074)
byi8220
committed
1 year ago
e13bef2c
Fix FSDP auto_wrap using characters instead of full str for layers (#3075)
muellerzr
committed
1 year ago
73a1531e
Release: v0.34.0
muellerzr
committed
1 year ago
159c0dd0
Remove `skip_first_batches` support for StatefulDataloader and fix all the tests (#3068)
muellerzr
committed
1 year ago
Verified
8931e5e4
Speed up tests by shaving off subprocess when not needed (#3042)
muellerzr
committed
1 year ago
Verified
a8485924
add set_epoch for MpDeviceLoaderWrapper (#3053)
append-only
committed
1 year ago
Verified
758d6243
Fix typo in comment (#3045)
mokizzz
committed
1 year ago
Verified
b07ad2ad
use duck-typing to ensure underlying optimizer supports schedulefree hooks (#3055)
tmm1
committed
1 year ago
Verified
1d09a20f
Do not import `transformer_engine` on import (#3056)
oraluben
committed
1 year ago
Verified
3fcc9461
Update torchpippy (#2938)
muellerzr
committed
1 year ago
Verified
939ce400
Add FP8 docker images (#3048)
muellerzr
committed
1 year ago
Verified
c2120927
Add a SLURM example with minimal config (#2950)
muellerzr
committed
1 year ago
Verified
654e1d99
Update CONTRIBUTING.md Setup Instructions (#3046)
siddk
committed
1 year ago
Verified
8c3aded2
Decouple `prepare_data_loader()` from Accelerator (#3047)
siddk
committed
1 year ago
Verified
27899339
Fixup dataloader state dict bugs + incorporate load/save_state API (#3034)
muellerzr
committed
1 year ago
Verified
726140ca
Fix batch_sampler maybe None error (#3025)
candlewill
committed
1 year ago
Verified
2d4f1dda
Fix fp8 benchmark on single GPU (#3032)
muellerzr
committed
1 year ago
Verified
c0cf860d
Add early support for `torchdata.stateful_dataloader.StatefulDataLoader` within the `Accelerator` (#2895)
byi8220
committed
1 year ago
Verified
ad3f574a
Improve config handling and add a zoo (#3029)
muellerzr
committed
1 year ago
Verified
1a6af0bd
Add end_training/destroy_pg to everything and unpin numpy (#3030)
muellerzr
committed
1 year ago
Verified
52fae096
Fix torch version check (#3024)
muellerzr
committed
1 year ago
Verified
7ffe7662
Set correct NPU backend and distributed_type when using transfer_to_npu (#3021)
ArthurinRUC
committed
1 year ago
Verified
5536a3a8
Tweak defaults for quantized-typed FP8 TE weights (#3018)
muellerzr
committed
1 year ago
Verified
7ec8eab9
destroy process group in `end_training` (#3012)
SunMarc
committed
1 year ago
Verified
589fddd3
Wrong import check for TE (#3016)
muellerzr
committed
1 year ago
Verified
99c69aaf
fix default value for rank size in cpu threads_per_process assignment logic (#3009)
rbrugaro
committed
1 year ago
Verified
00785cd9
Enable FSDP & Deepspeed + FP8 (#2983)
muellerzr
committed
1 year ago
Verified
a452327e
Fix `find_tied_params` for models with shared layers (#2986)
qubvel
committed
1 year ago
Verified
851cf343
Older