Go
Home
Pricing
FAQ
Install
Home
Pricing
FAQ
Install
Login
via GitHub
microsoft/DeepSpeed
Pull Requests
Commits
Open
Closed
Auto-detect CUTLASS for EvoformerAttention
#8000 opened 2026-05-08 18:26 by
MaxTretikov
zero3: SDMA allgather via mori (sdma_allgather)
#7999 opened 2026-05-07 13:10 by
inkcherry
Optimize singleton MoE collectives
#7997 opened 2026-05-07 02:28 by
Tianyi-Franklin-Wang
fix: use subprocess instead of os.system in data_analyzer.py
#7994 opened 2026-05-06 09:35 by
orbisai0security
docs: add test directory convention to AGENTS.md
#7993 opened 2026-05-06 06:26 by
delock
Add engine.coalesce_grad_reduction() for ZeRO 1/2/3 multi-backward
#7992 opened 2026-05-05 10:28 by
roycho96
fix gemma4 num attention head bugs (from #7975)
#7990 opened 2026-05-02 04:27 by
delock
Fix eigenvalue monitor logging
#7987 opened 2026-04-28 10:59 by
heurry
Add Qwen 3.5 preset to AutoTP
#7978 opened 2026-04-16 12:51 by
tohtana
fix gemma4 num attention head bugs
#7975 opened 2026-04-15 05:55 by
mingxiang1006
[Blog] Muon Optimizer Support in DeepSpeed
#7962 opened 2026-04-08 07:26 by
delock
Fix/warnings stacklevel mvapich runner
#7949 opened 2026-04-02 14:00 by
nathon-lee
Refactor/torch autocast encapsulate global state
#7946 opened 2026-04-02 06:06 by
nathon-lee
feat(zero2): add CPU offload support for Muon optimizer
#7939 opened 2026-03-31 06:59 by
delock
Add AutoEP
#7938 opened 2026-03-31 00:11 by
tohtana
Fix ZeRO-3 optimizer initialization validation (#7844)
#7929 opened 2026-03-28 16:20 by
amadhan882
[Feature] Enable AutoEP Compatibility with ZeRO-3
#7928 opened 2026-03-28 09:09 by
nathon-lee
Add torch_xla TPU support for ZeRO-1/2
#7917 opened 2026-03-21 18:43 by
PKUWZP
fix: add setup_context for torch.func compatibility
#7916 opened 2026-03-21 09:22 by
roycho96
doc: Remove suggestion to build extensions in parallel
#7899 opened 2026-03-12 15:58 by
Flamefire
[Bugfix] Validate fp16.loss_scale is finite in DeepSpeedFP16Config
#7892 opened 2026-03-08 20:00 by
s-zx
Fix Stage 0 + Ulysses crash: make bwc_tensor_model_parallel_rank() resilient to MP API absence
#7888 opened 2026-03-06 06:59 by
nathon-lee
fix(zero): Ensure full gradient reduction for Muon optimizer with reduce_scatter
#7878 opened 2026-02-27 06:46 by
nathon-lee
fix: correct DistributedAttention output shape and pad uneven sequence lengths (#7842)
#7868 opened 2026-02-22 11:00 by
harshang03
fix: keep fp32-pinned parameters out of the bf16 cast path in ZeRO-3 (#7747)
#7867 opened 2026-02-22 10:52 by
harshang03
Revert "fix: remove premature MPI environment variable check in OpenMPIRunner"
#7864 opened 2026-02-21 01:39 by
mikloorbi-sys
Fix global .cuh ignore and enforce tracked CUDA headers
#7858 opened 2026-02-18 04:38 by
harshang03
Fix ZeRO legacy grad-hook crash when next_functions is missing
#7857 opened 2026-02-17 22:07 by
harshang03
Reject non-finite fp16 loss_scale across config and ZeRO paths
#7856 opened 2026-02-17 18:13 by
harshang03
Fix zero/division safety gaps in utility and inference paths
#7855 opened 2026-02-17 18:05 by
harshang03
Older