Go
Home
Pricing
FAQ
Install
Home
Pricing
FAQ
Install
Login
via GitHub
bigscience-workshop/Megatron-DeepSpeed
Pull Requests
Commits
olruwase/sync_layer_norms
LS/alibi
LS/doc
Lucile/add-eval-only-arg
Lucile/delete_unnecessary_brackets
Lucile/useless-parenthesis
add-valid-data
bitfit
bloom-ds-inference-repos2
bloom-inference-meta
bnb-resume-2x
bseval_harness
cc-concurrency
chpt-conversion-fix
ckptavg
cluster_benchmark
consumed_samples_per_valid_dataset
cyclic_valid_dataloaders
debug_with_new_dataset
dependabot/pip/black-24.3.0
ds_ckpt_reshape-with-layer-norm-auto-sync
ds-version-check
fix-sample-ids
fp32-checkpoint-extraction
gpu-direct
hadyelsahar/main
launch-debug
license
log-grad-norm
lumi_eval
lumi_mtf
main
master
megatron-2.4-ds-pipe
mtf_p3
mtf-multival
new-dataset
no-shuffling-option
nozero_reshape
olruwase/ds_ckpt_reshape
olruwase/sync_layer_norms
prefixbseval
preprocess_from_HF_dataset
rm-duplicate-param-count
samson/spm
scratchpad
self_attention_stable_corby
skip-broken-tests
sync4
t0loading
test-conversion
thomas/add_shared_t5
thomas/evaluate_gpt_on_prefix_lm_loss
thomas/evaluate_gpt_speed_if_we_pass_attention_mask
thomas/fix_installation
thomas/fix_layer_norm
thomas/improve_test_to_test_custom_kernel
thomas/mlm_train_script
thomas/opt
thomas/test_different_layer_norm
tp-ln-debug
tr1-13B
tr8-104B
train-no-eval-restart
training_flos_rebase
training_flos
universal_ckpt_info
universal_to_fp32_checkpoint
val_args
Merge remote-tracking branch 'origin/main' into olruwase/sync_layer_norms
stas00
committed
3 years ago
5b368846
Eval harness (#212)
DanielHesslow
committed
3 years ago
Verified
3ab0ad18
Fixed MLM dataset arguments(#290)
thomasw21
committed
3 years ago
Verified
55f8cf8b
Mlm adaptation (#287)
Lintang Sutawika
committed
3 years ago
Verified
9d264312
Fix DS init (#285)
Quentin-Anthony
committed
3 years ago
Verified
987663c1
Fix tflops glu computation (#283)
Muennighoff
committed
3 years ago
Verified
e23393fb
[valid] deadlock workaround (#282)
stas00
committed
3 years ago
Verified
cb48bd2c
Fix mixed fused layer norm to mimick nn.LayerNorm for torch>1.11 (#281)
thomasw21
committed
3 years ago
Verified
908dc9cb
Update CODEOWNERS
TevenLeScao
committed
3 years ago
Verified
c85b7c25
Update CODEOWNERS
TevenLeScao
committed
3 years ago
Verified
b6266f55
Update CODEOWNERS
TevenLeScao
committed
3 years ago
Verified
c6f22c41
Create CODEOWNERS
TevenLeScao
committed
3 years ago
Verified
89c343f7
tweak the doc
stas00
committed
3 years ago
4c13c617
add start-fast doc (#278)
stas00
committed
3 years ago
Verified
40bd933d
Fix device issue when using torch.broadcast
thomasw21
committed
3 years ago
475f3730
Sync torch_rng_state (#277)
thomasw21
committed
3 years ago
Verified
d576775c
add stop alarm instructions
stas00
committed
3 years ago
Verified
640c8180
typo
stas00
committed
3 years ago
2ac141b1
improve the doc, and comment out the demo
stas00
committed
3 years ago
86b726cb
fp32 accessors
tjruwase
committed
3 years ago
84825956
add 2 more weights to sync
stas00
committed
3 years ago
bf7eeb3a
compare on cpu
stas00
committed
3 years ago
d64a947e
add the test script
stas00
committed
3 years ago
d2aa4f18
fix regex
stas00
committed
3 years ago
4443e6d2
dynamically discovered layer norm weights / refactor
stas00
committed
3 years ago
fc8f813d
fix requirements
stas00
committed
3 years ago
8f2ea60b
Sync lp/hp/optim for layer norms
tjruwase
committed
3 years ago
c7f20066
fix
stas00
committed
3 years ago
a5b5edc0
Merge remote-tracking branch 'origin/main' into thomas/test_different_layer_norm
stas00
committed
3 years ago
3c5e4914
`torch.testing.assert_equal` didn't make it (#273)
stas00
committed
3 years ago
Verified
87a9dba0
Older