Go
Home
Pricing
FAQ
Install
Home
Pricing
FAQ
Install
Login
via GitHub
bigscience-workshop/Megatron-DeepSpeed
Pull Requests
Commits
sync4
LS/alibi
LS/doc
Lucile/add-eval-only-arg
Lucile/delete_unnecessary_brackets
Lucile/useless-parenthesis
add-valid-data
bitfit
bloom-ds-inference-repos2
bloom-inference-meta
bnb-resume-2x
bseval_harness
cc-concurrency
chpt-conversion-fix
ckptavg
cluster_benchmark
consumed_samples_per_valid_dataset
cyclic_valid_dataloaders
debug_with_new_dataset
dependabot/pip/black-24.3.0
ds_ckpt_reshape-with-layer-norm-auto-sync
ds-version-check
fix-sample-ids
fp32-checkpoint-extraction
gpu-direct
hadyelsahar/main
launch-debug
license
log-grad-norm
lumi_eval
lumi_mtf
main
master
megatron-2.4-ds-pipe
mtf_p3
mtf-multival
new-dataset
no-shuffling-option
nozero_reshape
olruwase/ds_ckpt_reshape
olruwase/sync_layer_norms
prefixbseval
preprocess_from_HF_dataset
rm-duplicate-param-count
samson/spm
scratchpad
self_attention_stable_corby
skip-broken-tests
sync4
t0loading
test-conversion
thomas/add_shared_t5
thomas/evaluate_gpt_on_prefix_lm_loss
thomas/evaluate_gpt_speed_if_we_pass_attention_mask
thomas/fix_installation
thomas/fix_layer_norm
thomas/improve_test_to_test_custom_kernel
thomas/mlm_train_script
thomas/opt
thomas/test_different_layer_norm
tp-ln-debug
tr1-13B
tr8-104B
train-no-eval-restart
training_flos_rebase
training_flos
universal_ckpt_info
universal_to_fp32_checkpoint
val_args
fix header
stas00
committed
4 years ago
1c1423ed
add testing and contribute info
stas00
committed
4 years ago
0d404320
fix arg help (#65)
stas00
committed
4 years ago
Verified
56c2983b
Add test suite (#64)
stas00
committed
4 years ago
Verified
55b7036c
use HuggingFace Datasets as source to build Megatron data files (#48)
adammoody
committed
4 years ago
Verified
5069622a
[microsoft/Megatron-DeepSpeed sync] Commits including 2021-08-09 (#58)
stas00
committed
4 years ago
Verified
3c9d748b
Add GLU variants (#47)
jaketae
committed
4 years ago
Verified
effb2fb8
better comment on TB writer (`is_last_rank`)
VictorSanh
committed
4 years ago
5e3963db
Fix pretrain_gpt_single_node example script to have only one occurence of lr-decay-style
thomasw21
committed
4 years ago
7228b89e
chore: add deepspeed as comment
jaketae
committed
4 years ago
Verified
f4566b85
Update README.md (#51)
lintangsutawika
committed
4 years ago
Verified
a6cbb3bb
Add LRU cache, add faster tokenization (#37)
huu4ontocord
committed
4 years ago
Verified
36284576
Update requirements.txt (#46)
jaketae
committed
4 years ago
Verified
f919d0cf
rm `(s)` that slipped through
VictorSanh
committed
4 years ago
49b9d7de
Group tensorboard metrics (#39)
VictorSanh
committed
4 years ago
Verified
9e75429d
chore: update gitignore (#45)
jaketae
committed
4 years ago
Verified
752e958c
fix attn_mask (#50)
stas00
committed
4 years ago
Verified
42fe3b39
add .bs to the version to help check we are on the right repo/branch
stas00
committed
4 years ago
bdef48f7
fix the deepspeed example
stas00
committed
4 years ago
1cef056c
add a section on how we use deepspeed with Meg
stas00
committed
4 years ago
faf9809f
Faster preprocessing (#18)
thomasw21
committed
4 years ago
Verified
7b998814
Create README.md
stas00
committed
4 years ago
Verified
6394089f
Revert "Revert "use pp engine even for pp=1 (#6) (#34)""
stas00
committed
4 years ago
781676b5
Revert "use pp engine even for pp=1 (#6) (#34)"
stas00
committed
4 years ago
ad2e3d3a
use pp engine even for pp=1 (#6) (#34)
stas00
committed
4 years ago
Verified
6c6c64ac
Revert "Update gpt2_tokenization.py"
thomasw21
committed
4 years ago
8938d7ac
Update gpt2_tokenization.py
huu4ontocord
committed
4 years ago
Verified
a40d8163
fix timing (#31)
stas00
committed
4 years ago
Verified
feca57ac
Implement rotary embeddings (#7)
thomasw21
committed
4 years ago
Verified
dc4e0cba
adding missing prefix on merge
stas00
committed
4 years ago
5563ec6f
Newer
Older