flax
637b9f6c
- Add LAMB optimizer
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
6 years ago
Add LAMB optimizer This is similar to LARS but with adam instead of momentum as the wrapped update rule (and a couple other differences). It's popular for large-batch transformer training.
References
#108 - Add LAMB optimizer
Author
jekbradbury
Parents
15bcf6cc
Loading