transformers
8e08acad
- Support `num_attention_heads` != `num_key_value_heads` in Flax Llama Implementation (#29557)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
Support `num_attention_heads` != `num_key_value_heads` in Flax Llama Implementation (#29557) * fix tinyllama flax modelling * rename vars to minimize changes * move * formatting * remove unused var
References
#29557 - Support `num_attention_heads` != `num_key_value_heads` in Flax Llama Implementation
Author
bminixhofer
Parents
f01e1609
Loading