transformers
8e08acad - Support `num_attention_heads` != `num_key_value_heads` in Flax Llama Implementation (#29557)

Commit
1 year ago
Support `num_attention_heads` != `num_key_value_heads` in Flax Llama Implementation (#29557) * fix tinyllama flax modelling * rename vars to minimize changes * move * formatting * remove unused var
Author
Parents
Loading