transformers
[gpt-neox] Add attention_bias config to support model trained without attention biases
#28126
Merged

Loading