Megatron-LM
Megatron-LM copied to clipboard
add core_attention_bias_type to TransformerConfig
core_attention_bias_type
is needed to use alibi from transformer engine https://docs.nvidia.com/deeplearning/transformer-engine/user-guide/api/pytorch.html?highlight=alibi#transformer_engine.pytorch.DotProductAttention.forward