ERNIE
ERNIE copied to clipboard
Fix Core Attn in ernie 4.5
Fix Error in These two situation (This PR only affects the self_attention in modeling when use_flash_attention is set to False):
- When using the ./tools/inference/infer.py script for inference, setting use_flash_attention results in the following error