tutuandyang
Results
1
comments of
tutuandyang
我在跑mini-internvl-4b预训练模型的时候也遇到了这样的问题:I get warning: You are not running the flash-attention implementation, expect numerical differences. A100服务器。torch version: 2.1.0a0+4136153,flash-attn version: 2.3.6,transformers version: 4.41.2