Janus icon indicating copy to clipboard operation
Janus copied to clipboard

Errors with Attention Mechanisms in Janus Inference Script (FlashAttention 2.0 & _flash_supports_window_size)

Open AlanPonnachan opened this issue 1 year ago • 3 comments

I encountered multiple issues when running the Janus-1.3B model inference script, both with and without enabling FlashAttention. These errors prevent successful execution of the model in a standard environment, such as Google Colab. Below are the details of the issues:

Problem 1: Error Without FlashAttention

When running the inference script without FlashAttention, the following error is raised during the generate function call: NameError: name '_flash_supports_window_size' is not defined

Problem 2: FlashAttention 2.0 Unsupported

To address the above error, I attempted to install FlashAttention-2 and enable it via the attn_implementation="flash_attention_2" argument. However, this raises the following error: ValueError: MultiModalityCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co/deepseek-ai/Janus-1.3B/discussions/new

Steps to Reproduce:

  1. Install FlashAttention 2.0 via: pip install flash-attn --no-build-isolation 2.Modify the script to include:
vl_gpt: MultiModalityCausalLM = AutoModelForCausalLM.from_pretrained(
    model_path, trust_remote_code=True, attn_implementation="flash_attention_2"
)

Thank you for your assistance! Let me know if you require additional details or logs to reproduce the issues.

AlanPonnachan avatar Dec 21 '24 04:12 AlanPonnachan

I am getting the following error:

MultiModalityCausalLM does not support Flash Attention 2.0 yet.

HussamCheema avatar Jan 28 '25 10:01 HussamCheema

我收到以下错误

MultiModalityCausalLM 尚不支持 Flash Attention 2.0。

hey! I have the same question. Have you done it?

Purshow avatar Feb 01 '25 09:02 Purshow

same error

OHaiYo-lzy avatar Feb 10 '25 06:02 OHaiYo-lzy