ZebangCheng

Results 75 comments of ZebangCheng

如果您希望在本地部署demo,你需要执行app.py代码,请注意gradio的版本应该是3.47.1,如果是最新的版本,会导致报错,无法加载页面。 app_EmotionLlamaClient.py是我们用来提供api调用功能的代码。由于开发历史的原因,app_EmotionLlamaClient.py和app.py要求的环境是有冲突的。您如果能执行app_EmotionLlamaClient.py没问题,那么执行app.py就会出现问题。您需要重新创建一个新的环境,设置gradio的版本为3.47.1。

I noticed that the official Google Drive download link appears to be invalid. Fortunately, the corresponding model parameters (minigptv2_checkpoint.pth) can still be downloaded from the Hugging Face Space: > https://huggingface.co/spaces/Vision-CAIR/MiniGPT-v2/tree/main

是的,两个阶段都是使用lora训练,只是训练的数据不同。

Thank you for your interest in our work! I have a clarification to make: are you concerned about how to extract features from different encoders, or are you looking to...

Sure, I understand your question. In our paper, we discussed prompts related to extracting visual information using MiniGPT-v2 and audio information with Qwen-Audio. You can try setting up the corresponding...

We are currently organizing our code and plan to make everything open source. The script for building the MERR dataset involves multiple projects, and we need to consolidate everything before...

**Vision:** Use the open-source OpenFace tool to extract facial expression AU (Action Unit) information in advance, then map the AU data into corresponding textual descriptions through character mapping. **Audio:** Use...

Thank you for your kind words and for using Emotion-LLaMA! To answer your question: Yes, **we strongly recommend extracting `.npy` features in advance** for evaluation, as this setup is designed...

Here is your refined and translated response: --- Thank you for your follow-up question! In fact, the `eval_emotion.py` script **does include HuBERT audio features**—they are part of the `video_features` variable....

Apologies, progress has been slower than anticipated. Recently, our research efforts have focused on integrating with related projects within our research group. We experimented with several approaches, but the results...