nomadlx

Results 6 comments of nomadlx

The error traceback: ``` File "/home/exp_or_nmt/code/OR-NMT/OR-Transformer/fairseq/trainer.py", line 380, in train_step raise e File "/home/exp_or_nmt/code/OR-NMT/OR-Transformer/fairseq/trainer.py", line 358, in train_step ignore_grad=is_dummy_batch, File "/home/exp_or_nmt/code/OR-NMT/OR-Transformer/fairseq/tasks/fairseq_task.py", line 337, in train_step loss, sample_size, logging_output = criterion(model,...

The BUG seems to have been caused here: https://github.com/ictnlp/OR-NMT/blob/239e05e48c2ed4748de01e8909f919e836d821db/OR-Transformer/fairseq/search.py#L78 right is: ``` beams_buf = indices_buf // vocab_size ``` This bug was fixed, but another bug popped up... ``` File "/home/exp_or_nmt/code/OR-NMT/OR-Transformer/fairseq/models/oracle_transformer.py",...

> https://github.com/uclaml/SPIN/blob/e84b7be111b41b388367e591bdc23e327725c869/spin/alignment/trainer.py#L405 > > 在 spin_loss 定义中,在步骤 0 处,当 p_theta 等于 p_theta_t 时,损失值从固定值 0.6931 开始 I know that the initial loss is not equal to 0 in the actual code,...

> https://llava-vl.github.io/blog/2024-05-10-llava-next-stronger-llms/ How should I set `model_name` and `conv_template` for llava-next-72b/110b? Is there anything else I should be aware of?

> > For llava-next-72b/110b at `lmms-eval`, `conv_template=qwen_1_5`. > > At slgang, it's slightly different, you can check the `examples/usage` folder at sglang's repo. > > I've trying to find this...

> It's using same code structure but with different content, you could regard llava-next-72b/110b relies on upgraded llava repo. > > Since it's team efforts, we release it to a...