Haoqin Tu
Haoqin Tu
> > Can you do fine-tune using 7B? > > @yPanStupidog I try to change the llm. But I do not know if my pretrain loss is normal. Because I...
Hi @A9isha, does Maxtext support the other way round now? That's converting HF's Llama or Mistral weights to MaxText checkpoints. Thanks
> We have the script [llama_or_mistral_ckpt.py](https://github.com/google/maxtext/blob/main/MaxText/llama_or_mistral_ckpt.py) to convert the original PyTorch Llama2 checkpoint that Meta provides into MaxText checkpoint. > > You can see the usage [here](https://github.com/google/maxtext/blob/main/end_to_end/tpu/llama2/7b/test_llama2_7b.sh#L35) for Llama2-7b for...
> I see, unfortunately no there isn't the conversion script at the moment. It should be a modification of [llama_or_mistral_ckpt](https://github.com/google/maxtext/blob/main/MaxText/llama_or_mistral_ckpt.py). If you are interested, please feel free to send across...
So, is there a schedule now to release the code of metalm? It would be really nice to try it!
你好,我也在尝试使用自己的数据集进行训练,但是我的category(上衣、裤、裙)类型的分类只有一个,代码报错(估计是category类匹配不上),没能找出错误来,请问你解决了在自己数据集上训练的问题吗?能否分享经验?多谢
Same feature request here. It would be really helpful if you guys can integrate wandb or tensorboard to the training process! Thanks.
Same issue here for BEiTv3 finetuning, wondering what caused this problem..
> @ImKeTT, could you provide your training command? yes, here's my command for finetuning the model on VQAv2 task: ```bash CUDA_VISIBLE_DEVICES=0,1 python -m torch.distributed.launch --nproc_per_node=2 --master_port=25641 run_beit3_finetuning.py \ --model beit3_base_patch16_480...
Is there a solution for running LLaVA on TPU right now?