xmc-andy

Results 18 comments of xmc-andy

> May I know your task type and which version of Otter model you are using for initialization? I am doing a classification task, with multiple images and a single...

export PYTHONPATH=. accelerate launch --config_file=./pipeline/accelerate_configs/accelerate_config_fsdp.yaml \ pipeline/train/instruction_following.py \ --pretrained_model_name_or_path /mnt/large_model/weights/OTTER-Image-MPT7B_git \ --mimicit_vt_path /mnt/large_model/output/XX/SD_instruction.json \ --images_vt_path /mnt/large_model/output/XX/SD.json \ --external_save_dir /mnt/large_model/output/XX/OTTER-Identify-Image-MPT7B-BC4-partScale-negAug3 \ --batch_size 1 \ --num_epochs 15 \ --run_name OTTER-Identify-Image-MPT7B-BC4-partScale-negAug3 \ --workers...

When loading the pre-trained weights you posted and the baseline weights I trained, there will be no log loss of weights, but there will be when loading the newly trained...

By the way, due to network problems, the network cannot download tokenizer_config.json from huggingface's MPT, so I downloaded it offline through "https://huggingface.co/mosaicml/mpt-7b-instruct", except for the bin file and in modeling_otter.py...

I compared the config.json. Except for "_name_or_path" and "transformers_version", the rest are consistent with what you posted. This should not be the problem. I used to convert the trained weights...

I checked the save_pretrained part as you said, I'm using the version about a month ago, the save code is as follows" unwrapped_model = accelerator.unwrap_model(model) checkpoint_dict = get_checkpoint(model=unwrapped_model) accelerator.save( checkpoint_dict,...

> I would suggest you to use the `save_pretrained` method (it's a function of Huggingface Transformers). This method directly dump all things of your current trained model to a `path`...

Hey, I used to convert the trained weights final_weights.pt through otter/converting_otter_pt_to_hf.py, and then load the weights from pretrained. Please tell me if this is correct? I found that when converting...

> "transformers_version" Got it, the generated config.json only has "_name_or_path" and ""transformers_version"" different from what you posted.

> It could be right. If you confirm that the `config.json` is the same. Thank you very much for your careful answer. I have solved this bug. The cause of...