gritlm icon indicating copy to clipboard operation
gritlm copied to clipboard

Generative Representational Instruction Tuning

Results 38 gritlm issues
Sort by recently updated
recently updated
newest added

I understand that the GritLM fine-tuning uses both in-batch negative and hard negatives for contrastive learning. We can use in-batch negatives only by setting train group size to 1. However,...

Formatted and added typing for the main file gritlm.py. No significant changes

The original run.py saves the model in pytorch_model.bin, which cannot be loaded directly using the code provided in this repository. After replacing line 422 `trainer.save_model()` in training/run.py with `model.model.save_pretrained(training_args.output_dir)`, the...

I was trying to finetuning Meta-Llama-3-8B-Instruct using 4 gpus with the following command: `torchrun --nproc_per_node 4 -m training.run --output_dir llama3test --model_name_or_path meta-llama/Meta-Llama-3-8B-Instruct --train_data training/toy_data --learning_rate 1e-5 --num_train_epochs 5 --per_device_train_batch_size 1...

Work in progress! Not tested yet. `is_causal` support added. Uploading for comments.

I'm now trying to train llama3.1 with GRIT pipeline. At first I directly change ``--model_name_or_path`` and run the training code (the training script I used is as follows) ``` #!/bin/bash...

I am trying to evaluate GritLM-7B on MTEB datasets using the provided script. ``` #!/bin/bash python /home/e/e1347696/unified_encoder_decoder/src/eval/MTEB/eval_mteb.py \ --model_name_or_path /home/e/e1347696/unified_encoder_decoder/model/GritLM-7B \ --output_folder /home/e/e1347696/unified_encoder_decoder/src/results/GritLM-7B-mteb \ --task_types Classification,Clustering,PairClassification,Reranking,Retrieval,STS,Summarization \ --batch_size 32 ```...

Hi, I’m working with the GRITLM repository. And I'm training Mistral 7B and evaluating performance on the MTEB benchmark with NVIDIA RTX A6000. I first tested the pretrained mistralai/Mistral-7B-v0.1 model...