LESS
LESS copied to clipboard
[ICML 2024] LESS: Selecting Influential Data for Targeted Instruction Tuning
Hi, When I run "Step 2: Building the gradient datastore" FileNotFoundError: [Errno 2] No such file or directory: '../out/llama2-7b-p0.05-lora-seed3/checkpoint-1688/optimizer.bin' I check the folder "llama2-7b-p0.05-lora-seed3" generate from Step 1, only files...
when load the optimizer.pt display the key is different KeyError: 'base_model.model.model.layers.0.self_attn.q_proj.lora_A.default.weight' the items in optimizer.pt state is 0~255.
Following the script provided in the second step of "Selecting data for a task" in your readme, I have a command line that needs to be run as shown below:...
How can we utilize multiple GPUs for the gradient feature collection step? The current implementation only works with a single GPU.
thanks for sharing your code, https://github.com/princeton-nlp/LESS/blob/main/less/data_selection/write_selected_data.py#L76 In this code version, a small mistake made sorted.csv goes wrong, to make it correct, line 76 and line 77 should exchange their position
CUDA error: the provided PTX was compiled with an unsupported toolchain. What might be the cause for this mistake?
``` N_SUBTASKS = {"mmlu": 57, "bbh": 27, "tydiqa": 9} influence_score = influence_score.reshape( influence_score.shape[0], N_SUBTASKS[target_task_name], `-1).mean(-1).max(-1)[0] ``` what is meaning N_SUBTASKS , why do this? Can I change it to "...
Hello, I have some questions about the accuracy of llama2-7b. In the Table 5, the accuracy of llama2-7b-base on MMLU/TYDIQA/BBH are 46.7/52.1/39.8, but we use llama2-7b from "https://huggingface.co/meta-llama/Llama-2-7b-hf/tree/main" to test...
Hi, I'm trying to run experiments following the instructions given in README. I find that in Step 1 warmup training, 5% of samples are randomly selected to train $M_S$. But...
 Hi, I've been attempting to reproduce an experiment involving the finetuning of the Llama-2-7b-hf model, specifically using a random 5% of training data, using open-instruct [finetune_with_accelerate.sh](https://github.com/allenai/open-instruct/blob/main/scripts/finetune_with_accelerate.sh). I adhered to...