Prajwal Ganugula

Results 2 issues of Prajwal Ganugula

currently after merging pruning masks and LoRA weights, LLaMA-7B size is increasing from 15GB to 26GB. Please provide support to remove pruned weights from the model

Hi, I was wondering if this method can be used for trimming large vocabulary in LLMs. Can vocab trimmer be extended to LLMs?