unify-parameter-efficient-tuning
unify-parameter-efficient-tuning copied to clipboard
Implementation of paper "Towards a Unified View of Parameter-Efficient Transfer Learning" (ICLR 2022)
Thank you for the shared code. When I train to get a checkpoint, how to make the model load the parameters at this time?
Hi, thanks for publishing the paper and sharing the source code. I found that the "attn_output" is not used after definition. When learning roberta for parameter efficient learning, the paper...
Hi, thanks for sharing the source code. Could you please share the training log file, i.e., `log.txt`, with me? I just encountered some training problems, and the loss score decreased...
Thanks for your great work! I have read your paper, but I am a bit confused about two things. (1) The instantiation of Multi-head PA. How can we instantiate Multi-head...
 transformers.adapter only have prefix_tuning and adapter,where is the lora
Hi, You have mentioned that you used mosesdecoder for computing bleu score, can you explain that? maybe this is the source of difference between my results and yours. Thanks.
I am not familiar with the theoretic derivation, but I am interested in the range of suitability of the formula。Thank you。