pyreft
pyreft copied to clipboard
[P1] TGI and vLLM support
-
Are there plans for inference support. This is needed if it's to be used by devs in production.
-
Is fine tuning much faster than LoRA?
- Optimization and backward pass are MUCH faster, but surely forward pass is similar (technically, slightly slower)
- Why so many epochs?
- I was surprised to see 10-12 epochs in the paper.
- in practice with LoRA I find less is more (often just do one epoch with constant LR) because it stops overfitting