Sebastian Raschka
Sebastian Raschka
I think exporting to `.safetensors` would be nice in the future. In the meantime, to address your issue, you could load it via `state_dicts` -- I just had wanted to...
Thanks for the interest @srsawant34 ! Sorry, we had lot of ongoing PRs and issues in the last 2 weeks for a bit of a LitGPT redesign and must have...
@carmocca Sorry, I only used LoRA recently. One thing to try for experimentation purposes is to truncate LIMA, i.e, preparing it with `--max_seq_len 1028` so it's more comparable to Alpaca...
sry, I would have to rerun it again but plan to do so. Probably early next week! I'll make a reminder to give an update
I am getting similarly noise loss curves but the model performance with LIMA is actually quite good despite this. Maybe the examples in LIMA are so different that it's normal...
I agree. I think `generate.py` is more useful when you want to execute the end-to-end example in a bash script, but for the general tutorials let's use `chat.py`
Sorry for the late response, I've been traveling last week and I haven't fully caught up yet. The adapter methods are not very popular so we haven't prioritized them lately....
Arg you were right, the problem is the gating factor in the adapter models. That's currently not supported by our conversion tools for the Evaluation Harness. Sorry, this might be...
@TonAnh You are correct, this should be the main file. It's essentially converting the LitGPT checkpoint to a HF model to be used in the evaluation harness. In Lit-Llama, the...
Sorry for the hassle here. It's a bit frustrating that it doesn't work out of the box via torchvision anymore. This was also shared as an issue here: https://github.com/pytorch/vision/issues/1920 What...