Daniel Han

Results 781 comments of Daniel Han

@mouhsineguet You can try lllamfying it via llama-factory also try searching on HF models for `qwen 14b llama` and there might be some llamafied versions

@JIBSIL What's the generation speed without Unsloth in Kaggle? Also why 150/? Shouldn't it be len(output)/?

@JIBSIL Whoops my bad - I think I fixed it now!

@HirCoir Apologies sorry was extremely busy this week so didn't have time to look at this! I'll see what I can do! @JIBSIL Also sorry did not respond until now!...

@JIBSIL Fixed batched inference yesterday (after your comment!!) See https://github.com/unslothai/unsloth/issues/267#issuecomment-2034047189 for more info. You'll need to update Unsloth without any dependency updates via `pip install --upgrade --force-reinstall --no-cache-dir git+https://github.com/unslothai/unsloth.git` for...

@JIBSIL Hmmm could be im not freeing something - let me check

@JIBSIL Much apologies on the delay - its possible there are small memory fragmentations over time, which will cause OOMs, but ye its possibly cause ur doing 7B and not...

Oh I think I did a typo whoops - let me first check why I did it lol

@pdurasie Sorry whoops long time on the PR - I did an overhaul, and I actually allow you to use `map_eos_token = False` now :)

@thedarkzeno Oh wait full finetuning - did you make all layers (Q, K, V, O, gate, up, down) + layernorms + lm_head, embeddings all trainable? I was gonna say I...