compilade

Results 109 comments of compilade

> Re using bigger batch sizes - does this mean if memory allows, imatrix should be in fact faster to process via PP? @danielhanchen Currently, with `llama-imatrix` from the `master`...

To address some feedback I got recently, I've added a warning when writing using the legacy format so that it's more obvious what is happening. ``` save_imatrix: saving to legacy...

> @compilade Time to merge this (and adapt #12718 afterwards)? @CISC Sure. I hope I've tested enough edge cases. Will merge at 16:00 UTC on 2025-07-19 (in around 10 hours),...

I'm currently working on a big refactor of how Mamba (and Jamba) works to make all sequences of a sub-batch be of the same length (initially only for models with...

I've pushed the refactor to use equal-sequence-length sub-batch splitting for recurrent models. This greatly simplifies the SSM operations, no need for `inp_s_seq` anymore. And recurrent state slot allocation is now...

> @compilade Do you have local changes in this branch? Would like to merge latest `master` here @ggerganov I do have local changes, which I've pushed now. I was in...

> The change is quite big and I'm having a bit of trouble to merge it all at once. Wonder if we should take a more step-by-step approach. I agree...

Now that variable GQA support is in `master` (because of #7359 which has been merged), I plan to separate the advanced batch splits feature in its own PR for easier...

> Any updates on this since Jamba 1.5 is now out? @Autumnlight02 Basically, since was merged, now I need to resolve a *very big* merge conflict because I didn't keep...

Some progress update on Jamba: I began resolving the merge conflicts, and there were at least 2000+ lines of conflicts (basically half of this PR). This is manageable. While I've...