Nick

Results 11 comments of Nick

mode use the function with the argument : `model = torch.compile(model, mode="reduce overhead")` _**Or:**_ `model = torch.compile(model, mode="max-autotune")` _**OR:**_ `model = torch.compile(model, mode="reduce-overhead", fullgraph=True, backend='Eager')` - fastest

> @NeuralAIM did you manage to actually run it with YOLOv5 model? Yes 💯

| Model | Test Size | APval | AP50val | AP75val | Param. | FLOPs | | :-- | :-: | :-: | :-: | :-: | :-: | :-:...

**When will this be applied in the main branch?**

> > **When will this be applied in the main branch?** > > We have to rework this feature in a new pr since this pr is incompatible with the...

Same with `sentence-transformers/LaBSE` [b2636](https://github.com/ggerganov/llama.cpp/releases/tag/b2636) - working

**Output:** ![image](https://github.com/user-attachments/assets/af52513f-81fe-433a-830c-960bbed766d9) **API Info:** ![image](https://github.com/user-attachments/assets/e162f961-7792-43c7-8af4-8cc163e6b4b1) https://github.com/paul-gauthier/aider/issues/705#issuecomment-2195777748, Why is there no auto-continue for other models?

> > Why is there no auto-continue for other models? > > Because Anthropic provides prefill of assistant responses, so the output can be resumed. https://docs.anthropic.com/en/docs/build-with-claude/prompt-engineering/prefill-claudes-response **Added:** https://platform.deepseek.com/api-docs/news/news0725 ![image](https://github.com/user-attachments/assets/8fe101c0-f400-422d-883b-f0246446b8da) @paul-gauthier

**DeepSeek API introduces Context Caching on Disk** ![image](https://github.com/user-attachments/assets/a171dd56-0b75-4d22-a028-8f7d7c7539f6) ![image](https://github.com/user-attachments/assets/201a31a2-2a13-42cf-bc4c-b8b467cff9c7) https://platform.deepseek.com/api-docs/news/news0802/