Write a 0.2 changelog
There's been quite a few changes from 0.1. We should document them for people updating their applications.
Thoughts on getting this into the next release? https://github.com/rustformers/llm/pull/224
Fine by me!
I guess we'll want to look at this in order to figure out what goes into the changelog https://github.com/rustformers/llm/compare/v0.1.1...main
Looking at the history for the inference.rs example is another good way to track changes to the public API https://github.com/rustformers/llm/commits/main/crates/llm/examples/inference.rs
we can do so automatically here https://docs.github.com/en/repositories/releasing-projects-on-github/automatically-generated-release-notes
Example here
What's Changed
- refactor: move load to KnownModel trait by @philpax in https://github.com/rustformers/llm/pull/199
- Refactors from llm-chain integration by @danforbes in https://github.com/rustformers/llm/pull/200
- Fine-grain inference feedback by @danforbes in https://github.com/rustformers/llm/pull/206
- Update upstream GGML dependency by @danforbes in https://github.com/rustformers/llm/pull/215
- Implement RedPajama as a NeoX Specialization by @danforbes in https://github.com/rustformers/llm/pull/204
- Implement LoRA patching via the Loader by @LLukas22 in https://github.com/rustformers/llm/pull/211
- Dynamically calculate the LoRA patching context size by @LLukas22 in https://github.com/rustformers/llm/pull/225
- NeoX -> GPT-NeoX by @danforbes in https://github.com/rustformers/llm/pull/233
- Update GGML dependency by @danforbes in https://github.com/rustformers/llm/pull/226
- fix: cancel the loading spinner in error cases by @tehmatt in https://github.com/rustformers/llm/pull/236
- Added support for multiple LoRA adapters by @LLukas22 in https://github.com/rustformers/llm/pull/232
- fix: qnt0 build by @philpax in https://github.com/rustformers/llm/pull/242
- Typo fix in extern function name by @tanmaysachan in https://github.com/rustformers/llm/pull/247
- Implement MPT Model by @danforbes in https://github.com/rustformers/llm/pull/218
- feat: support quantizing with a GGML container by @philpax in https://github.com/rustformers/llm/pull/245
- Add perplexity calculation and cli-arg by @tanmaysachan in https://github.com/rustformers/llm/pull/248
- Update Gpt-NeoX implementation by @LLukas22 in https://github.com/rustformers/llm/pull/237
- Fix perplexity calculation by @tanmaysachan in https://github.com/rustformers/llm/pull/251
- GGML README by @danforbes in https://github.com/rustformers/llm/pull/224
- Lingering NeoX reference by @danforbes in https://github.com/rustformers/llm/pull/254
- Model architectures should appear in alphabetic order by @danforbes in https://github.com/rustformers/llm/pull/255
- More model standardization by @danforbes in https://github.com/rustformers/llm/pull/256
- README updates for v0.2 by @danforbes in https://github.com/rustformers/llm/pull/253
- More CONTRIBUTING updates by @danforbes in https://github.com/rustformers/llm/pull/258
- Update headings by @danforbes in https://github.com/rustformers/llm/pull/259
- Context size consistency & fixes by @danforbes in https://github.com/rustformers/llm/pull/260
- Remove LLaMA "old_loader" by @danforbes in https://github.com/rustformers/llm/pull/262
- refactor: remove model-associated inference params by @philpax in https://github.com/rustformers/llm/pull/272
- Pull upstream changes into GptNeox by @LLukas22 in https://github.com/rustformers/llm/pull/270
- Fix embeddings extraction by @skirodev in https://github.com/rustformers/llm/pull/273
- Update GGML Dependency & Bindings by @danforbes in https://github.com/rustformers/llm/pull/275
- Add scratchbuffers to MPT by @LLukas22 in https://github.com/rustformers/llm/pull/278
- Fixed broken GPT-J quantization by @LLukas22 in https://github.com/rustformers/llm/pull/283
- Add HuggingFace's Tokenizer by @RedBoxing in https://github.com/rustformers/llm/pull/271
- Sampler rework by @philpax in https://github.com/rustformers/llm/pull/280
- Add alternative key shortcuts for shift enter by @pdufour in https://github.com/rustformers/llm/pull/287
- Fix error in Usage section of README by @danforbes in https://github.com/rustformers/llm/pull/290
- Made Sampler and InferenceParameters threadsafe by @LLukas22 in https://github.com/rustformers/llm/pull/292
- Fix embeddings extraction for all models by @skirodev in https://github.com/rustformers/llm/pull/291
- Fix
--helppanicing in debug mode by @tehmatt in https://github.com/rustformers/llm/pull/300 - Default to f16 model memory k/v in llm CLI and InferenceSessionConfig by @KerfuffleV2 in https://github.com/rustformers/llm/pull/296
- Switch from GGML to LLama.cpp as GGML source. by @LLukas22 in https://github.com/rustformers/llm/pull/302
- NixOS flake stuff needs to be in root by @danforbes in https://github.com/rustformers/llm/pull/306
- Add bindings section to README by @danforbes in https://github.com/rustformers/llm/pull/309
- adding secondbrain to applications section on Readme by @juliooa in https://github.com/rustformers/llm/pull/310
- Added support for cuBLAS and CLBlast in ggml. by @darxkies in https://github.com/rustformers/llm/pull/282
- Fix GPT-2 Positional Embeddings by @LLukas22 in https://github.com/rustformers/llm/pull/322
- Implement Metal support by @pixelspark in https://github.com/rustformers/llm/pull/311
- fix(metal): patch ggml-metal.m with shader code by @philpax in https://github.com/rustformers/llm/pull/328
- K-quants by @nightscape in https://github.com/rustformers/llm/pull/326
- feat: automatically run checks before commit by @philpax in https://github.com/rustformers/llm/pull/329
- Added InferenceSessionRef to exports, for easier serializing and deserializing. by @JuliaMerz in https://github.com/rustformers/llm/pull/335
- Add Falcon Support by @LLukas22 in https://github.com/rustformers/llm/pull/313
- Added tokenizers-remote by @jafioti in https://github.com/rustformers/llm/pull/339
- Basic Integration Testing by @LLukas22 in https://github.com/rustformers/llm/pull/337
- feat(test): check hyperparameters can roundtrip by @philpax in https://github.com/rustformers/llm/pull/344
- Improve model architecture error message by @viirya in https://github.com/rustformers/llm/pull/348
- fix #216 - split prompts for chat mode in CLI by @philpax in https://github.com/rustformers/llm/pull/346
- Add placeholder check for message prompt by @viirya in https://github.com/rustformers/llm/pull/351
- Fix linux CLBlast build and add CI for accelerated builds by @LLukas22 in https://github.com/rustformers/llm/pull/358
- Add CI for cuda builds by @LLukas22 in https://github.com/rustformers/llm/pull/360
- Add ability to delete tokens (undo feed) by @steventrouble in https://github.com/rustformers/llm/pull/334
- Update gpt2 to use wte if no lm_head by @steventrouble in https://github.com/rustformers/llm/pull/362
- Add CI for OpenCL on windows by @LLukas22 in https://github.com/rustformers/llm/pull/366
- allow chat to halt new token generation on
stop_sequenceby @averypelle in https://github.com/rustformers/llm/pull/364 - feat(tracing): add tracing to
llmandllm-basecrates by @radu-matei in https://github.com/rustformers/llm/pull/367 - CUDA/OpenCL Acceleration by @LLukas22 in https://github.com/rustformers/llm/pull/325
- Fix MSRV to 1.65.0 by @philpax in https://github.com/rustformers/llm/pull/375
- Peg CUDA architecture by @LLukas22 in https://github.com/rustformers/llm/pull/376
- Update README.md by @basraayman in https://github.com/rustformers/llm/pull/379
- Update README.md by @baskrahmer in https://github.com/rustformers/llm/pull/380
- Better GPU Documentation by @LLukas22 in https://github.com/rustformers/llm/pull/385
- add Falcon 40B model support by @skirodev in https://github.com/rustformers/llm/pull/368
- Custom RoPE Scaling by @LLukas22 in https://github.com/rustformers/llm/pull/389
- Fix #393 halting feed_prompt by @clarkmcc in https://github.com/rustformers/llm/pull/394
- Use llm_samplers crate for sampler backend by @KerfuffleV2 in https://github.com/rustformers/llm/pull/359
New Contributors
- @skirodev made their first contribution in https://github.com/rustformers/llm/pull/273
- @RedBoxing made their first contribution in https://github.com/rustformers/llm/pull/271
- @pdufour made their first contribution in https://github.com/rustformers/llm/pull/287
- @juliooa made their first contribution in https://github.com/rustformers/llm/pull/310
- @darxkies made their first contribution in https://github.com/rustformers/llm/pull/282
- @pixelspark made their first contribution in https://github.com/rustformers/llm/pull/311
- @nightscape made their first contribution in https://github.com/rustformers/llm/pull/326
- @JuliaMerz made their first contribution in https://github.com/rustformers/llm/pull/335
- @jafioti made their first contribution in https://github.com/rustformers/llm/pull/339
- @viirya made their first contribution in https://github.com/rustformers/llm/pull/348
- @steventrouble made their first contribution in https://github.com/rustformers/llm/pull/334
- @averypelle made their first contribution in https://github.com/rustformers/llm/pull/364
- @radu-matei made their first contribution in https://github.com/rustformers/llm/pull/367
- @basraayman made their first contribution in https://github.com/rustformers/llm/pull/379
- @baskrahmer made their first contribution in https://github.com/rustformers/llm/pull/380
- @clarkmcc made their first contribution in https://github.com/rustformers/llm/pull/394
Full Changelog: https://github.com/rustformers/llm/compare/v0.1.1...v0.2
Though I think this ci add on looks interesting https://intuit.github.io/auto/docs
Aye, that's a good start; we'll likely need to edit it to remove the redundancies and focus on the things that are relevant to users, but I really like the new contributors list!