0.13.0
Major Changes
- June 2024 Binary Update by @martindevans in #751
- Add Rosetta2 Binaries by @abhiaagarwal in #755
- Optimize the native library loading. by @AsakusaRinne in #688
- Add experimental auto-download support. by @AsakusaRinne in #692
- Upload native libraries to huggingface when release new version. by @AsakusaRinne in #749
- Batch Queueing by @martindevans in #761
- BatchedExecutor LLava by @martindevans in #770
- StreamingTextDecoder Special Token Support by @martindevans in #777
Documentation
- Updated READMEs. by @Lamothe in #742
- Add documentation for the change of native library loading. by @AsakusaRinne in #747
- Add OAI demo link by @sangyuxiaowu in #778
Bug Fixes
- KernelMemory bug fix by @zsogitbe in #726
- Fixes for website when running on case-sensitive filesystems. by @Lamothe in #753
Other Changes
- Exposed basic timing information from llama.cpp by @martindevans in #736
- Less Sampler Allocations by @martindevans in #735
- Several updates to web project by @Lamothe in #718
- Conversation Generate All Logits by @martindevans in #743
- Removed .NET 7 now that it's reached EOL. by @Lamothe in #741
- BatchedExecutor Double Buffering by @martindevans in #748
- Minor BatchedExecutor Example Improvements by @martindevans in #750
- LLava API Improvements by @martindevans in #757
- SemanticKernel HomeAutomation Example by @zsogitbe in #769
New Contributors
- @Lamothe made their first contribution in #718
- @sangyuxiaowu made their first contribution in #778
- @abhiaagarwal made their first contribution in #755
Full Changelog: v0.12.0...v0.13.0