Releases: SciSharp/LLamaSharp
Releases · SciSharp/LLamaSharp
v0.4.0 - Executor and ChatSession
Version 0.4.0 introduces many break changes. However we strongly recommend to upgrade to 0.4.0 because it provides better abstractions and stability by refactoring the framework. The backend v0.3.0
and v0.3.1
still works for LLamaSharp v0.4.0
.
The main changes:
- Add three-level abstractions:
LLamaModel
,LLamaExecutor
andChatSession
. - Fix the BUG of saving and loading state.
- Support saving/loading chat session directly.
- Add more flexible APIs in the chat session.
- Add detailed documentations: https://scisharp.github.io/LLamaSharp/0.4/
Acknowledge
During the development, thanks a lot for the help from @TheTerrasque ! His/Her fork gives us many inspirations. Besides, many thanks for the following contributors!
- MacOS Arm64 support by @SignalRT in #24
- Fixed a typo in FixedSizeQueue by @mlof in #25
- Document interfaces by @mlof in #26
New Contributors
v0.3.0 - Load and save state
- Support loading and saving state.
- Support tokenization and detokenization.
- Fix BUGs of instruct mode.
- break change:
n_parts
param is removed. - break change:
LLamaModelV1
is dropped. - Remove dependencies for third-party loggers.
- Verified model repo is added on huggingface.
- Optimize the examples.
v0.2.3 - Inference BUG Fix
Fix some strange behaviors of model inference.
v0.2.2 - Embedder
- Sync with the latest llama.cpp master branch.
- Add
LLamaEmbedder
to support to get the embeddings only. - Add
n_gpu_layers
andprompt_cache_all
param. - Split the package into main package + backend package.
v0.2.1 - Chat session, quantization and Web API
- Add basic APIs and chat session.
- Support quantization.
- Add Web API support.