This repository has been archived by the owner on Sep 12, 2024. It is now read-only.
Releases: Atome-FE/llama-node
Releases · Atome-FE/llama-node
v0.1.6
What's Changed
- feat: implement chat feature to rwkv by @yorkzero831 in #63
- feat: update cuda dynamic compiling by @hlhr202 in #66
- feature: optimiza llama.cpp loading, fix llama.cpp tokenizer, unify logger by @hlhr202 in #75
- update: refractor onnx by @hlhr202 in #87
- update: upgrade llm to 0.2.0-dev by @fardjad in #86
New Contributors
Full Changelog: v0.1.4...v0.1.6
v0.1.4
v0.1.1
v0.1.0
v0.0.37
v0.0.36
What's Changed
- feat: impl cache session and load session by @hlhr202 in #35
- feature: llama cpp new sampling by @hlhr202 in #38
- feat: new inference resolve/reject type by @hlhr202 in #39
- Typo fix by @Dinex-dev in #40
- feat: temp workaround for cublas build by @hlhr202 in #42
- fix: rwkv cpp windows failing by @hlhr202 in #48
New Contributors
- @Dinex-dev made their first contribution in #40
- @yorkzero831 made their first contribution in #46
Full Changelog: v0.0.33...v0.0.36