docs: add v0.5.5 CHANGELOG.md
parent
8c669dee8e
commit
e6fb1b6ac0
|
|
@ -2,14 +2,10 @@
|
||||||
|
|
||||||
## Features
|
## Features
|
||||||
|
|
||||||
# v0.5.4
|
# v0.5.5
|
||||||
|
|
||||||
## Fixes and Improvements
|
## Fixes and Improvements
|
||||||
|
|
||||||
* Fix deadlock issue reported in https://github.com/TabbyML/tabby/issues/718
|
|
||||||
|
|
||||||
# v0.5.3
|
|
||||||
|
|
||||||
## Notice
|
## Notice
|
||||||
|
|
||||||
* llama.cpp backend (CPU, Metal) now requires a redownload of gguf model due to upstream format changes: https://github.com/TabbyML/tabby/pull/645 https://github.com/ggerganov/llama.cpp/pull/3252
|
* llama.cpp backend (CPU, Metal) now requires a redownload of gguf model due to upstream format changes: https://github.com/TabbyML/tabby/pull/645 https://github.com/ggerganov/llama.cpp/pull/3252
|
||||||
|
|
@ -26,6 +22,7 @@
|
||||||
* add `server.completion_timeout` to control the code completion interface timeout: https://github.com/TabbyML/tabby/pull/637
|
* add `server.completion_timeout` to control the code completion interface timeout: https://github.com/TabbyML/tabby/pull/637
|
||||||
* Cuda backend is switched to llama.cpp: https://github.com/TabbyML/tabby/pull/656
|
* Cuda backend is switched to llama.cpp: https://github.com/TabbyML/tabby/pull/656
|
||||||
* Tokenizer implementation is switched to llama.cpp, so tabby no longer need to download additional tokenizer file: https://github.com/TabbyML/tabby/pull/683
|
* Tokenizer implementation is switched to llama.cpp, so tabby no longer need to download additional tokenizer file: https://github.com/TabbyML/tabby/pull/683
|
||||||
|
* Fix deadlock issue reported in https://github.com/TabbyML/tabby/issues/718
|
||||||
|
|
||||||
# v0.4.0
|
# v0.4.0
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue