Commit Graph

251 Commits (e87e78b74ccbcf2cf2cc73ade5ff54067220dd37)

Author SHA1 Message Date
Meng Zhang e87e78b74c fix: llama.cpp requires kv cache to be N_CTX * parallelism (#714) 2023-11-06 23:02:28 -08:00
Meng Zhang 7ca3221b52 fix: when there's an error happens in background inference loop, it should exit the process (#713) 2023-11-06 23:02:23 -08:00
Meng Zhang 03a9c7dac3 chore: add machete check to ensure no unused dependencies (#701)
* refactor: remove useless dependencies

* add machete
2023-11-06 23:02:12 -08:00
Meng Zhang b4fe249636 feat: support downloading resume (#700) 2023-11-04 20:12:18 -07:00
Meng Zhang 36d13d2837 fix(llama.cpp): wrongly index for n_seq in warmup 2023-11-04 20:12:13 -07:00
Meng Zhang 01ce18fe1a fix: llama.cpp warmp logic 2023-11-04 20:12:09 -07:00
Meng Zhang 536c7e86a0 Release 0.5.0
http-api-bindings@0.5.0
llama-cpp-bindings@0.5.0
tabby@0.5.0
tabby-common@0.5.0
tabby-download@0.5.0
tabby-inference@0.5.0
tabby-scheduler@0.5.0

Generated by cargo-workspaces
2023-11-03 18:00:38 -07:00
Meng Zhang 281d189848 Release 0.5.0-rc.4
http-api-bindings@0.5.0-rc.4
llama-cpp-bindings@0.5.0-rc.4
tabby@0.5.0-rc.4
tabby-common@0.5.0-rc.4
tabby-download@0.5.0-rc.4
tabby-inference@0.5.0-rc.4
tabby-scheduler@0.5.0-rc.4

Generated by cargo-workspaces
2023-11-03 17:36:51 -07:00
Meng Zhang c21ea483f4 fix: collect_snippet should handle NotReady error 2023-11-03 17:35:06 -07:00
Meng Zhang 03fe1e9f6b Release 0.5.0-rc.3
http-api-bindings@0.5.0-rc.3
llama-cpp-bindings@0.5.0-rc.3
tabby@0.5.0-rc.3
tabby-common@0.5.0-rc.3
tabby-download@0.5.0-rc.3
tabby-inference@0.5.0-rc.3
tabby-scheduler@0.5.0-rc.3

Generated by cargo-workspaces
2023-11-03 13:53:31 -07:00
Meng Zhang a5bb20becb Release 0.5.0-rc.2
http-api-bindings@0.5.0-rc.2
llama-cpp-bindings@0.5.0-rc.2
tabby@0.5.0-rc.2
tabby-common@0.5.0-rc.2
tabby-download@0.5.0-rc.2
tabby-inference@0.5.0-rc.2
tabby-scheduler@0.5.0-rc.2

Generated by cargo-workspaces
2023-11-03 13:26:51 -07:00
Meng Zhang f97cdf2ad9 Release 0.5.0-rc.1
http-api-bindings@0.5.0-rc.1
llama-cpp-bindings@0.5.0-rc.1
tabby@0.5.0-rc.1
tabby-common@0.5.0-rc.1
tabby-download@0.5.0-rc.1
tabby-inference@0.5.0-rc.1
tabby-scheduler@0.5.0-rc.1

Generated by cargo-workspaces
2023-11-03 13:23:25 -07:00
Meng Zhang 61605ca553 Release 0.5.0-rc.0
http-api-bindings@0.5.0-rc.0
llama-cpp-bindings@0.5.0-rc.0
tabby@0.5.0-rc.0
tabby-common@0.5.0-rc.0
tabby-download@0.5.0-rc.0
tabby-inference@0.5.0-rc.0
tabby-scheduler@0.5.0-rc.0

Generated by cargo-workspaces
2023-11-03 11:35:26 -07:00
Meng Zhang e4efcc4091
fix: avoid special keywords (e.g AND) failed the query parsing (#695) 2023-11-03 01:13:28 +00:00
Meng Zhang 2adcc0726c
feat: support prefix query on name field (#694)
* feat: support prefix phase query on name field

* update changelog
2023-11-03 01:04:33 +00:00
Meng Zhang acb3a33d78 fix: handle non utf-8 / utf-16 error 2023-11-02 16:29:30 -07:00
Meng Zhang eb34850a5e fix: output err if step failed 2023-11-02 16:15:11 -07:00
Meng Zhang 4c7eae584e
feat: add model warmup logic (#693) 2023-11-02 23:07:32 +00:00
Meng Zhang 0e4a2d2a12
feat: simplify download management, model file should be able to indi… (#690)
* feat: simplify download management, model file should be able to individually introduced

* fix typo

* update local model support

* update spec back

* update spec

* update

* update
2023-11-02 16:01:04 -07:00
Meng Zhang 36ffeb63f1 refactor: remove useless rust-cxx-cmake-bridge 2023-10-31 17:58:21 -07:00
Meng Zhang 296342efd8
refactor: use llama.cpp tokenizer (#683)
* refactor: switch to llama.cpp tokenizer to simplify implementation

* refactor: remove tokenizer dependency from tabby

* refactor: renaming decoding to stop condition

* refactor: remove tokenizer dependency

* refactor: remove submodule

* chore: update formatting

* move tokenization to c++
2023-10-31 22:16:09 +00:00
Meng Zhang f15926f233 chore(ui): update crates/tabby/ui 2023-10-31 08:51:22 -07:00
Meng Zhang 3f1f8bfd30 fix: add /swagger to tabby ui handler 2023-10-30 17:36:22 -07:00
Meng Zhang 73758e207d
feat: improve dashboard UI (#677) 2023-10-30 21:47:38 +00:00
Meng Zhang b4772fbcd0
feat(ui): add dashboard (#674)
* feat(ui): add dashboard

* handle path
2023-10-30 07:29:50 +00:00
Meng Zhang 89a63dbf33
fix: when send failed, treat the request as stopped (#673) 2023-10-30 06:27:09 +00:00
Meng Zhang de827b1e74
Revert "feat: make --model optional (#668)" (#672)
This reverts commit c55e4481ba.
2023-10-29 21:44:11 -07:00
leiwen83 b47bdd5d77
fix: align with fastchat mainstream (#670)
Fastchat mainstream change its return format, and text now is only
string in choices structure.

So make this change, to work with mainstream fastchat.

Signed-off-by: Lei Wen <wenlei03@qiyi.com>
Co-authored-by: Lei Wen <wenlei03@qiyi.com>
2023-10-29 19:31:46 -07:00
Meng Zhang 88d2617a34
fix: move events writer to individual thread (#669) 2023-10-30 01:31:41 +00:00
Meng Zhang c55e4481ba
feat: make --model optional (#668) 2023-10-30 00:04:42 +00:00
Meng Zhang 7330d75de6 chore: clear cache when there's no active requests 2023-10-29 16:30:30 -07:00
Meng Zhang 2ee5dbfd4f
chore: move tabby-ui under ee license. (#667)
* chore: introduce tabby-ui EE license.

* update
2023-10-29 15:56:57 -07:00
Meng Zhang 8c680a73fb
feat(ui): add /api page (#665)
* refactor(tabby-ui): extract tabby-fetcher

* feat(tabby-ui): add /api page

* feat(tabby-ui): add chat model badge

* fix: add components.json for shadcn

* chore: release tabby-ui
2023-10-29 14:55:50 -07:00
Meng Zhang 7bd99d14c0
feat: support continuous batching in llama.cpp backend (#659)
* refactor: switch back to llama batch interface

* feat: support cont batching
2023-10-28 23:37:05 -07:00
Meng Zhang 14d03b6826 fix(ui): handle invalid semver error 2023-10-28 23:30:35 -07:00
Meng Zhang 8dc5526091 Revert "feat: supports PHP (#634)"
This reverts commit 688e7d75b5.
2023-10-28 23:02:10 -07:00
Meng Zhang 43cc5f38cc feat: do not download ctranslate2 files in downloader 2023-10-28 02:27:09 -07:00
Meng Zhang 444222683a fix(llama.cpp): bump upstream fix for starcoder model on cuda 2023-10-28 02:03:34 -07:00
Meng Zhang 9309e0314f fix: fix docker build 2023-10-27 21:25:45 -07:00
Meng Zhang 6dd12ce1ec
fix: adding cuda search path to docker build. 2023-10-27 19:40:35 -07:00
Meng Zhang 2d948639be
fix: docker build for llama cuda backend 2023-10-27 16:36:54 -07:00
Meng Zhang 361091385d
Update CMakeLists.txt 2023-10-27 15:40:52 -07:00
Meng Zhang 23bd542cec
feat: switch cuda backend to llama.cpp (#656)
* feat: switch cuda backend to llama.cpp

* fix

* fix
2023-10-27 13:41:22 -07:00
Meng Zhang f37840566b
feat: upgrade llama.cpp (#645)
* feat: upgrade llama.cpp

* update download files

* update changelog

* Update CHANGELOG.md

* Update CHANGELOG.md
2023-10-27 12:18:46 -07:00
Oleksandr 688e7d75b5
feat: supports PHP (#634)
* feat: add php language configuration

* feat: add php repository context support

* docs: update documentation to indicate php support

* Update programming-languages.md

* added rev to tree-sitter-php

---------

Co-authored-by: Aleksandr Tarabaka <aleksandr.tarabaka@privatbank.ua>
Co-authored-by: Meng Zhang <meng@tabbyml.com>
2023-10-26 19:27:57 +00:00
Meng Zhang 62054cb4f1
feat: add new version notice in tabby-ui (#640)
* refactor: extract useHealth hook

* feat: add new version notice
2023-10-26 00:48:18 +00:00
Meng Zhang 9ecbf9031f
refactor: rename tabby-playground to tabby-ui (#639) 2023-10-25 23:05:11 +00:00
Meng Zhang 1a4c2aa71f
feat: swtich cpu backend to llama.cpp (#638)
* feat: swtich Cpu backend to llama.cpp

* feat: switch cpu serving to ggml

* fix cargo.toml

* use optional dependency

* fix compliation

* update ci target
2023-10-25 15:40:11 -07:00
Meng Zhang 21ec60eddf
feat: add server.completion_timeout to control timeout of /v1/completion (#637)
* feat: add server.completion_timeout to control timeout of /v1/completion

* Update config.rs
2023-10-25 22:05:23 +00:00
Meng Zhang d6296bb121 refactor: remove useless SwaggerConfig 2023-10-25 13:46:46 -07:00