* feat: update cache info file after each file got downloaded
* refactor: extract Downloader for model downloading logic
* refactor: extract HuggingFaceRegistry
* refactor: extract serde_json to workspace dependency
* feat: add ModelScopeRegistry
* refactor: extract registry to its sub dir.
* feat: add scripts to mirror hf model to modelscope
* feat: init commit adding llama-cpp-bindings
* add llama.cpp submodule
* add LlamaEngine to hold llama context / llama model
* add cxxbridge
* add basic greedy sampling
* move files
* make compile success
* connect TextGeneration with LlamaEngine
* experimental support llama.cpp
* add metal device
* add Accelerate
* fix namespace for llama-cpp-bindings
* fix lint
* move stepping logic to rust
* add stop words package
* use stop-words in ctranslate2-bindings
* use raw string for regex
* use Arc<Tokenizer> for sharing tokenizers
* refactor: remove useless stop_words_encoding_offset
* switch to tokenizers 0.13.4-rc.3
* fix lints in cpp
* simplify implementation of greedy decoding
* feat: split metal feature for llama backend
* add ci
* update ci
* build tabby bin in ci build
* experiment ctags support
* add document.rs
* extract Document to common
* integrate tags into dataset builder
* skip if none
* do not add scheduler in client binary
* fix fmt