Go to file
Meng Zhang 1f860aae79
Update issue templates
2023-04-09 21:08:43 +08:00
.github Update issue templates 2023-04-09 21:08:43 +08:00
clients feat: improve vim extension user experience (#71) 2023-04-09 20:26:54 +08:00
deployment Update README.md 2023-04-07 05:34:10 +08:00
development feat: hide streamlit menu in production (#50) 2023-04-06 17:21:26 +08:00
docs feat: add star badge for admin / openai page. 2023-04-06 20:36:04 +08:00
tabby chore: add pre-commit workflow (#72) 2023-04-09 20:17:56 +08:00
tests test: support TABBY_API_HOST in k6 tests 2023-04-04 11:14:22 +08:00
.dockerignore Add gptj converter (#19) 2023-03-27 11:12:52 +08:00
.gitattributes Add docker compose (#3) 2023-03-22 02:42:47 +08:00
.gitignore Add supervisord.pid to gitignore 2023-03-29 16:41:18 +08:00
.pre-commit-config.yaml feat: support stopping words in python backend. (#32) 2023-03-29 20:23:11 +08:00
Dockerfile feat: integrate caddy, re-org paths (#49) 2023-04-06 17:02:10 +08:00
LICENSE Create LICENSE 2023-03-16 17:28:10 +08:00
Makefile test: support TABBY_API_HOST in k6 tests 2023-04-04 11:14:22 +08:00
README.md Update README.md 2023-04-09 18:31:32 +08:00
poetry.lock feat: api endpoint url in admin page. (#70) 2023-04-09 20:00:22 +08:00
pyproject.toml feat: api endpoint url in admin page. (#70) 2023-04-09 20:00:22 +08:00

README.md

🐾 Tabby

License Code style: black Docker build status Docker pulls

architecture

Self-hosted AI coding assistant. An opensource / on-prem alternative to GitHub Copilot.

Warning Tabby is still in the alpha phase

Features

  • Self-contained, with no need for a DBMS or cloud service
  • Web UI for visualizing and configuration models and MLOps.
  • OpenAPI interface, easy to integrate with existing infrastructure (e.g Cloud IDE).
  • Consumer level GPU supports (FP-16 weight loading with various optimization).

Demo

Open in Spaces

Demo

Get started

Docker

NOTE: Tabby requires Pascal or newer NVIDIA GPU.

Before running Tabby, ensure the installation of the NVIDIA Container Toolkit. We suggest using NVIDIA drivers that are compatible with CUDA version 11.8 or higher.

# Create data dir and grant owner to 1000 (Tabby run as uid 1000 in container)
mkdir -p data/hf_cache && chown -R 1000 data

docker run \
  --gpus all \
  -it --rm \
  -v "./data:/data" \
  -v "./data/hf_cache:/home/app/.cache/huggingface" \
  -p 5000:5000 \
  -e MODEL_NAME=TabbyML/J-350M \
  -e MODEL_BACKEND=triton \
  --name=tabby \
  tabbyml/tabby

You can then query the server using /v1/completions endpoint:

curl -X POST http://localhost:5000/v1/completions -H 'Content-Type: application/json' --data '{
    "prompt": "def binarySearch(arr, left, right, x):\n    mid = (left +"
}'

We also provides an interactive playground in admin panel localhost:5000/_admin

image

Skypilot

See deployment/skypilot/README.md

API documentation

Tabby opens an FastAPI server at localhost:5000, which embeds an OpenAPI documentation of the HTTP API.

Development

Go to development directory.

make dev

or

make dev-triton # Turn on triton backend (for cuda env developers)