tabby/website/docs/self-hosting/01-docker.md

1.0 KiB

Docker

There is a supplied docker image to make deploying a server as a container easier.

CPU

Command line

docker run \
  -p 8080:8080 -v $HOME/.tabby:/data \
  tabbyml/tabby serve --model TabbyML/SantaCoder-1B

Docker Compose

version: '3.5'

services:
  tabby:
    restart: always
    image: tabbyml/tabby
    command: serve --model TabbyML/SantaCoder-1B
    volumes:
      - "$HOME/.tabby:/data"
    ports:
      - 8080:8080

CUDA (requires NVIDIA Container Toolkit)

Command line

docker run \
  --gpus all -p 8080:8080 -v $HOME/.tabby:/data \
  tabbyml/tabby \
  serve --model TabbyML/SantaCoder-1B --device cuda

Docker Compose

version: '3.5'
services:
  tabby:
    restart: always
    image: tabbyml/tabby
    command: serve --model TabbyML/SantaCoder-1B --device cuda
    volumes:
      - "$HOME/.tabby:/data"
    ports:
      - 8080:8080
    resources:
    reservations:
      devices:
      - driver: nvidia
        count: 1
        capabilities: [gpu]