60 lines
1.0 KiB
Markdown
60 lines
1.0 KiB
Markdown
|
|
# Docker
|
||
|
|
|
||
|
|
There is a supplied docker image to make deploying a server as a container easier.
|
||
|
|
|
||
|
|
|
||
|
|
## CPU
|
||
|
|
|
||
|
|
**Command line**
|
||
|
|
```bash
|
||
|
|
docker run \
|
||
|
|
-p 8080:8080 -v $HOME/.tabby:/data \
|
||
|
|
tabbyml/tabby serve --model TabbyML/SantaCoder-1B
|
||
|
|
```
|
||
|
|
|
||
|
|
|
||
|
|
**Docker Compose**
|
||
|
|
```yaml
|
||
|
|
version: '3.5'
|
||
|
|
|
||
|
|
services:
|
||
|
|
tabby:
|
||
|
|
restart: always
|
||
|
|
image: tabbyml/tabby
|
||
|
|
command: serve --model TabbyML/SantaCoder-1B
|
||
|
|
volumes:
|
||
|
|
- "$HOME/.tabby:/data"
|
||
|
|
ports:
|
||
|
|
- 8080:8080
|
||
|
|
```
|
||
|
|
|
||
|
|
## CUDA (requires NVIDIA Container Toolkit)
|
||
|
|
|
||
|
|
**Command line**
|
||
|
|
```bash
|
||
|
|
docker run \
|
||
|
|
--gpus all -p 8080:8080 -v $HOME/.tabby:/data \
|
||
|
|
tabbyml/tabby \
|
||
|
|
serve --model TabbyML/SantaCoder-1B --device cuda
|
||
|
|
```
|
||
|
|
|
||
|
|
**Docker Compose**
|
||
|
|
```yaml
|
||
|
|
version: '3.5'
|
||
|
|
services:
|
||
|
|
tabby:
|
||
|
|
restart: always
|
||
|
|
image: tabbyml/tabby
|
||
|
|
command: serve --model TabbyML/SantaCoder-1B --device cuda
|
||
|
|
volumes:
|
||
|
|
- "$HOME/.tabby:/data"
|
||
|
|
ports:
|
||
|
|
- 8080:8080
|
||
|
|
resources:
|
||
|
|
reservations:
|
||
|
|
devices:
|
||
|
|
- driver: nvidia
|
||
|
|
count: 1
|
||
|
|
capabilities: [gpu]
|
||
|
|
```
|