diff --git a/clients/vscode/README.md b/clients/vscode/README.md index 1f9d6ed..a1eeecb 100644 --- a/clients/vscode/README.md +++ b/clients/vscode/README.md @@ -22,7 +22,7 @@ Try our online demo [here](https://tabbyml.github.io/tabby/playground). If you have installed the Tabby VSCode extension, you can follow the built-in walkthrough guides to get started. You can also reopen walkthrough page anytime by using command `Tabby: Getting Started`. -1. Setup the Tabby server: you can get a Tabby Cloud hosted server [here](https://app.tabbyml.com), or build your self-hosted Tabby server following [this guide](https://tabbyml.github.io/tabby/docs/self-hosting/). +1. Setup the Tabby server: you can get a Tabby Cloud hosted server [here](https://app.tabbyml.com), or build your self-hosted Tabby server following [this guide](https://tabby.tabbyml.com/docs/installation). 2. Use the command `Tabby: Specify API Endpoint of Tabby` to connect the extension to your Tabby server. If you are using a Tabby Cloud server endpoint, please follow the popup messages to complete authorization. Once setup is complete, Tabby will provide inline suggestions automatically, and you can accept suggestions by just pressing the Tab key. You can hover on the inline suggestion text to find more useful actions such as partially accepting by word or by line. diff --git a/website/docs/02-self-hosting/01-docker.mdx b/website/docs/02-self-hosting/01-docker.mdx deleted file mode 100644 index 3e9aa10..0000000 --- a/website/docs/02-self-hosting/01-docker.mdx +++ /dev/null @@ -1,77 +0,0 @@ -# Docker - -There is a supplied docker image to make deploying a server as a container easier. - -import Tabs from '@theme/Tabs'; -import TabItem from '@theme/TabItem'; - -## CPU - - - - - ```bash title="run.sh" - docker run -it \ - -p 8080:8080 -v $HOME/.tabby:/data \ - tabbyml/tabby serve --model TabbyML/SantaCoder-1B - ``` - - - - -```yaml title="docker-compose.yml" -version: '3.5' - -services: - tabby: - restart: always - image: tabbyml/tabby - command: serve --model TabbyML/SantaCoder-1B - volumes: - - "$HOME/.tabby:/data" - ports: - - 8080:8080 -``` - - - - - -## CUDA (requires NVIDIA Container Toolkit) - - - - - ```bash title="run.sh" - docker run -it \ - --gpus all -p 8080:8080 -v $HOME/.tabby:/data \ - tabbyml/tabby \ - serve --model TabbyML/SantaCoder-1B --device cuda - ``` - - - - - ```yaml title="docker-compose.yml" - version: '3.5' - services: - tabby: - restart: always - image: tabbyml/tabby - command: serve --model TabbyML/SantaCoder-1B --device cuda - volumes: - - "$HOME/.tabby:/data" - ports: - - 8080:8080 - deploy: - resources: - reservations: - devices: - - driver: nvidia - count: 1 - capabilities: [gpu] - ``` - - - - diff --git a/website/docs/02-self-hosting/index.md b/website/docs/02-self-hosting/index.md deleted file mode 100644 index e0b8322..0000000 --- a/website/docs/02-self-hosting/index.md +++ /dev/null @@ -1,6 +0,0 @@ -# 📚 Self Hosting - -Tabby can be deployed on the server side using Docker or as client-side software, harnessing the computing power of Mac M1/M2. Please refer to each individual page for more information. - -* [Docker](./docker) -* [Mac M1/M2 (Preview)](./apple) diff --git a/website/docs/02-self-hosting/03-configuration.md b/website/docs/configuration.md similarity index 98% rename from website/docs/02-self-hosting/03-configuration.md rename to website/docs/configuration.md index ff57294..9020be6 100644 --- a/website/docs/02-self-hosting/03-configuration.md +++ b/website/docs/configuration.md @@ -1,4 +1,4 @@ -# Configuration +# ⚙️ Configuration :::tip The configuration file is not mandatory; Tabby can be run with just a single line of command. diff --git a/website/docs/extensions/index.mdx b/website/docs/extensions/index.mdx new file mode 100644 index 0000000..9c12acf --- /dev/null +++ b/website/docs/extensions/index.mdx @@ -0,0 +1,9 @@ +--- +sidebar_position: 3 +--- + +# 💻 IDE / Editor Extensions + +import DocCardList from '@theme/DocCardList'; + + diff --git a/website/docs/extensions/intellij.mdx b/website/docs/extensions/intellij.mdx new file mode 100644 index 0000000..c15f1ab --- /dev/null +++ b/website/docs/extensions/intellij.mdx @@ -0,0 +1,5 @@ +# IntelliJ Platform +# +import IntelliJ from "../../../clients/intellij/README.md"; + + diff --git a/website/docs/extensions/vim.mdx b/website/docs/extensions/vim.mdx new file mode 100644 index 0000000..be9d84b --- /dev/null +++ b/website/docs/extensions/vim.mdx @@ -0,0 +1,5 @@ +# VIM / NeoVIM + +import VIM from "../../../clients/vim/README.md"; + + diff --git a/website/docs/extensions/vscode.mdx b/website/docs/extensions/vscode.mdx new file mode 100644 index 0000000..0c73781 --- /dev/null +++ b/website/docs/extensions/vscode.mdx @@ -0,0 +1,8 @@ +--- +sidebar_position: 0 +--- +# Visual Studio Code + +import VSCode from "../../../clients/vscode/README.md"; + + diff --git a/website/docs/03-faq.mdx b/website/docs/faq.mdx similarity index 100% rename from website/docs/03-faq.mdx rename to website/docs/faq.mdx diff --git a/website/docs/01-getting-started.md b/website/docs/getting-started.md similarity index 98% rename from website/docs/01-getting-started.md rename to website/docs/getting-started.md index 67a4546..78f6d45 100644 --- a/website/docs/01-getting-started.md +++ b/website/docs/getting-started.md @@ -1,3 +1,6 @@ +--- +sidebar_position: 0 +--- # 👋 Getting Started Tabby is an open-source, self-hosted AI coding assistant. With Tabby, every team can set up its own LLM-powered code completion server with ease. diff --git a/website/docs/02-self-hosting/02-apple.md b/website/docs/installation/apple.md similarity index 77% rename from website/docs/02-self-hosting/02-apple.md rename to website/docs/installation/apple.md index cb4e6b0..1bdc0be 100644 --- a/website/docs/02-self-hosting/02-apple.md +++ b/website/docs/installation/apple.md @@ -1,4 +1,13 @@ -# Mac M1/M2 (Preview) +--- +sidebar_position: 3 +--- + +# Homebrew (Apple M1/M2) +This guide explains how to install Tabby using homebrew. + +:::info +Apple M1/M2 support is under **alpha** test. +::: Thanks to Apple's Accelerate and CoreML frameworks, we can now run Tabby on edge devices with reasonable inference speed. Follow the steps below to set it up using homebrew: @@ -10,6 +19,4 @@ brew install --HEAD tabby tabby serve --device metal --model TabbyML/CodeLlama-7B ``` -:::tip The compute power of M1/M2 is limited and is likely to be sufficient only for individual usage. If you require a shared instance for a team, we recommend considering Docker hosting with CUDA. You can find more information about Docker [here](./docker). -::: diff --git a/website/docs/installation/docker-compose.mdx b/website/docs/installation/docker-compose.mdx new file mode 100644 index 0000000..23ef7ef --- /dev/null +++ b/website/docs/installation/docker-compose.mdx @@ -0,0 +1,53 @@ +--- +sidebar_position: 1 +--- + +# Docker Compose +This guide explains how to launch Tabby using docker-compose. + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + + + + +```yaml title="docker-compose.yml" +version: '3.5' + +services: + tabby: + restart: always + image: tabbyml/tabby + command: serve --model TabbyML/SantaCoder-1B + volumes: + - "$HOME/.tabby:/data" + ports: + - 8080:8080 +``` + + + + +```yaml title="docker-compose.yml" +version: '3.5' +services: + tabby: + restart: always + image: tabbyml/tabby + command: serve --model TabbyML/SantaCoder-1B --device cuda + volumes: + - "$HOME/.tabby:/data" + ports: + - 8080:8080 + deploy: + resources: + reservations: + devices: + - driver: nvidia + count: 1 + capabilities: [gpu] +``` + + + + diff --git a/website/docs/installation/docker.mdx b/website/docs/installation/docker.mdx new file mode 100644 index 0000000..ab8cfb5 --- /dev/null +++ b/website/docs/installation/docker.mdx @@ -0,0 +1,32 @@ +--- +sidebar_position: 0 +--- + +# Docker + +This guide explains how to launch Tabby using docker. + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + + + + + ```bash title="run.sh" + docker run -it \ + -p 8080:8080 -v $HOME/.tabby:/data \ + tabbyml/tabby serve --model TabbyML/SantaCoder-1B + ``` + + + + + ```bash title="run.sh" + docker run -it \ + --gpus all -p 8080:8080 -v $HOME/.tabby:/data \ + tabbyml/tabby \ + serve --model TabbyML/SantaCoder-1B --device cuda + ``` + + + diff --git a/website/docs/installation/index.mdx b/website/docs/installation/index.mdx new file mode 100644 index 0000000..6c10555 --- /dev/null +++ b/website/docs/installation/index.mdx @@ -0,0 +1,9 @@ +--- +sidebar_position: 1 +--- + +import DocCardList from '@theme/DocCardList'; + +# 📚 Installation + + diff --git a/website/docs/04-roadmap.md b/website/docs/roadmap.md similarity index 100% rename from website/docs/04-roadmap.md rename to website/docs/roadmap.md