23 lines
747 B
Markdown
23 lines
747 B
Markdown
---
|
|
sidebar_position: 3
|
|
---
|
|
|
|
# Homebrew (Apple M1/M2)
|
|
This guide explains how to install Tabby using homebrew.
|
|
|
|
:::info
|
|
Apple M1/M2 support is under **alpha** test.
|
|
:::
|
|
|
|
Thanks to Apple's Accelerate and CoreML frameworks, we can now run Tabby on edge devices with reasonable inference speed. Follow the steps below to set it up using homebrew:
|
|
|
|
```bash
|
|
brew tap TabbyML/tabby
|
|
brew install --HEAD tabby
|
|
|
|
# Start server with CodeLlama
|
|
tabby serve --device metal --model TabbyML/CodeLlama-7B
|
|
```
|
|
|
|
The compute power of M1/M2 is limited and is likely to be sufficient only for individual usage. If you require a shared instance for a team, we recommend considering Docker hosting with CUDA. You can find more information about Docker [here](./docker).
|