# Mac M1/M2 (Preview) Thanks to Apple's Accelerate and CoreML frameworks, we can now run Tabby on edge devices with reasonable inference speed. Follow the steps below to set it up: 1. Download the `tabby` binary from the Release page [here](https://github.com/TabbyML/tabby/releases/tag/latest). 2. Run `tabby --help` to verify successful installation. 3. Start the server with `tabby serve model --models TabbyML/T5P-220M`.