|
--- |
|
license: apache-2.0 |
|
--- |
|
|
|
# Zephyr 7B Beta Llamafiles |
|
## See [here](https://dev.to/timesurgelabs/llamafile-ai-integration-deployment-made-easy-44cg#how-to-use-llamafiles) for a guide on how to use llamafiles! |
|
* Original Model: [Zephyr 7B Beta](hhttps://huggingface.co/HuggingFaceH4/zephyr-7b-beta) |
|
* Quantized Model: [Zephyr 7B Beta GGUF](https://huggingface.co/TheBloke/zephyr-7B-beta-GGUF) |
|
* Llamafile Source Code: [Mozilla-Ocho/llamafile](https://github.com/Mozilla-Ocho/llamafile) |
|
+ Built with [Llamafile `5ea929c`](https://github.com/Mozilla-Ocho/llamafile/tree/5ea929c618e9a2b162d39d8cc1c91cb564934a9f) |
|
|
|
Both the server and the CLI are based on [TheBloke's Zephyr 7B Beta GGUF Q4_K_M](https://huggingface.co/TheBloke/zephyr-7B-beta-GGUF) model. |
|
|
|
## Usage |
|
|
|
**NOTE:** Due to the executable being greater than 4GB, it is currently not compatible with Windows. I will update with a Windows friendly version of Zephyr 7B Beta when I can. |
|
|
|
```bash |
|
# replace with the CLI if you prefer |
|
wget https://huggingface.co/TimeSurgeLabs/zephyr-7b-beta-llamafile/resolve/main/zephyr-beta-server.llamafile |
|
chmod +x zephyr-beta-server.llamafile |
|
./zephyr-beta-server.llamafile |
|
``` |
|
|