Spaces:
Runtime error
Runtime error
ztime
commited on
Commit
·
2a6826a
1
Parent(s):
0dc41c6
change port export
Browse files- Dockerfile +3 -2
- README.md +4 -4
- start_server.sh +1 -1
Dockerfile
CHANGED
|
@@ -18,9 +18,10 @@ RUN git clone https://github.com/ggerganov/llama.cpp.git llamacpp --depth 1 && \
|
|
| 18 |
cd llamacpp && \
|
| 19 |
cmake -B build -DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=OpenBLAS && \
|
| 20 |
cmake --build build --config Release --target main server && \
|
| 21 |
-
cp build/bin/*
|
| 22 |
touch /llama.log && \
|
| 23 |
-
chmod 777 /llama.log
|
|
|
|
| 24 |
|
| 25 |
# Download model
|
| 26 |
RUN mkdir model && \
|
|
|
|
| 18 |
cd llamacpp && \
|
| 19 |
cmake -B build -DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=OpenBLAS && \
|
| 20 |
cmake --build build --config Release --target main server && \
|
| 21 |
+
cp build/bin/* ../ && \
|
| 22 |
touch /llama.log && \
|
| 23 |
+
chmod 777 /llama.log && \
|
| 24 |
+
ls -lt
|
| 25 |
|
| 26 |
# Download model
|
| 27 |
RUN mkdir model && \
|
README.md
CHANGED
|
@@ -1,20 +1,20 @@
|
|
| 1 |
---
|
| 2 |
-
title:
|
| 3 |
colorFrom: purple
|
| 4 |
colorTo: blue
|
| 5 |
sdk: docker
|
| 6 |
models:
|
| 7 |
- teknium/OpenHermes-2.5-Mistral-7B
|
| 8 |
-
- TheBloke/
|
| 9 |
tags:
|
| 10 |
- inference api
|
| 11 |
- openai-api compatible
|
| 12 |
- llama-cpp-python
|
| 13 |
-
-
|
| 14 |
- gguf
|
| 15 |
pinned: false
|
| 16 |
---
|
| 17 |
|
| 18 |
-
#
|
| 19 |
|
| 20 |
Please refer to the [index.html](index.html) for more information.
|
|
|
|
| 1 |
---
|
| 2 |
+
title: openchat-3.5-0106-GGUF (Q4_K_M)
|
| 3 |
colorFrom: purple
|
| 4 |
colorTo: blue
|
| 5 |
sdk: docker
|
| 6 |
models:
|
| 7 |
- teknium/OpenHermes-2.5-Mistral-7B
|
| 8 |
+
- TheBloke/openchat-3.5-0106-GGUF
|
| 9 |
tags:
|
| 10 |
- inference api
|
| 11 |
- openai-api compatible
|
| 12 |
- llama-cpp-python
|
| 13 |
+
- openchat-3.5-0106-GGUF
|
| 14 |
- gguf
|
| 15 |
pinned: false
|
| 16 |
---
|
| 17 |
|
| 18 |
+
# openchat-3.5-0106-GGUF (Q4_K_M)
|
| 19 |
|
| 20 |
Please refer to the [index.html](index.html) for more information.
|
start_server.sh
CHANGED
|
@@ -3,7 +3,7 @@
|
|
| 3 |
# For mlock support
|
| 4 |
ulimit -l unlimited
|
| 5 |
|
| 6 |
-
|
| 7 |
|
| 8 |
/server --host 0.0.0.0 --port 7860 -m model/gguf-model.bin
|
| 9 |
/llamacpp/build/bin/server --host 0.0.0.0 --port 7860 -m model/gguf-model.bin
|
|
|
|
| 3 |
# For mlock support
|
| 4 |
ulimit -l unlimited
|
| 5 |
|
| 6 |
+
/main -m model/gguf-model.bin --color -p '你好,你是谁'
|
| 7 |
|
| 8 |
/server --host 0.0.0.0 --port 7860 -m model/gguf-model.bin
|
| 9 |
/llamacpp/build/bin/server --host 0.0.0.0 --port 7860 -m model/gguf-model.bin
|