Spaces:
Runtime error
Runtime error
ztime
commited on
Commit
•
2a6826a
1
Parent(s):
0dc41c6
change port export
Browse files- Dockerfile +3 -2
- README.md +4 -4
- start_server.sh +1 -1
Dockerfile
CHANGED
@@ -18,9 +18,10 @@ RUN git clone https://github.com/ggerganov/llama.cpp.git llamacpp --depth 1 && \
|
|
18 |
cd llamacpp && \
|
19 |
cmake -B build -DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=OpenBLAS && \
|
20 |
cmake --build build --config Release --target main server && \
|
21 |
-
cp build/bin/*
|
22 |
touch /llama.log && \
|
23 |
-
chmod 777 /llama.log
|
|
|
24 |
|
25 |
# Download model
|
26 |
RUN mkdir model && \
|
|
|
18 |
cd llamacpp && \
|
19 |
cmake -B build -DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=OpenBLAS && \
|
20 |
cmake --build build --config Release --target main server && \
|
21 |
+
cp build/bin/* ../ && \
|
22 |
touch /llama.log && \
|
23 |
+
chmod 777 /llama.log && \
|
24 |
+
ls -lt
|
25 |
|
26 |
# Download model
|
27 |
RUN mkdir model && \
|
README.md
CHANGED
@@ -1,20 +1,20 @@
|
|
1 |
---
|
2 |
-
title:
|
3 |
colorFrom: purple
|
4 |
colorTo: blue
|
5 |
sdk: docker
|
6 |
models:
|
7 |
- teknium/OpenHermes-2.5-Mistral-7B
|
8 |
-
- TheBloke/
|
9 |
tags:
|
10 |
- inference api
|
11 |
- openai-api compatible
|
12 |
- llama-cpp-python
|
13 |
-
-
|
14 |
- gguf
|
15 |
pinned: false
|
16 |
---
|
17 |
|
18 |
-
#
|
19 |
|
20 |
Please refer to the [index.html](index.html) for more information.
|
|
|
1 |
---
|
2 |
+
title: openchat-3.5-0106-GGUF (Q4_K_M)
|
3 |
colorFrom: purple
|
4 |
colorTo: blue
|
5 |
sdk: docker
|
6 |
models:
|
7 |
- teknium/OpenHermes-2.5-Mistral-7B
|
8 |
+
- TheBloke/openchat-3.5-0106-GGUF
|
9 |
tags:
|
10 |
- inference api
|
11 |
- openai-api compatible
|
12 |
- llama-cpp-python
|
13 |
+
- openchat-3.5-0106-GGUF
|
14 |
- gguf
|
15 |
pinned: false
|
16 |
---
|
17 |
|
18 |
+
# openchat-3.5-0106-GGUF (Q4_K_M)
|
19 |
|
20 |
Please refer to the [index.html](index.html) for more information.
|
start_server.sh
CHANGED
@@ -3,7 +3,7 @@
|
|
3 |
# For mlock support
|
4 |
ulimit -l unlimited
|
5 |
|
6 |
-
|
7 |
|
8 |
/server --host 0.0.0.0 --port 7860 -m model/gguf-model.bin
|
9 |
/llamacpp/build/bin/server --host 0.0.0.0 --port 7860 -m model/gguf-model.bin
|
|
|
3 |
# For mlock support
|
4 |
ulimit -l unlimited
|
5 |
|
6 |
+
/main -m model/gguf-model.bin --color -p '你好,你是谁'
|
7 |
|
8 |
/server --host 0.0.0.0 --port 7860 -m model/gguf-model.bin
|
9 |
/llamacpp/build/bin/server --host 0.0.0.0 --port 7860 -m model/gguf-model.bin
|