Update README.md
Browse files
README.md
CHANGED
@@ -40,6 +40,12 @@ tags:
|
|
40 |
|
41 |
This repo contains GGML format model files for [NousResearch's Nous Hermes Llama2 70B](https://huggingface.co/NousResearch/Nous-Hermes-Llama2-70b).
|
42 |
|
|
|
|
|
|
|
|
|
|
|
|
|
43 |
GPU acceleration is now available for Llama 2 70B GGML files, with both CUDA (NVidia) and Metal (macOS). The following clients/libraries are known to work with these files, including with GPU acceleration:
|
44 |
* [llama.cpp](https://github.com/ggerganov/llama.cpp), commit `e76d630` and later.
|
45 |
* [text-generation-webui](https://github.com/oobabooga/text-generation-webui), the most widely used web UI.
|
@@ -68,7 +74,7 @@ GPU acceleration is now available for Llama 2 70B GGML files, with both CUDA (NV
|
|
68 |
<!-- compatibility_ggml start -->
|
69 |
## Compatibility
|
70 |
|
71 |
-
###
|
72 |
|
73 |
Or one of the other tools and libraries listed above.
|
74 |
|
@@ -142,6 +148,10 @@ sudo apt update -y && sudo apt install 7zip
|
|
142 |
|
143 |
## How to run in `llama.cpp`
|
144 |
|
|
|
|
|
|
|
|
|
145 |
I use the following command line; adjust for your tastes and needs:
|
146 |
|
147 |
```
|
|
|
40 |
|
41 |
This repo contains GGML format model files for [NousResearch's Nous Hermes Llama2 70B](https://huggingface.co/NousResearch/Nous-Hermes-Llama2-70b).
|
42 |
|
43 |
+
### Important note regarding GGML files.
|
44 |
+
|
45 |
+
The GGML format has now been superseded by GGUF. As of August 21st 2023, [llama.cpp](https://github.com/ggerganov/llama.cpp) no longer supports GGML models. Third party clients and libraries are expected to still support it for a time, but many may also drop support.
|
46 |
+
|
47 |
+
### About GGML
|
48 |
+
|
49 |
GPU acceleration is now available for Llama 2 70B GGML files, with both CUDA (NVidia) and Metal (macOS). The following clients/libraries are known to work with these files, including with GPU acceleration:
|
50 |
* [llama.cpp](https://github.com/ggerganov/llama.cpp), commit `e76d630` and later.
|
51 |
* [text-generation-webui](https://github.com/oobabooga/text-generation-webui), the most widely used web UI.
|
|
|
74 |
<!-- compatibility_ggml start -->
|
75 |
## Compatibility
|
76 |
|
77 |
+
### Compatible with [commit `e76d630`](https://github.com/ggerganov/llama.cpp/commit/e76d630df17e235e6b9ef416c45996765d2e36fb) until August 21st 2023.
|
78 |
|
79 |
Or one of the other tools and libraries listed above.
|
80 |
|
|
|
148 |
|
149 |
## How to run in `llama.cpp`
|
150 |
|
151 |
+
Make sure you are using `llama.cpp` from commit [dadbed99e65252d79f81101a392d0d6497b86caa](https://github.com/ggerganov/llama.cpp/commit/dadbed99e65252d79f81101a392d0d6497b86caa) or earlier.
|
152 |
+
|
153 |
+
For compatibility with latest llama.cpp, please use GGUF files instead.
|
154 |
+
|
155 |
I use the following command line; adjust for your tastes and needs:
|
156 |
|
157 |
```
|