freefallr commited on
Commit
534e719
·
1 Parent(s): 528232f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -6
README.md CHANGED
@@ -21,15 +21,14 @@ The original model was created by [jphme](https://huggingface.co/jphme) and is a
21
 
22
 
23
  ## Model Profile
24
-
25
- | | |
26
  |----------------------------|--------------------------------------------------------------------------------------------------------------|
27
  | **Model** | [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) |
28
  | **Format** | GGUF |
29
- | **Quantization Levels** | 8 Bit<br> 5 Bit K_M |
30
- | **Conversion Tool used** | llama.cpp (Commit: 9e20231) |
31
- | **Original Model Creator** | [jphme](https://huggingface.co/jphme) |
32
- | **Training Data Info** | Prorietary German Conversation Dataset, German SQuAD, German legal SQuAD data, augmented with "wrong" contexts, to improve factual RAG |
33
 
34
  ## Replication Steps
35
  Clone and install llama.cpp *(Commit: 9e20231)* and use the provided `convert.py` file to convert the original model to GGUF with FP16 precision. The converted model will then be used to do further quantization.
 
21
 
22
 
23
  ## Model Profile
24
+ The model profile describes the properties
 
25
  |----------------------------|--------------------------------------------------------------------------------------------------------------|
26
  | **Model** | [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) |
27
  | **Format** | GGUF |
28
+ | **Quantization Types** | - 8 Bit <br>- 5 Bit K_M |
29
+ | **Conversion Tool** | llama.cpp (Commit: 9e20231) |
30
+ | **Original Model Creator** | [jphme](https://huggingface.co/jphme) |
31
+ | **Training Data** | Prorietary German Conversation Dataset, German SQuAD, German legal SQuAD data, augmented with "wrong" contexts, to improve factual RAG |
32
 
33
  ## Replication Steps
34
  Clone and install llama.cpp *(Commit: 9e20231)* and use the provided `convert.py` file to convert the original model to GGUF with FP16 precision. The converted model will then be used to do further quantization.