Update README.md
Browse files
README.md
CHANGED
@@ -14,38 +14,35 @@ datasets:
|
|
14 |
- Christoph911/German-legal-SQuAD
|
15 |
- philschmid/test_german_squad
|
16 |
---
|
17 |
-
#
|
18 |
|
19 |
-
This repository contains the model [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german)
|
20 |
-
|
21 |
|
22 |
-
## Model Profile
|
23 |
-
The Model Profile describes relevant properties of the model in a standardized, digestible and easy-to-read way.
|
24 |
|
25 |
-
|
|
26 |
|----------------------------|--------------------------------------------------------------------------------------------------------------|
|
|
|
27 |
| **Model** | [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) |
|
28 |
-
| **
|
29 |
-
| **Type** |
|
30 |
-
| **
|
31 |
-
| **
|
32 |
-
| **
|
33 |
-
| **Finetuning Method** | Full |
|
34 |
-
| **Finetuning Datasets** | Prorietary German Conversation Dataset, German SQuAD, German legal SQuAD data, augmented with "wrong" contexts, to improve factual RAG. For details see original model link. |
|
35 |
| **Architecture** | Transformers |
|
36 |
-
| **
|
|
|
|
|
|
|
|
|
37 |
| **Quantization Types** | 8 Bit <br>5 Bit (K_M) |
|
38 |
| **Utilized Tools** | llama.cpp (Commit 9e20231) for quantization to 8, 5 and 4 bit |
|
39 |
| **Deployment** | #/bin/sh<br/>chmod +x run.sh && ./run.sh # Script to be published soon |
|
40 |
|
41 |
-
**
|
42 |
-
We are
|
43 |
-
|
44 |
-
AIDF is a novel format proposed by Morgendigital that aims to summarize and standardize the information needed to run an AI model. More information on AIDF is coming soon.
|
45 |
-
|
46 |
-
**Metadata**
|
47 |
-
*Profile Type:* General AI Profile
|
48 |
-
*Profile Version:* AIDF v1.0
|
49 |
|
50 |
## Replicate
|
51 |
1. Clone and install llama.cpp *(Commit: 9e20231)*.
|
|
|
14 |
- Christoph911/German-legal-SQuAD
|
15 |
- philschmid/test_german_squad
|
16 |
---
|
17 |
+
# Summary
|
18 |
|
19 |
+
This repository contains the model [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) in GGUF format, for fast and easy inference with llama.cpp and similar LLM inference tools.
|
20 |
+
This model was created and trained by [jphme](https://huggingface.co/jphme). It is a fine-tuned variant of Meta's [Llama2 13b Chat](https://huggingface.co/meta-llama/Llama-2-13b-chat) with a compilation of multiple instruction datasets in German language.
|
21 |
|
22 |
+
## Model Profile (.aiml)
|
|
|
23 |
|
24 |
+
|Attribute|Details|
|
25 |
|----------------------------|--------------------------------------------------------------------------------------------------------------|
|
26 |
+
| **ID** | 1 |
|
27 |
| **Model** | [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) |
|
28 |
+
| **Source** | https://huggingface.co/ |
|
29 |
+
| **Type** | Large Language Model |
|
30 |
+
| **Functions** | Text Generation |
|
31 |
+
| **Filetype** | GGUF |
|
32 |
+
| **Modifications** | Quantized FP16-accuracy model to 8, 4 (K_M) and 5 (K_M) Bit |
|
|
|
|
|
33 |
| **Architecture** | Transformers |
|
34 |
+
| **Model** | [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) |
|
35 |
+
| **Creator** | [jphme](https://huggingface.co/jphme) |
|
36 |
+
| **Compression** | GGUF 8 Bit<br/>GGUF 5 Bit (K_M) |
|
37 |
+
| **Tuning** | full_finetune |
|
38 |
+
| **Tuning Datasets** | Prorietary German Conversation Dataset, German SQuAD, German legal SQuAD data, augmented with "wrong" contexts to improve factual RAG. |
|
39 |
| **Quantization Types** | 8 Bit <br>5 Bit (K_M) |
|
40 |
| **Utilized Tools** | llama.cpp (Commit 9e20231) for quantization to 8, 5 and 4 bit |
|
41 |
| **Deployment** | #/bin/sh<br/>chmod +x run.sh && ./run.sh # Script to be published soon |
|
42 |
|
43 |
+
**About AI Model Profiles (.aiml Files)**
|
44 |
+
We are experimenting with a novel descriptive format called **AI Markdown Language *(.aiml)***. An AIML file contains all the necessary configuration parameters and rules to automatically deploy and serve production-ready AI models.
|
45 |
+
Our aim is to unify and standardize the description of AI models and make deployments and inference of AI much easier and quicker.
|
|
|
|
|
|
|
|
|
|
|
46 |
|
47 |
## Replicate
|
48 |
1. Clone and install llama.cpp *(Commit: 9e20231)*.
|