freefallr commited on
Commit
107ca64
·
1 Parent(s): 9b6bc35

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -22
README.md CHANGED
@@ -14,38 +14,35 @@ datasets:
14
  - Christoph911/German-legal-SQuAD
15
  - philschmid/test_german_squad
16
  ---
17
- # Llama 2 13b Chat German | GGUF
18
 
19
- This repository contains the model [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) converted in GGUF format.
20
- The original model was created by [jphme](https://huggingface.co/jphme). It is a fine-tune of Meta's [Llama2 13b Chat](https://huggingface.co/meta-llama/Llama-2-13b-chat), trained on a dataset of instructions in German language.
21
 
22
- ## Model Profile
23
- The Model Profile describes relevant properties of the model in a standardized, digestible and easy-to-read way.
24
 
25
- |Property|Value|
26
  |----------------------------|--------------------------------------------------------------------------------------------------------------|
 
27
  | **Model** | [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) |
28
- | **Creator** | [jphme](https://huggingface.co/jphme) |
29
- | **Type** | LLM |
30
- | **Use Case** | Text Generation |
31
- | **Class** | Fine-tuned model |
32
- | **Parameters** | 70B |
33
- | **Finetuning Method** | Full |
34
- | **Finetuning Datasets** | Prorietary German Conversation Dataset, German SQuAD, German legal SQuAD data, augmented with "wrong" contexts, to improve factual RAG. For details see original model link. |
35
  | **Architecture** | Transformers |
36
- | **File Format** | GGUF |
 
 
 
 
37
  | **Quantization Types** | 8 Bit <br>5 Bit (K_M) |
38
  | **Utilized Tools** | llama.cpp (Commit 9e20231) for quantization to 8, 5 and 4 bit |
39
  | **Deployment** | #/bin/sh<br/>chmod +x run.sh && ./run.sh # Script to be published soon |
40
 
41
- **Details**
42
- We are using **AI Markdown Language *(.aiml)*** to standardize the description of our AI models, to make deploying and inferencing as easy and quick as possible.
43
-
44
- AIDF is a novel format proposed by Morgendigital that aims to summarize and standardize the information needed to run an AI model. More information on AIDF is coming soon.
45
-
46
- **Metadata**
47
- *Profile Type:* General AI Profile
48
- *Profile Version:* AIDF v1.0
49
 
50
  ## Replicate
51
  1. Clone and install llama.cpp *(Commit: 9e20231)*.
 
14
  - Christoph911/German-legal-SQuAD
15
  - philschmid/test_german_squad
16
  ---
17
+ # Summary
18
 
19
+ This repository contains the model [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) in GGUF format, for fast and easy inference with llama.cpp and similar LLM inference tools.
20
+ This model was created and trained by [jphme](https://huggingface.co/jphme). It is a fine-tuned variant of Meta's [Llama2 13b Chat](https://huggingface.co/meta-llama/Llama-2-13b-chat) with a compilation of multiple instruction datasets in German language.
21
 
22
+ ## Model Profile (.aiml)
 
23
 
24
+ |Attribute|Details|
25
  |----------------------------|--------------------------------------------------------------------------------------------------------------|
26
+ | **ID** | 1 |
27
  | **Model** | [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) |
28
+ | **Source** | https://huggingface.co/ |
29
+ | **Type** | Large Language Model |
30
+ | **Functions** | Text Generation |
31
+ | **Filetype** | GGUF |
32
+ | **Modifications** | Quantized FP16-accuracy model to 8, 4 (K_M) and 5 (K_M) Bit |
 
 
33
  | **Architecture** | Transformers |
34
+ | **Model** | [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) |
35
+ | **Creator** | [jphme](https://huggingface.co/jphme) |
36
+ | **Compression** | GGUF 8 Bit<br/>GGUF 5 Bit (K_M) |
37
+ | **Tuning** | full_finetune |
38
+ | **Tuning Datasets** | Prorietary German Conversation Dataset, German SQuAD, German legal SQuAD data, augmented with "wrong" contexts to improve factual RAG. |
39
  | **Quantization Types** | 8 Bit <br>5 Bit (K_M) |
40
  | **Utilized Tools** | llama.cpp (Commit 9e20231) for quantization to 8, 5 and 4 bit |
41
  | **Deployment** | #/bin/sh<br/>chmod +x run.sh && ./run.sh # Script to be published soon |
42
 
43
+ **About AI Model Profiles (.aiml Files)**
44
+ We are experimenting with a novel descriptive format called **AI Markdown Language *(.aiml)***. An AIML file contains all the necessary configuration parameters and rules to automatically deploy and serve production-ready AI models.
45
+ Our aim is to unify and standardize the description of AI models and make deployments and inference of AI much easier and quicker.
 
 
 
 
 
46
 
47
  ## Replicate
48
  1. Clone and install llama.cpp *(Commit: 9e20231)*.