Update README.md
Browse files
README.md
CHANGED
@@ -7,7 +7,10 @@ language:
|
|
7 |
- en
|
8 |
---
|
9 |
|
10 |
-
# Model
|
|
|
|
|
|
|
11 |
|
12 |
## Model Description
|
13 |
Base model: [Zephyr-7b-alpha](https://huggingface.co/HuggingFaceH4/zephyr-7b-alpha) finetuned on [DDSC/partial-danish-gigaword-no-twitter](https://huggingface.co/datasets/DDSC/partial-danish-gigaword-no-twitter). The training involved a maximum length of 512. QLora completion finetuning of all linear layers was also implemented. This model is mostly fun tinkering for personal learning purpose. A heavily instruction-tuned model was used in the hope of transferring some of its behaviors into the Danish finetune.
|
|
|
7 |
- en
|
8 |
---
|
9 |
|
10 |
+
# Model Card for kanelsnegl-v0.1
|
11 |
+
|
12 |
+
|
13 |
+
<img src="https://huggingface.co/RJuro/kanelsnegl-v0.1/resolve/main/kanelsnegl_banner.png" alt="Kanelsnegl Logo" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
|
14 |
|
15 |
## Model Description
|
16 |
Base model: [Zephyr-7b-alpha](https://huggingface.co/HuggingFaceH4/zephyr-7b-alpha) finetuned on [DDSC/partial-danish-gigaword-no-twitter](https://huggingface.co/datasets/DDSC/partial-danish-gigaword-no-twitter). The training involved a maximum length of 512. QLora completion finetuning of all linear layers was also implemented. This model is mostly fun tinkering for personal learning purpose. A heavily instruction-tuned model was used in the hope of transferring some of its behaviors into the Danish finetune.
|