RJuro commited on
Commit
4f16bbe
·
1 Parent(s): 11ceb40

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -1
README.md CHANGED
@@ -7,7 +7,10 @@ language:
7
  - en
8
  ---
9
 
10
- # Model name: kanelsnegl-v0.1
 
 
 
11
 
12
  ## Model Description
13
  Base model: [Zephyr-7b-alpha](https://huggingface.co/HuggingFaceH4/zephyr-7b-alpha) finetuned on [DDSC/partial-danish-gigaword-no-twitter](https://huggingface.co/datasets/DDSC/partial-danish-gigaword-no-twitter). The training involved a maximum length of 512. QLora completion finetuning of all linear layers was also implemented. This model is mostly fun tinkering for personal learning purpose. A heavily instruction-tuned model was used in the hope of transferring some of its behaviors into the Danish finetune.
 
7
  - en
8
  ---
9
 
10
+ # Model Card for kanelsnegl-v0.1
11
+
12
+
13
+ <img src="https://huggingface.co/RJuro/kanelsnegl-v0.1/resolve/main/kanelsnegl_banner.png" alt="Kanelsnegl Logo" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
14
 
15
  ## Model Description
16
  Base model: [Zephyr-7b-alpha](https://huggingface.co/HuggingFaceH4/zephyr-7b-alpha) finetuned on [DDSC/partial-danish-gigaword-no-twitter](https://huggingface.co/datasets/DDSC/partial-danish-gigaword-no-twitter). The training involved a maximum length of 512. QLora completion finetuning of all linear layers was also implemented. This model is mostly fun tinkering for personal learning purpose. A heavily instruction-tuned model was used in the hope of transferring some of its behaviors into the Danish finetune.