Update README.md
Browse files
README.md
CHANGED
@@ -11,14 +11,18 @@ tags:
|
|
11 |
- CEREBORN
|
12 |
- Conversations
|
13 |
- Classification
|
|
|
|
|
|
|
|
|
14 |
model-index:
|
15 |
- name: CEREBORN-german
|
16 |
results: []
|
17 |
-
|
18 |
---
|
19 |
# ✅ Model Card for CEREBORN_german
|
20 |
|
21 |
-
|
|
|
22 |
|
23 |
**CEREBORN-german** is a neat little model built on top of **Phi 3.5 4B Instruct**, fine-tuned via LoRA on an A100 using carefully curated data.
|
24 |
We ended up adjusting about **5.5%** of the parameters, hit a **0.76 loss** on our eval set, and chugged through **1.2 billion tokens** during training.
|
@@ -72,4 +76,4 @@ Here are some unedited examples:
|
|
72 |
The model was trained **entirely sustainable** on hyperstack.
|
73 |
|
74 |
# ✅ Sources
|
75 |
-
CEREBORN-german is a fine-tuned version of [microsoft/Phi-3.5-mini-instruct](https://huggingface.co/microsoft/Phi-3.5-mini-instruct).
|
|
|
11 |
- CEREBORN
|
12 |
- Conversations
|
13 |
- Classification
|
14 |
+
- reasoning
|
15 |
+
- memory
|
16 |
+
- ger
|
17 |
+
- ggup
|
18 |
model-index:
|
19 |
- name: CEREBORN-german
|
20 |
results: []
|
|
|
21 |
---
|
22 |
# ✅ Model Card for CEREBORN_german
|
23 |
|
24 |
+
**3Step Version**
|
25 |
+
I just (14.03.25) added a new "3step Model" als **GGUP** that implements a *3 step reasoning, answer and remembering process* for cereborn-german.
|
26 |
|
27 |
**CEREBORN-german** is a neat little model built on top of **Phi 3.5 4B Instruct**, fine-tuned via LoRA on an A100 using carefully curated data.
|
28 |
We ended up adjusting about **5.5%** of the parameters, hit a **0.76 loss** on our eval set, and chugged through **1.2 billion tokens** during training.
|
|
|
76 |
The model was trained **entirely sustainable** on hyperstack.
|
77 |
|
78 |
# ✅ Sources
|
79 |
+
CEREBORN-german is a fine-tuned version of [microsoft/Phi-3.5-mini-instruct](https://huggingface.co/microsoft/Phi-3.5-mini-instruct).
|