doubledsbv commited on
Commit
b7a8b53
·
verified ·
1 Parent(s): 3a3eaaf

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -1
README.md CHANGED
@@ -17,6 +17,9 @@ base_model:
17
  pipeline_tag: text-generation
18
  ---
19
 
 
 
 
20
  # Model Description
21
 
22
  **KafkaLM‑15B‑Base** is a 15‑billion‑parameter, sparsity‑aware language model distilled from *Mistral‑Small‑24B‑Base‑2501* and further post trained (SFT + DPO + GRPO /w verifiable rewards).
@@ -86,7 +89,8 @@ Up to 40 % parameter reduction (24 B → 15 B) delivers 2× lower TTFT
86
  | Tokens / s | 579 | **812** | +40% |
87
 
88
 
89
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/645ded34a45b4182d7f5c385/4rDhaeC-1GMj6KWbB27f9.png)
 
90
 
91
  ### Training scalability (distillation run, MI300A cluster)
92
 
 
17
  pipeline_tag: text-generation
18
  ---
19
 
20
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/645ded34a45b4182d7f5c385/EgsjPDWd37LjAtamiICxk.png" width="480" height="480" alt="image/png">
21
+
22
+
23
  # Model Description
24
 
25
  **KafkaLM‑15B‑Base** is a 15‑billion‑parameter, sparsity‑aware language model distilled from *Mistral‑Small‑24B‑Base‑2501* and further post trained (SFT + DPO + GRPO /w verifiable rewards).
 
89
  | Tokens / s | 579 | **812** | +40% |
90
 
91
 
92
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/645ded34a45b4182d7f5c385/4rDhaeC-1GMj6KWbB27f9.png" width="480" height="480" alt="image/png">
93
+
94
 
95
  ### Training scalability (distillation run, MI300A cluster)
96