Update README.md
Browse files
README.md
CHANGED
@@ -2,12 +2,16 @@
|
|
2 |
license: apache-2.0
|
3 |
language:
|
4 |
- en
|
|
|
|
|
|
|
|
|
5 |
pipeline_tag: text-generation
|
6 |
---
|
7 |
|
8 |
<img alt="OLMo Logo" src="https://huggingface.co/datasets/allenai/blog-images/resolve/main/olmo2/olmo.png" width="242px">
|
9 |
|
10 |
-
OLMo 2 32B Instruct March 2025 is post-trained variant of the [OLMo-2 32B March 2025](https://huggingface.co/allenai/OLMo-2-0325-32B/) model, which has undergone supervised finetuning on an OLMo-specific variant of the [Tülu 3 dataset](https://huggingface.co/datasets/allenai/tulu-3-sft-olmo-2-mixture) and further DPO training on [this dataset](https://huggingface.co/datasets/allenai/olmo-2-32b-
|
11 |
Tülu 3 is designed for state-of-the-art performance on a diversity of tasks in addition to chat, such as MATH, GSM8K, and IFEval.
|
12 |
Check out the [OLMo 2 paper](https://arxiv.org/abs/2501.00656) or [Tülu 3 paper](https://arxiv.org/abs/2411.15124) for more details!
|
13 |
|
@@ -20,7 +24,7 @@ These models are trained on the Dolma dataset. We are releasing all code, checkp
|
|
20 |
- **Model type:** A model trained on a mix of publicly available, synthetic and human-created datasets.
|
21 |
- **Language(s) (NLP):** Primarily English
|
22 |
- **License:** Apache 2.0
|
23 |
-
- **Finetuned from model:** allenai/OLMo-2-0325-32B
|
24 |
|
25 |
### Model Sources
|
26 |
|
|
|
2 |
license: apache-2.0
|
3 |
language:
|
4 |
- en
|
5 |
+
datasets:
|
6 |
+
- allenai/olmo-2-0325-32b-preference-mix
|
7 |
+
base_model:
|
8 |
+
- allenai/OLMo-2-0325-32B-SFT
|
9 |
pipeline_tag: text-generation
|
10 |
---
|
11 |
|
12 |
<img alt="OLMo Logo" src="https://huggingface.co/datasets/allenai/blog-images/resolve/main/olmo2/olmo.png" width="242px">
|
13 |
|
14 |
+
OLMo 2 32B Instruct March 2025 is post-trained variant of the [OLMo-2 32B March 2025](https://huggingface.co/allenai/OLMo-2-0325-32B/) model, which has undergone supervised finetuning on an OLMo-specific variant of the [Tülu 3 dataset](https://huggingface.co/datasets/allenai/tulu-3-sft-olmo-2-mixture) and further DPO training on [this dataset](https://huggingface.co/datasets/allenai/olmo-2-0325-32b-preference-mix).
|
15 |
Tülu 3 is designed for state-of-the-art performance on a diversity of tasks in addition to chat, such as MATH, GSM8K, and IFEval.
|
16 |
Check out the [OLMo 2 paper](https://arxiv.org/abs/2501.00656) or [Tülu 3 paper](https://arxiv.org/abs/2411.15124) for more details!
|
17 |
|
|
|
24 |
- **Model type:** A model trained on a mix of publicly available, synthetic and human-created datasets.
|
25 |
- **Language(s) (NLP):** Primarily English
|
26 |
- **License:** Apache 2.0
|
27 |
+
- **Finetuned from model:** allenai/OLMo-2-0325-32B-SFT
|
28 |
|
29 |
### Model Sources
|
30 |
|