Text Generation
PyTorch
English
olmo2
conversational
vwxyzjn commited on
Commit
c6be2af
·
verified ·
1 Parent(s): 8a83ddb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -2
README.md CHANGED
@@ -2,12 +2,16 @@
2
  license: apache-2.0
3
  language:
4
  - en
 
 
 
 
5
  pipeline_tag: text-generation
6
  ---
7
 
8
  <img alt="OLMo Logo" src="https://huggingface.co/datasets/allenai/blog-images/resolve/main/olmo2/olmo.png" width="242px">
9
 
10
- OLMo 2 32B Instruct March 2025 is post-trained variant of the [OLMo-2 32B March 2025](https://huggingface.co/allenai/OLMo-2-0325-32B/) model, which has undergone supervised finetuning on an OLMo-specific variant of the [Tülu 3 dataset](https://huggingface.co/datasets/allenai/tulu-3-sft-olmo-2-mixture) and further DPO training on [this dataset](https://huggingface.co/datasets/allenai/olmo-2-32b-pref-mix-v1).
11
  Tülu 3 is designed for state-of-the-art performance on a diversity of tasks in addition to chat, such as MATH, GSM8K, and IFEval.
12
  Check out the [OLMo 2 paper](https://arxiv.org/abs/2501.00656) or [Tülu 3 paper](https://arxiv.org/abs/2411.15124) for more details!
13
 
@@ -20,7 +24,7 @@ These models are trained on the Dolma dataset. We are releasing all code, checkp
20
  - **Model type:** A model trained on a mix of publicly available, synthetic and human-created datasets.
21
  - **Language(s) (NLP):** Primarily English
22
  - **License:** Apache 2.0
23
- - **Finetuned from model:** allenai/OLMo-2-0325-32B
24
 
25
  ### Model Sources
26
 
 
2
  license: apache-2.0
3
  language:
4
  - en
5
+ datasets:
6
+ - allenai/olmo-2-0325-32b-preference-mix
7
+ base_model:
8
+ - allenai/OLMo-2-0325-32B-SFT
9
  pipeline_tag: text-generation
10
  ---
11
 
12
  <img alt="OLMo Logo" src="https://huggingface.co/datasets/allenai/blog-images/resolve/main/olmo2/olmo.png" width="242px">
13
 
14
+ OLMo 2 32B Instruct March 2025 is post-trained variant of the [OLMo-2 32B March 2025](https://huggingface.co/allenai/OLMo-2-0325-32B/) model, which has undergone supervised finetuning on an OLMo-specific variant of the [Tülu 3 dataset](https://huggingface.co/datasets/allenai/tulu-3-sft-olmo-2-mixture) and further DPO training on [this dataset](https://huggingface.co/datasets/allenai/olmo-2-0325-32b-preference-mix).
15
  Tülu 3 is designed for state-of-the-art performance on a diversity of tasks in addition to chat, such as MATH, GSM8K, and IFEval.
16
  Check out the [OLMo 2 paper](https://arxiv.org/abs/2501.00656) or [Tülu 3 paper](https://arxiv.org/abs/2411.15124) for more details!
17
 
 
24
  - **Model type:** A model trained on a mix of publicly available, synthetic and human-created datasets.
25
  - **Language(s) (NLP):** Primarily English
26
  - **License:** Apache 2.0
27
+ - **Finetuned from model:** allenai/OLMo-2-0325-32B-SFT
28
 
29
  ### Model Sources
30