nielsr HF Staff commited on
Commit
48777f6
·
verified ·
1 Parent(s): 4ba9dee

Set pipeline tag to text-generation

Browse files

This PR sets the pipeline tag to `text-generation` so that the model is discoverable by text generation users.

Files changed (1) hide show
  1. README.md +6 -5
README.md CHANGED
@@ -1,7 +1,9 @@
1
  ---
 
 
 
2
  library_name: transformers
3
  license: apache-2.0
4
- base_model: Qwen/Qwen2.5-7B-Instruct
5
  tags:
6
  - llama-factory
7
  - full
@@ -9,8 +11,7 @@ tags:
9
  model-index:
10
  - name: OpenThinker-7B
11
  results: []
12
- datasets:
13
- - open-thoughts/open-thoughts-114k
14
  ---
15
 
16
  <p align="center">
@@ -26,7 +27,7 @@ This model is a fine-tuned version of [Qwen/Qwen2.5-7B-Instruct](https://hugging
26
  [OpenThoughts-114k dataset](https://huggingface.co/datasets/open-thoughts/OpenThoughts-114k) dataset.
27
 
28
  The dataset is derived by distilling DeepSeek-R1 using the [data pipeline available on github](https://github.com/open-thoughts/open-thoughts).
29
- More info about the dataset can be found on the dataset card at [OpenThoughts-114k dataset](https://huggingface.co/datasets/open-thoughts/open-thoughts-114k).
30
 
31
  This model improves upon the [Bespoke-Stratos-7B model](https://huggingface.co/bespokelabs/Bespoke-Stratos-7B), which used 17k examples ([Bespoke-Stratos-17k dataset](https://huggingface.co/datasets/bespokelabs/Bespoke-Stratos-17k)).
32
  The numbers reported in the table below are evaluated with our open-source tool [Evalchemy](https://github.com/mlfoundations/Evalchemy).
@@ -107,4 +108,4 @@ More info can be found in our repository: [https://github.com/open-thoughts/open
107
  primaryClass={cs.LG},
108
  url={https://arxiv.org/abs/2506.04178},
109
  }
110
- ```
 
1
  ---
2
+ base_model: Qwen/Qwen2.5-7B-Instruct
3
+ datasets:
4
+ - open-thoughts/open-thoughts-114k
5
  library_name: transformers
6
  license: apache-2.0
 
7
  tags:
8
  - llama-factory
9
  - full
 
11
  model-index:
12
  - name: OpenThinker-7B
13
  results: []
14
+ pipeline_tag: text-generation
 
15
  ---
16
 
17
  <p align="center">
 
27
  [OpenThoughts-114k dataset](https://huggingface.co/datasets/open-thoughts/OpenThoughts-114k) dataset.
28
 
29
  The dataset is derived by distilling DeepSeek-R1 using the [data pipeline available on github](https://github.com/open-thoughts/open-thoughts).
30
+ More info about the dataset can be found on the dataset card at [OpenThoughts-114k dataset](https://huggingface.co/datasets/open-thoughts/OpenThoughts-114k).
31
 
32
  This model improves upon the [Bespoke-Stratos-7B model](https://huggingface.co/bespokelabs/Bespoke-Stratos-7B), which used 17k examples ([Bespoke-Stratos-17k dataset](https://huggingface.co/datasets/bespokelabs/Bespoke-Stratos-17k)).
33
  The numbers reported in the table below are evaluated with our open-source tool [Evalchemy](https://github.com/mlfoundations/Evalchemy).
 
108
  primaryClass={cs.LG},
109
  url={https://arxiv.org/abs/2506.04178},
110
  }
111
+ ```