File size: 5,229 Bytes
18220c5
db600ce
18220c5
 
 
 
 
 
db600ce
 
 
 
18220c5
 
 
db600ce
 
 
 
 
 
 
18220c5
 
db600ce
18220c5
db600ce
18220c5
db600ce
 
18220c5
db600ce
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
---
base_model: unsloth/deepscaler-1.5b-preview
tags:
- text-generation-inference
- transformers
- unsloth
- qwen2
- trl
- reasoning 
- finetune 
- edge-device 
- research 
license: apache-2.0
language:
- en
datasets:
- bespokelabs/Bespoke-Stratos-17k
- simplescaling/s1K
- cognitivecomputations/dolphin-r1
- openai/gsm8k
- PrimeIntellect/NuminaMath-QwQ-CoT-5M
library_name: transformers
---

![image](./image.webp)

# **mini-Cogito-R1**

## **Overview**
The **mini-Cogito-R1** is a lightweight, high-performance language model fine-tuned for **text generation**, **mathematical reasoning**, and **edge-device optimization**. Developed by **Daemontatox**, this model is based on the **unsloth/deepscaler-1.5b-preview** architecture and fine-tuned using the **Unsloth** framework and Huggingface's **TRL** library, achieving **2x faster training speeds** without compromising performance.

---

## **Key Features**
- **Efficient Training:** Leverages [Unsloth](https://github.com/unslothai/unsloth) for faster and more efficient fine-tuning.  
- **Optimized for Edge Devices:** Designed to run efficiently on resource-constrained devices, making it ideal for edge computing applications.  
- **Mathematical Reasoning:** Excels in tasks requiring logical and mathematical reasoning.  
- **Text Generation:** Capable of generating high-quality, coherent text for a variety of applications.  
- **Lightweight:** Despite its compact size (1.5B parameters), it delivers robust performance.  

---

## **Model Details**
- **Developed by:** Daemontatox  
- **Model Name:** mini-Cogito-R1  
- **License:** Apache-2.0  
- **Base Model:** unsloth/deepscaler-1.5b-preview  
- **Fine-Tuned From:** unsloth/deepscaler-1.5b-preview-unsloth-bnb-4bit  
- **Framework:** Unsloth + Huggingface TRL  
- **Language:** English  

---

## **Training Datasets**
The **mini-Cogito-R1** model was fine-tuned on a diverse set of high-quality datasets to enhance its reasoning, mathematical, and text-generation capabilities. These datasets include:

1. **PrimeIntellect/NuminaMath-QwQ-CoT-5M**  
   - A large-scale dataset focused on mathematical reasoning and chain-of-thought (CoT) problem-solving.  

2. **openai/gsm8k**  
   - A dataset of grade-school math problems designed to test mathematical reasoning and problem-solving skills.  

3. **cognitivecomputations/dolphin-r1**  
   - A dataset for instruction-following and reasoning tasks, enhancing the model's ability to understand and execute complex instructions.  

4. **simplescaling/s1K**  
   - A lightweight dataset for general-purpose text generation and reasoning tasks.  

5. **bespokelabs/Bespoke-Stratos-17k**  
   - A dataset tailored for edge-device optimization and efficient text generation.  

---

## **Use Cases**
- **Edge Computing:** Deploy on edge devices for real-time text generation and reasoning tasks.  
- **Educational Tools:** Assist in solving mathematical problems and logical reasoning exercises.  
- **Content Creation:** Generate high-quality text for blogs, articles, and creative writing.  
- **Research:** Explore efficient training techniques and lightweight model architectures.  

---

## **Performance**
The **mini-Cogito-R1** was fine-tuned **2x faster** using Unsloth's optimized training pipeline, making it a cost-effective solution for developers and researchers. It maintains high accuracy and efficiency, particularly in mathematical reasoning and text generation tasks.

---

## **How to Use**
You can load and use the model with Huggingface's `transformers` library:

```python
from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "Daemontatox/mini-Cogito-R1"
model = AutoModelForCausalLM.from_pretrained(model_name)
tokenizer = AutoTokenizer.from_pretrained(model_name)

inputs = tokenizer("Solve for x: 2x + 5 = 15", return_tensors="pt")
outputs = model.generate(**inputs)
print(tokenizer.decode(outputs[0]))
```

---

## **Acknowledgments**
- **Unsloth Team:** For their groundbreaking work on efficient model training.  
- **Huggingface:** For providing the TRL library and ecosystem.  
- **Open Source Community:** For continuous support and contributions.  

---

## **License**
This model is licensed under the **Apache-2.0** license. For more details, see the [LICENSE](https://www.apache.org/licenses/LICENSE-2.0) file.

---

## **Connect with the Developer**
- **GitHub:** [Daemontatox](https://github.com/Daemontatox)  
- **Huggingface Model Hub:** [mini-Cogito-R1](https://huggingface.co/Daemontatox/mini-Cogito-R1)  

---

[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)  

---

### **Dataset References**
- **NuminaMath-QwQ-CoT-5M:** [PrimeIntellect](https://huggingface.co/datasets/PrimeIntellect/NuminaMath-QwQ-CoT-5M)  
- **GSM8K:** [OpenAI](https://huggingface.co/datasets/openai/gsm8k)  
- **Dolphin-R1:** [Cognitive Computations](https://huggingface.co/datasets/cognitivecomputations/dolphin-r1)  
- **S1K:** [Simple Scaling](https://huggingface.co/datasets/simplescaling/s1K)  
- **Bespoke-Stratos-17k:** [Bespoke Labs](https://huggingface.co/datasets/bespokelabs/Bespoke-Stratos-17k)  

---