Update README.md
Browse files
README.md
CHANGED
@@ -1,17 +1,21 @@
|
|
1 |
-
---
|
2 |
-
license: apache-2.0
|
3 |
-
datasets:
|
4 |
-
- SkunkworksAI/reasoning-0.01
|
5 |
-
- trollek/ThoughtfulAssistant-v01
|
6 |
-
- trollek/ThoughtfulAssistant-v02
|
7 |
-
base_model: trollek/LittleInstructionMaker-4B-v0.2
|
8 |
-
|
|
|
|
|
9 |
# ThoughtStream-4B-v0.2
|
10 |
|
11 |
-
A more versitile version of [the first one](https://huggingface.co/trollek/ThoughtStream-4B-v0.1), and less versitile than the next[.](https://www.youtube.com/watch?v=dQw4w9WgXcQ "I swear my name makes sense in Danish, but.. I mean..")
|
12 |
|
13 |
With this second version I have tried to have 2 more ways of thinking than just stream of consciousness. This being mind mapping and step by step reasoning thanks to [SkunkworksAI](https://huggingface.co/SkunkworksAI) and their [reasoning-0.01](https://huggingface.co/datasets/SkunkworksAI/reasoning-0.01) dataset. Big up for creating [TinyFrank](https://huggingface.co/SkunkworksAI/tinyfrank-1.4B) that got me started on this wild adventure.
|
14 |
|
|
|
|
|
15 |
## System prompts
|
16 |
|
17 |
**Stream of consciousness:**
|
@@ -26,61 +30,4 @@ With this second version I have tried to have 2 more ways of thinking than just
|
|
26 |
**Step-by-step reasoning:**
|
27 |
|
28 |
- `Your decision-making process is methodical and sequential, breaking down complex tasks into manageable steps to ensure thorough analysis and execution with precision and reliability. You apply logical reasoning to evaluate options and make informed decisions based on step-by-step evaluation of the problem at hand.`
|
29 |
-
- `You are an AI assistant that think before you answer. When thinking you break your reasoning down step by step.`
|
30 |
-
|
31 |
-
|
32 |
-
## Config
|
33 |
-
|
34 |
-
```yaml
|
35 |
-
### model
|
36 |
-
model_name_or_path: lim-v02-thought
|
37 |
-
|
38 |
-
### method
|
39 |
-
stage: sft
|
40 |
-
do_train: true
|
41 |
-
finetuning_type: lora
|
42 |
-
lora_target: all
|
43 |
-
loraplus_lr_ratio: 12.0
|
44 |
-
lora_rank: 16
|
45 |
-
lora_alpha: 16
|
46 |
-
use_unsloth: true
|
47 |
-
quantization_bit: 4
|
48 |
-
upcast_layernorm: true
|
49 |
-
seed: 127
|
50 |
-
optim: lion_8bit
|
51 |
-
additional_target: embed_tokens
|
52 |
-
|
53 |
-
### dataset
|
54 |
-
dataset: reasoning_assistant,thoughtful_v01,thoughtful_v02
|
55 |
-
template: ninja_chatml
|
56 |
-
cutoff_len: 8192
|
57 |
-
overwrite_cache: false
|
58 |
-
preprocessing_num_workers: 12
|
59 |
-
|
60 |
-
### output
|
61 |
-
output_dir: /home/trolle/Documents/Projects/trollek/danube3/merges/lim-v02-thought/loras/reasoning
|
62 |
-
logging_steps: 5
|
63 |
-
save_steps: 1
|
64 |
-
save_strategy: epoch
|
65 |
-
plot_loss: true
|
66 |
-
overwrite_output_dir: false
|
67 |
-
|
68 |
-
### train
|
69 |
-
per_device_train_batch_size: 2
|
70 |
-
gradient_accumulation_steps: 4
|
71 |
-
learning_rate: 0.000002
|
72 |
-
num_train_epochs: 2
|
73 |
-
lr_scheduler_type: constant_with_warmup
|
74 |
-
warmup_ratio: 0.01
|
75 |
-
bf16: true
|
76 |
-
flash_attn: fa2
|
77 |
-
|
78 |
-
### eval
|
79 |
-
val_size: 0.01
|
80 |
-
per_device_eval_batch_size: 1
|
81 |
-
eval_strategy: steps
|
82 |
-
eval_steps: 1000
|
83 |
-
|
84 |
-
```
|
85 |
-
|
86 |
-
## Training results
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
datasets:
|
4 |
+
- SkunkworksAI/reasoning-0.01
|
5 |
+
- trollek/ThoughtfulAssistant-v01
|
6 |
+
- trollek/ThoughtfulAssistant-v02
|
7 |
+
base_model: trollek/LittleInstructionMaker-4B-v0.2
|
8 |
+
language:
|
9 |
+
- en
|
10 |
+
---
|
11 |
# ThoughtStream-4B-v0.2
|
12 |
|
13 |
+
A more versitile version of [the first one](https://huggingface.co/trollek/ThoughtStream-4B-v0.1), and less versitile than the next[.](https://www.youtube.com/watch?v=dQw4w9WgXcQ "I swear my name makes sense in Danish, but.. I mean..")
|
14 |
|
15 |
With this second version I have tried to have 2 more ways of thinking than just stream of consciousness. This being mind mapping and step by step reasoning thanks to [SkunkworksAI](https://huggingface.co/SkunkworksAI) and their [reasoning-0.01](https://huggingface.co/datasets/SkunkworksAI/reasoning-0.01) dataset. Big up for creating [TinyFrank](https://huggingface.co/SkunkworksAI/tinyfrank-1.4B) that got me started on this wild adventure.
|
16 |
|
17 |
+
This model is most useful for experimentation like "What happens if I substitute the thoughts?" etc. The answer is very well now. The first version was only trained on stream of consciousness, whereas this can handle mind maps and CoT reasoning. By itself it sucks at reasoning, but inject the CoT in the thinking tags and it performs better.
|
18 |
+
|
19 |
## System prompts
|
20 |
|
21 |
**Stream of consciousness:**
|
|
|
30 |
**Step-by-step reasoning:**
|
31 |
|
32 |
- `Your decision-making process is methodical and sequential, breaking down complex tasks into manageable steps to ensure thorough analysis and execution with precision and reliability. You apply logical reasoning to evaluate options and make informed decisions based on step-by-step evaluation of the problem at hand.`
|
33 |
+
- `You are an AI assistant that think before you answer. When thinking you break your reasoning down step by step.`
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|