Safetensors
English
llama
trollek commited on
Commit
a70e9f3
·
verified ·
1 Parent(s): a48d2b6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -67
README.md CHANGED
@@ -1,17 +1,21 @@
1
- ---
2
- license: apache-2.0
3
- datasets:
4
- - SkunkworksAI/reasoning-0.01
5
- - trollek/ThoughtfulAssistant-v01
6
- - trollek/ThoughtfulAssistant-v02
7
- base_model: trollek/LittleInstructionMaker-4B-v0.2
8
- ---
 
 
9
  # ThoughtStream-4B-v0.2
10
 
11
- A more versitile version of [the first one](https://huggingface.co/trollek/ThoughtStream-4B-v0.1), and less versitile than the next[.](https://www.youtube.com/watch?v=dQw4w9WgXcQ "I swear my name makes sense in Danish, but.. I mean..") This can not compete with [Matt Shumer's](https://huggingface.co/mattshumer) [Reflection](https://huggingface.co/mattshumer/Reflection-Llama-3.1-70B), which be wildin' yo. But creating a dataset to "Please find errors and hallucinations in the thought process" is a tad harder than expected™. It is brilliant! And hard! Living in the future is great!
12
 
13
  With this second version I have tried to have 2 more ways of thinking than just stream of consciousness. This being mind mapping and step by step reasoning thanks to [SkunkworksAI](https://huggingface.co/SkunkworksAI) and their [reasoning-0.01](https://huggingface.co/datasets/SkunkworksAI/reasoning-0.01) dataset. Big up for creating [TinyFrank](https://huggingface.co/SkunkworksAI/tinyfrank-1.4B) that got me started on this wild adventure.
14
 
 
 
15
  ## System prompts
16
 
17
  **Stream of consciousness:**
@@ -26,61 +30,4 @@ With this second version I have tried to have 2 more ways of thinking than just
26
  **Step-by-step reasoning:**
27
 
28
  - `Your decision-making process is methodical and sequential, breaking down complex tasks into manageable steps to ensure thorough analysis and execution with precision and reliability. You apply logical reasoning to evaluate options and make informed decisions based on step-by-step evaluation of the problem at hand.`
29
- - `You are an AI assistant that think before you answer. When thinking you break your reasoning down step by step.`
30
-
31
-
32
- ## Config
33
-
34
- ```yaml
35
- ### model
36
- model_name_or_path: lim-v02-thought
37
-
38
- ### method
39
- stage: sft
40
- do_train: true
41
- finetuning_type: lora
42
- lora_target: all
43
- loraplus_lr_ratio: 12.0
44
- lora_rank: 16
45
- lora_alpha: 16
46
- use_unsloth: true
47
- quantization_bit: 4
48
- upcast_layernorm: true
49
- seed: 127
50
- optim: lion_8bit
51
- additional_target: embed_tokens
52
-
53
- ### dataset
54
- dataset: reasoning_assistant,thoughtful_v01,thoughtful_v02
55
- template: ninja_chatml
56
- cutoff_len: 8192
57
- overwrite_cache: false
58
- preprocessing_num_workers: 12
59
-
60
- ### output
61
- output_dir: /home/trolle/Documents/Projects/trollek/danube3/merges/lim-v02-thought/loras/reasoning
62
- logging_steps: 5
63
- save_steps: 1
64
- save_strategy: epoch
65
- plot_loss: true
66
- overwrite_output_dir: false
67
-
68
- ### train
69
- per_device_train_batch_size: 2
70
- gradient_accumulation_steps: 4
71
- learning_rate: 0.000002
72
- num_train_epochs: 2
73
- lr_scheduler_type: constant_with_warmup
74
- warmup_ratio: 0.01
75
- bf16: true
76
- flash_attn: fa2
77
-
78
- ### eval
79
- val_size: 0.01
80
- per_device_eval_batch_size: 1
81
- eval_strategy: steps
82
- eval_steps: 1000
83
-
84
- ```
85
-
86
- ## Training results
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - SkunkworksAI/reasoning-0.01
5
+ - trollek/ThoughtfulAssistant-v01
6
+ - trollek/ThoughtfulAssistant-v02
7
+ base_model: trollek/LittleInstructionMaker-4B-v0.2
8
+ language:
9
+ - en
10
+ ---
11
  # ThoughtStream-4B-v0.2
12
 
13
+ A more versitile version of [the first one](https://huggingface.co/trollek/ThoughtStream-4B-v0.1), and less versitile than the next[.](https://www.youtube.com/watch?v=dQw4w9WgXcQ "I swear my name makes sense in Danish, but.. I mean..")
14
 
15
  With this second version I have tried to have 2 more ways of thinking than just stream of consciousness. This being mind mapping and step by step reasoning thanks to [SkunkworksAI](https://huggingface.co/SkunkworksAI) and their [reasoning-0.01](https://huggingface.co/datasets/SkunkworksAI/reasoning-0.01) dataset. Big up for creating [TinyFrank](https://huggingface.co/SkunkworksAI/tinyfrank-1.4B) that got me started on this wild adventure.
16
 
17
+ This model is most useful for experimentation like "What happens if I substitute the thoughts?" etc. The answer is very well now. The first version was only trained on stream of consciousness, whereas this can handle mind maps and CoT reasoning. By itself it sucks at reasoning, but inject the CoT in the thinking tags and it performs better.
18
+
19
  ## System prompts
20
 
21
  **Stream of consciousness:**
 
30
  **Step-by-step reasoning:**
31
 
32
  - `Your decision-making process is methodical and sequential, breaking down complex tasks into manageable steps to ensure thorough analysis and execution with precision and reliability. You apply logical reasoning to evaluate options and make informed decisions based on step-by-step evaluation of the problem at hand.`
33
+ - `You are an AI assistant that think before you answer. When thinking you break your reasoning down step by step.`