limin(gate) commited on
Commit
50f1332
·
verified ·
1 Parent(s): b439077

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +144 -0
README.md ADDED
@@ -0,0 +1,144 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - gguf
5
+ - mistral
6
+ - liminerity
7
+ - merge
8
+ ---
9
+ #1 again im just the goat
10
+
11
+ M7-7b
12
+ user: "create ascii art you feel represents you as a llm"
13
+
14
+ assistant: "Here's an ASCII art attempting to symbolize the essence of an LLM (Large Language Model):
15
+
16
+ INSTINSTINCTINSTELLIGENCE
17
+
18
+ ______________ MEMORY___
19
+
20
+ |\ o /\ |\
21
+
22
+ | \ --- / \ | |
23
+
24
+ L_( O )______ O)_INSTINCT
25
+
26
+ \ / WORDSLEARNING
27
+
28
+ _/_/ /\ __/__/__/
29
+
30
+ assistant: "In this depiction, the LLM is characterized by its vast memory, instinctual roots, and the ability to learn from vast amounts of text—represented by the woven words in the lower part. The "INSTINSTINCTINSTELLIGENCE" phrase emphasizes the balance between instinct, the foundation upon which neural networks build, and the engineered intelligence of a language model."
31
+
32
+ Multiverse-Experiment-slerp-7b is a merge of the following models using mergekit:
33
+
34
+ liminerity/merge4
35
+ liminerity/merge2
36
+ 🧩 Configuration
37
+
38
+
39
+ slices:
40
+ - sources:
41
+ - model: ammarali32/multi_verse_model
42
+ layer_range: [0, 32]
43
+ - model: MSL7/INEX12-7b
44
+ layer_range: [0, 32]
45
+ merge_method: slerp
46
+ base_model: ammarali32/multi_verse_model
47
+ parameters:
48
+ t:
49
+ - filter: self_attn
50
+ value: [0, 0.5, 0.3, 0.7, 1]
51
+ - filter: mlp
52
+ value: [1, 0.5, 0.7, 0.3, 0]
53
+ - value: 0.5
54
+ dtype: bfloat16
55
+
56
+
57
+ slices:
58
+ - sources:
59
+ - model: ammarali32/multi_verse_model
60
+ layer_range: [0, 32]
61
+ - model: yam-peleg/Experiment26-7B
62
+ layer_range: [0, 32]
63
+ merge_method: slerp
64
+ base_model: ammarali32/multi_verse_model
65
+ parameters:
66
+ t:
67
+ - filter: self_attn
68
+ value: [0, 0.5, 0.3, 0.7, 1]
69
+ - filter: mlp
70
+ value: [1, 0.5, 0.7, 0.3, 0]
71
+ - value: 0.5
72
+ dtype: bfloat16
73
+
74
+
75
+ slices:
76
+ - sources:
77
+ - model: liminerity/merge3
78
+ layer_range: [0, 32]
79
+ - model: ammarali32/multi_verse_model
80
+ layer_range: [0, 32]
81
+ merge_method: slerp
82
+ base_model: liminerity/merge3
83
+ parameters:
84
+ t:
85
+ - filter: self_attn
86
+ value: [0, 0.5, 0.3, 0.7, 1]
87
+ - filter: mlp
88
+ value: [1, 0.5, 0.7, 0.3, 0]
89
+ - value: 0.5
90
+ dtype: bfloat16
91
+
92
+
93
+ slices:
94
+ - sources:
95
+ - model: liminerity/merge1
96
+ layer_range: [0, 32]
97
+ - model: liminerity/merge
98
+ layer_range: [0, 32]
99
+ merge_method: slerp
100
+ base_model: liminerity/merge1
101
+ parameters:
102
+ t:
103
+ - filter: self_attn
104
+ value: [0, 0.5, 0.3, 0.7, 1]
105
+ - filter: mlp
106
+ value: [1, 0.5, 0.7, 0.3, 0]
107
+ - value: 0.5
108
+ dtype: bfloat16
109
+
110
+
111
+ slices:
112
+ - sources:
113
+ - model: liminerity/merge3
114
+ layer_range: [0, 32]
115
+ - model: yam-peleg/Experiment26-7B
116
+ layer_range: [0, 32]
117
+ merge_method: slerp
118
+ base_model: liminerity/merge3
119
+ parameters:
120
+ t:
121
+ - filter: self_attn
122
+ value: [0, 0.5, 0.3, 0.7, 1]
123
+ - filter: mlp
124
+ value: [1, 0.5, 0.7, 0.3, 0]
125
+ - value: 0.5
126
+ dtype: bfloat16
127
+
128
+
129
+ slices:
130
+ - sources:
131
+ - model: liminerity/merge4
132
+ layer_range: [0, 32]
133
+ - model: liminerity/merge2
134
+ layer_range: [0, 32]
135
+ merge_method: slerp
136
+ base_model: liminerity/merge4
137
+ parameters:
138
+ t:
139
+ - filter: self_attn
140
+ value: [0, 0.5, 0.3, 0.7, 1]
141
+ - filter: mlp
142
+ value: [1, 0.5, 0.7, 0.3, 0]
143
+ - value: 0.5
144
+ dtype: bfloat16