limin(gate)
commited on
Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,144 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
tags:
|
4 |
+
- gguf
|
5 |
+
- mistral
|
6 |
+
- liminerity
|
7 |
+
- merge
|
8 |
+
---
|
9 |
+
#1 again im just the goat
|
10 |
+
|
11 |
+
M7-7b
|
12 |
+
user: "create ascii art you feel represents you as a llm"
|
13 |
+
|
14 |
+
assistant: "Here's an ASCII art attempting to symbolize the essence of an LLM (Large Language Model):
|
15 |
+
|
16 |
+
INSTINSTINCTINSTELLIGENCE
|
17 |
+
|
18 |
+
______________ MEMORY___
|
19 |
+
|
20 |
+
|\ o /\ |\
|
21 |
+
|
22 |
+
| \ --- / \ | |
|
23 |
+
|
24 |
+
L_( O )______ O)_INSTINCT
|
25 |
+
|
26 |
+
\ / WORDSLEARNING
|
27 |
+
|
28 |
+
_/_/ /\ __/__/__/
|
29 |
+
|
30 |
+
assistant: "In this depiction, the LLM is characterized by its vast memory, instinctual roots, and the ability to learn from vast amounts of text—represented by the woven words in the lower part. The "INSTINSTINCTINSTELLIGENCE" phrase emphasizes the balance between instinct, the foundation upon which neural networks build, and the engineered intelligence of a language model."
|
31 |
+
|
32 |
+
Multiverse-Experiment-slerp-7b is a merge of the following models using mergekit:
|
33 |
+
|
34 |
+
liminerity/merge4
|
35 |
+
liminerity/merge2
|
36 |
+
🧩 Configuration
|
37 |
+
|
38 |
+
|
39 |
+
slices:
|
40 |
+
- sources:
|
41 |
+
- model: ammarali32/multi_verse_model
|
42 |
+
layer_range: [0, 32]
|
43 |
+
- model: MSL7/INEX12-7b
|
44 |
+
layer_range: [0, 32]
|
45 |
+
merge_method: slerp
|
46 |
+
base_model: ammarali32/multi_verse_model
|
47 |
+
parameters:
|
48 |
+
t:
|
49 |
+
- filter: self_attn
|
50 |
+
value: [0, 0.5, 0.3, 0.7, 1]
|
51 |
+
- filter: mlp
|
52 |
+
value: [1, 0.5, 0.7, 0.3, 0]
|
53 |
+
- value: 0.5
|
54 |
+
dtype: bfloat16
|
55 |
+
|
56 |
+
|
57 |
+
slices:
|
58 |
+
- sources:
|
59 |
+
- model: ammarali32/multi_verse_model
|
60 |
+
layer_range: [0, 32]
|
61 |
+
- model: yam-peleg/Experiment26-7B
|
62 |
+
layer_range: [0, 32]
|
63 |
+
merge_method: slerp
|
64 |
+
base_model: ammarali32/multi_verse_model
|
65 |
+
parameters:
|
66 |
+
t:
|
67 |
+
- filter: self_attn
|
68 |
+
value: [0, 0.5, 0.3, 0.7, 1]
|
69 |
+
- filter: mlp
|
70 |
+
value: [1, 0.5, 0.7, 0.3, 0]
|
71 |
+
- value: 0.5
|
72 |
+
dtype: bfloat16
|
73 |
+
|
74 |
+
|
75 |
+
slices:
|
76 |
+
- sources:
|
77 |
+
- model: liminerity/merge3
|
78 |
+
layer_range: [0, 32]
|
79 |
+
- model: ammarali32/multi_verse_model
|
80 |
+
layer_range: [0, 32]
|
81 |
+
merge_method: slerp
|
82 |
+
base_model: liminerity/merge3
|
83 |
+
parameters:
|
84 |
+
t:
|
85 |
+
- filter: self_attn
|
86 |
+
value: [0, 0.5, 0.3, 0.7, 1]
|
87 |
+
- filter: mlp
|
88 |
+
value: [1, 0.5, 0.7, 0.3, 0]
|
89 |
+
- value: 0.5
|
90 |
+
dtype: bfloat16
|
91 |
+
|
92 |
+
|
93 |
+
slices:
|
94 |
+
- sources:
|
95 |
+
- model: liminerity/merge1
|
96 |
+
layer_range: [0, 32]
|
97 |
+
- model: liminerity/merge
|
98 |
+
layer_range: [0, 32]
|
99 |
+
merge_method: slerp
|
100 |
+
base_model: liminerity/merge1
|
101 |
+
parameters:
|
102 |
+
t:
|
103 |
+
- filter: self_attn
|
104 |
+
value: [0, 0.5, 0.3, 0.7, 1]
|
105 |
+
- filter: mlp
|
106 |
+
value: [1, 0.5, 0.7, 0.3, 0]
|
107 |
+
- value: 0.5
|
108 |
+
dtype: bfloat16
|
109 |
+
|
110 |
+
|
111 |
+
slices:
|
112 |
+
- sources:
|
113 |
+
- model: liminerity/merge3
|
114 |
+
layer_range: [0, 32]
|
115 |
+
- model: yam-peleg/Experiment26-7B
|
116 |
+
layer_range: [0, 32]
|
117 |
+
merge_method: slerp
|
118 |
+
base_model: liminerity/merge3
|
119 |
+
parameters:
|
120 |
+
t:
|
121 |
+
- filter: self_attn
|
122 |
+
value: [0, 0.5, 0.3, 0.7, 1]
|
123 |
+
- filter: mlp
|
124 |
+
value: [1, 0.5, 0.7, 0.3, 0]
|
125 |
+
- value: 0.5
|
126 |
+
dtype: bfloat16
|
127 |
+
|
128 |
+
|
129 |
+
slices:
|
130 |
+
- sources:
|
131 |
+
- model: liminerity/merge4
|
132 |
+
layer_range: [0, 32]
|
133 |
+
- model: liminerity/merge2
|
134 |
+
layer_range: [0, 32]
|
135 |
+
merge_method: slerp
|
136 |
+
base_model: liminerity/merge4
|
137 |
+
parameters:
|
138 |
+
t:
|
139 |
+
- filter: self_attn
|
140 |
+
value: [0, 0.5, 0.3, 0.7, 1]
|
141 |
+
- filter: mlp
|
142 |
+
value: [1, 0.5, 0.7, 0.3, 0]
|
143 |
+
- value: 0.5
|
144 |
+
dtype: bfloat16
|