bartowski commited on
Commit
891f9cd
·
verified ·
1 Parent(s): 259fc46

measurement.json

Browse files
Files changed (2) hide show
  1. README.md +125 -0
  2. measurement.json +0 -0
README.md ADDED
@@ -0,0 +1,125 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: other
5
+ tags:
6
+ - axolotl
7
+ - instruct
8
+ - finetune
9
+ - chatml
10
+ - gpt4
11
+ - synthetic data
12
+ - science
13
+ - physics
14
+ - chemistry
15
+ - biology
16
+ - math
17
+ - qwen
18
+ - qwen2
19
+ base_model: Qwen/Qwen2-7B
20
+ datasets:
21
+ - allenai/ai2_arc
22
+ - camel-ai/physics
23
+ - camel-ai/chemistry
24
+ - camel-ai/biology
25
+ - camel-ai/math
26
+ - metaeval/reclor
27
+ - openbookqa
28
+ - mandyyyyii/scibench
29
+ - derek-thomas/ScienceQA
30
+ - TIGER-Lab/ScienceEval
31
+ - jondurbin/airoboros-3.2
32
+ - LDJnr/Capybara
33
+ - Cot-Alpaca-GPT4-From-OpenHermes-2.5
34
+ - STEM-AI-mtl/Electrical-engineering
35
+ - knowrohit07/saraswati-stem
36
+ - sablo/oasst2_curated
37
+ - lmsys/lmsys-chat-1m
38
+ - TIGER-Lab/MathInstruct
39
+ - bigbio/med_qa
40
+ - meta-math/MetaMathQA-40K
41
+ - openbookqa
42
+ - piqa
43
+ - metaeval/reclor
44
+ - derek-thomas/ScienceQA
45
+ - scibench
46
+ - sciq
47
+ - Open-Orca/SlimOrca
48
+ - migtissera/Synthia-v1.3
49
+ - TIGER-Lab/ScienceEval
50
+ - allenai/WildChat
51
+ - microsoft/orca-math-word-problems-200k
52
+ - openchat/openchat_sharegpt4_dataset
53
+ - teknium/GPTeacher-General-Instruct
54
+ - m-a-p/CodeFeedback-Filtered-Instruction
55
+ - totally-not-an-llm/EverythingLM-data-V3
56
+ - HuggingFaceH4/no_robots
57
+ - OpenAssistant/oasst_top1_2023-08-25
58
+ - WizardLM/WizardLM_evol_instruct_70k
59
+ - abacusai/SystemChat-1.1
60
+ - H-D-T/Buzz-V1.2
61
+ quantized_by: bartowski
62
+ pipeline_tag: text-generation
63
+ ---
64
+
65
+ ## Exllama v2 Quantizations of Einstein-v7-Qwen2-7B
66
+
67
+ Using <a href="https://github.com/turboderp/exllamav2/releases/tag/v0.1.6">turboderp's ExLlamaV2 v0.1.6</a> for quantization.
68
+
69
+ <b>The "main" branch only contains the measurement.json, download one of the other branches for the model (see below)</b>
70
+
71
+ Each branch contains an individual bits per weight, with the main one containing only the meaurement.json for further conversions.
72
+
73
+ Original model: https://huggingface.co/Weyaxi/Einstein-v7-Qwen2-7B
74
+
75
+ ## Prompt format
76
+
77
+ ```
78
+ <|im_start|>system
79
+ {system_prompt}<|im_end|>
80
+ <|im_start|>user
81
+ {prompt}<|im_end|>
82
+ <|im_start|>assistant
83
+
84
+ ```
85
+
86
+ ## Available sizes
87
+
88
+
89
+ | Branch | Bits | lm_head bits | VRAM (4k) | VRAM (16k) | VRAM (32k) | Description |
90
+ | ----- | ---- | ------- | ------ | ------ | ------ | ------------ |
91
+ | [8_0](https://huggingface.co/bartowski/Einstein-v7-Qwen2-7B-exl2/tree/8_0) | 8.0 | 8.0 | 8.4 GB | 9.8 GB | 11.8 GB | Maximum quality that ExLlamaV2 can produce, near unquantized performance. |
92
+ | [6_5](https://huggingface.co/bartowski/Einstein-v7-Qwen2-7B-exl2/tree/6_5) | 6.5 | 8.0 | 7.2 GB | 8.6 GB | 10.6 GB | Very similar to 8.0, good tradeoff of size vs performance, **recommended**. |
93
+ | [5_0](https://huggingface.co/bartowski/Einstein-v7-Qwen2-7B-exl2/tree/5_0) | 5.0 | 6.0 | 6.0 GB | 7.4 GB | 9.4 GB | Slightly lower quality vs 6.5, but usable on 8GB cards. |
94
+ | [4_25](https://huggingface.co/bartowski/Einstein-v7-Qwen2-7B-exl2/tree/4_25) | 4.25 | 6.0 | 5.3 GB | 6.7 GB | 8.7 GB | GPTQ equivalent bits per weight, slightly higher quality. |
95
+ | [3_5](https://huggingface.co/bartowski/Einstein-v7-Qwen2-7B-exl2/tree/3_5) | 3.5 | 6.0 | 4.7 GB | 6.1 GB | 8.1 GB | Lower quality, only use if you have to. |
96
+
97
+ ## Download instructions
98
+
99
+ With git:
100
+
101
+ ```shell
102
+ git clone --single-branch --branch 6_5 https://huggingface.co/bartowski/Einstein-v7-Qwen2-7B-exl2 Einstein-v7-Qwen2-7B-exl2-6_5
103
+ ```
104
+
105
+ With huggingface hub (credit to TheBloke for instructions):
106
+
107
+ ```shell
108
+ pip3 install huggingface-hub
109
+ ```
110
+
111
+ To download a specific branch, use the `--revision` parameter. For example, to download the 6.5 bpw branch:
112
+
113
+ Linux:
114
+
115
+ ```shell
116
+ huggingface-cli download bartowski/Einstein-v7-Qwen2-7B-exl2 --revision 6_5 --local-dir Einstein-v7-Qwen2-7B-exl2-6_5
117
+ ```
118
+
119
+ Windows (which apparently doesn't like _ in folders sometimes?):
120
+
121
+ ```shell
122
+ huggingface-cli download bartowski/Einstein-v7-Qwen2-7B-exl2 --revision 6_5 --local-dir Einstein-v7-Qwen2-7B-exl2-6.5
123
+ ```
124
+
125
+ Want to support my work? Visit my ko-fi page here: https://ko-fi.com/bartowski
measurement.json ADDED
The diff for this file is too large to render. See raw diff