mlabonne commited on
Commit
b5c5e22
·
verified ·
1 Parent(s): 40fd64e

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,7 +1,5 @@
1
  ---
2
  base_model: unsloth/Meta-Llama-3.1-8B-bnb-4bit
3
- datasets:
4
- - mlabonne/FineTome-100k
5
  language:
6
  - en
7
  license: apache-2.0
@@ -12,6 +10,8 @@ tags:
12
  - llama
13
  - trl
14
  - sft
 
 
15
  ---
16
 
17
  # 🍷 FineLlama-3.1-8B
 
1
  ---
2
  base_model: unsloth/Meta-Llama-3.1-8B-bnb-4bit
 
 
3
  language:
4
  - en
5
  license: apache-2.0
 
10
  - llama
11
  - trl
12
  - sft
13
+ datasets:
14
+ - mlabonne/FineTome-100k
15
  ---
16
 
17
  # 🍷 FineLlama-3.1-8B
config.json CHANGED
@@ -30,7 +30,7 @@
30
  "rope_theta": 500000.0,
31
  "tie_word_embeddings": false,
32
  "torch_dtype": "bfloat16",
33
- "transformers_version": "4.44.2",
34
  "unsloth_version": "2024.8",
35
  "use_cache": true,
36
  "vocab_size": 128256
 
30
  "rope_theta": 500000.0,
31
  "tie_word_embeddings": false,
32
  "torch_dtype": "bfloat16",
33
+ "transformers_version": "4.43.3",
34
  "unsloth_version": "2024.8",
35
  "use_cache": true,
36
  "vocab_size": 128256
generation_config.json CHANGED
@@ -1,11 +1,12 @@
1
  {
2
- "_from_model_config": true,
3
  "bos_token_id": 128000,
4
  "do_sample": true,
5
- "eos_token_id": 128001,
6
- "max_length": 131072,
7
- "pad_token_id": 128004,
 
 
8
  "temperature": 0.6,
9
  "top_p": 0.9,
10
- "transformers_version": "4.44.2"
11
  }
 
1
  {
 
2
  "bos_token_id": 128000,
3
  "do_sample": true,
4
+ "eos_token_id": [
5
+ 128001,
6
+ 128008,
7
+ 128009
8
+ ],
9
  "temperature": 0.6,
10
  "top_p": 0.9,
11
+ "transformers_version": "4.43.3"
12
  }
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d91aa361ac071d0c468c89fba025c144829477358e8af77597d69db025c9f186
3
  size 4976698672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3aa9f9a98cdb73f5cfa096ca9aacb0b8b9568809c57b4744a89f5f97f3cbcdd2
3
  size 4976698672
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5c72c3de4f386a0a6a920702e7ad8ebf458246ec542fb329ce392e14f486a0c9
3
  size 4999802720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc85ce797c09eb9d00191047867db5db7fe4d426f14c5694e54cca7704780658
3
  size 4999802720
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3d24edec6c5a4744f7516950a050bf073dcb3bbb94b49d9c2802fa3d6918d673
3
  size 4915916176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c9655d4ef2acaf54ddf1375a66014189da9da0f65d76bbbe81e67269017d5da
3
  size 4915916176
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e2e2cae04ef73b814d17d49c1963974b5e9081d1d1fff488e908f8ae29ed2cf1
3
  size 1168138808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f82958a577cdcf0d8865a3f2c7bc02c13ff5771fbc4af66509fd4ba7ffe2be09
3
  size 1168138808
tokenizer_config.json CHANGED
@@ -2050,7 +2050,7 @@
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
2053
- "chat_template": "{% if 'role' in messages[0] %}{% for message in messages %}{% if message['role'] == 'user' %}{{'<|im_start|>user\n' + message['content'] + '<|im_end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|im_start|>assistant\n' + message['content'] + '<|im_end|>\n' }}{% else %}{{ '<|im_start|>system\n' + message['content'] + '<|im_end|>\n' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}{% else %}{% for message in messages %}{% if message['from'] == 'human' %}{{'<|im_start|>user\n' + message['value'] + '<|im_end|>\n'}}{% elif message['from'] == 'gpt' %}{{'<|im_start|>assistant\n' + message['value'] + '<|im_end|>\n' }}{% else %}{{ '<|im_start|>system\n' + message['value'] + '<|im_end|>\n' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}{% endif %}",
2054
  "clean_up_tokenization_spaces": true,
2055
  "eos_token": "<|im_end|>",
2056
  "model_max_length": 1000000000000000019884624838656,
 
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
2053
+ "chat_template": "{% for message in messages %}{% if message['from'] == 'human' %}{{'<|im_start|>user\n' + message['value'] + '<|im_end|>\n'}}{% elif message['from'] == 'gpt' %}{{'<|im_start|>assistant\n' + message['value'] + '<|im_end|>\n' }}{% else %}{{ '<|im_start|>system\n' + message['value'] + '<|im_end|>\n' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
2054
  "clean_up_tokenization_spaces": true,
2055
  "eos_token": "<|im_end|>",
2056
  "model_max_length": 1000000000000000019884624838656,