obahamonde commited on
Commit
10e5d05
·
verified ·
1 Parent(s): 1abc3d7
README.md CHANGED
@@ -32,14 +32,14 @@ More information needed
32
  ### Training hyperparameters
33
 
34
  The following hyperparameters were used during training:
35
- - learning_rate: 1e-06
36
  - train_batch_size: 1
37
  - eval_batch_size: 8
38
  - seed: 42
39
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
  - lr_scheduler_type: constant
41
  - lr_scheduler_warmup_ratio: 0.03
42
- - num_epochs: 3
43
 
44
  ### Training results
45
 
 
32
  ### Training hyperparameters
33
 
34
  The following hyperparameters were used during training:
35
+ - learning_rate: 0.0002
36
  - train_batch_size: 1
37
  - eval_batch_size: 8
38
  - seed: 42
39
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
  - lr_scheduler_type: constant
41
  - lr_scheduler_warmup_ratio: 0.03
42
+ - num_epochs: 5
43
 
44
  ### Training results
45
 
adapter_config.json CHANGED
@@ -19,13 +19,13 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "o_proj",
23
- "v_proj",
24
  "gate_proj",
25
  "up_proj",
26
- "down_proj",
27
  "q_proj",
28
- "k_proj"
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "use_rslora": false
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "k_proj",
23
+ "down_proj",
24
  "gate_proj",
25
  "up_proj",
26
+ "v_proj",
27
  "q_proj",
28
+ "o_proj"
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "use_rslora": false
generation_config.json CHANGED
@@ -1,7 +1,6 @@
1
  {
2
  "_from_model_config": true,
3
- "bos_token_id": 32001,
4
- "eos_token_id": 32000,
5
- "pad_token_id": 32000,
6
  "transformers_version": "4.36.2"
7
  }
 
1
  {
2
  "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
 
5
  "transformers_version": "4.36.2"
6
  }
runs/Mar20_12-37-02_lab/events.out.tfevents.1710938223.lab.32302.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8499d8c373c99b6c8cccac7415ce14819b00c4e0ebabff82f42d1360e3bbc83
3
+ size 4733
runs/Mar20_12-38-11_lab/events.out.tfevents.1710938291.lab.32586.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:392edf39fc15b2160595c089bff4af3806647d1e9c2ede3a42091ce3ac6e3b68
3
+ size 5543
special_tokens_map.json CHANGED
@@ -1,23 +1,25 @@
1
  {
2
- "additional_special_tokens": [
3
- {
4
- "content": "<|im_end|>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false
9
- },
10
- {
11
- "content": "<|im_start|>",
12
- "lstrip": false,
13
- "normalized": false,
14
- "rstrip": false,
15
- "single_word": false
16
- }
17
- ],
18
- "bos_token": "<|im_start|>",
19
- "eos_token": "<|im_end|>",
20
- "pad_token": "<|im_end|>",
 
 
21
  "unk_token": {
22
  "content": "<unk>",
23
  "lstrip": true,
 
1
  {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": true,
5
+ "normalized": false,
6
+ "rstrip": true,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|im_end|>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
  "unk_token": {
24
  "content": "<unk>",
25
  "lstrip": true,
tokenizer.json CHANGED
@@ -50,8 +50,8 @@
50
  "id": 32001,
51
  "content": "<|im_start|>",
52
  "single_word": false,
53
- "lstrip": false,
54
- "rstrip": false,
55
  "normalized": false,
56
  "special": true
57
  }
 
50
  "id": 32001,
51
  "content": "<|im_start|>",
52
  "single_word": false,
53
+ "lstrip": true,
54
+ "rstrip": true,
55
  "normalized": false,
56
  "special": true
57
  }
tokenizer_config.json CHANGED
@@ -36,21 +36,18 @@
36
  },
37
  "32001": {
38
  "content": "<|im_start|>",
39
- "lstrip": false,
40
  "normalized": false,
41
- "rstrip": false,
42
  "single_word": false,
43
  "special": true
44
  }
45
  },
46
- "additional_special_tokens": [
47
- "<|im_end|>",
48
- "<|im_start|>"
49
- ],
50
- "bos_token": "<|im_start|>",
51
- "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
52
  "clean_up_tokenization_spaces": false,
53
- "eos_token": "<|im_end|>",
54
  "legacy": true,
55
  "max_length": null,
56
  "model_max_length": 1000000000000000019884624838656,
 
36
  },
37
  "32001": {
38
  "content": "<|im_start|>",
39
+ "lstrip": true,
40
  "normalized": false,
41
+ "rstrip": true,
42
  "single_word": false,
43
  "special": true
44
  }
45
  },
46
+ "additional_special_tokens": [],
47
+ "bos_token": "<s>",
48
+ "chat_template": "{{ bos_token }}{% for message in messages %}<|im_start|> {{ message.role }}\n{{ message.content }}{% if message.role == 'assistant' %}{{ eos_token }}{% endif %}",
 
 
 
49
  "clean_up_tokenization_spaces": false,
50
+ "eos_token": "</s>",
51
  "legacy": true,
52
  "max_length": null,
53
  "model_max_length": 1000000000000000019884624838656,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:990aae14fbee58b82bc9875a5070f80d8226c4bdaed43fd024ecca7d1820e735
3
  size 4664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad6ea7e984b94d60cd1ce82807509a4605d551c369dae971949ade7690f31b5b
3
  size 4664