ballcos commited on
Commit
dc96e7e
·
1 Parent(s): 4c88ee3
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ model.safetensors filter=lfs diff=lfs merge=lfs -text
config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "out/pretrained_300M_515000",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 1,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 1024,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 5632,
14
+ "max_position_embeddings": 2048,
15
+ "model_type": "llama",
16
+ "num_attention_heads": 16,
17
+ "num_hidden_layers": 12,
18
+ "num_key_value_heads": 4,
19
+ "pretraining_tp": 1,
20
+ "rms_norm_eps": 1e-05,
21
+ "rope_scaling": null,
22
+ "rope_theta": 10000.0,
23
+ "tie_word_embeddings": false,
24
+ "torch_dtype": "float32",
25
+ "transformers_version": "4.39.1",
26
+ "use_cache": true,
27
+ "vocab_size": 32000
28
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.39.1"
6
+ }
hyperparameters.yaml ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ checkpoint_dir: keeeeenw/MicroLlama
2
+ out_dir: out/escho
3
+ devices: 1
4
+ num_nodes: 1
5
+ lora_r: 8
6
+ lora_alpha: 16
7
+ lora_dropout: 0.05
8
+ lora_query: true
9
+ lora_key: false
10
+ lora_value: true
11
+ lora_projection: false
12
+ lora_mlp: false
13
+ lora_head: false
14
+ data:
15
+ class_path: litgpt.data.JSON
16
+ init_args:
17
+ json_path: data/ko_alpaca_data.json
18
+ mask_prompt: false
19
+ val_split_fraction: 0.5
20
+ prompt_style: alpaca
21
+ ignore_index: -100
22
+ seed: 42
23
+ num_workers: 4
24
+ train:
25
+ save_interval: 1000
26
+ log_interval: 1
27
+ global_batch_size: 16
28
+ micro_batch_size: 1
29
+ lr_warmup_steps: 100
30
+ epochs: 5
31
+ min_lr: 6.0e-05
32
+ eval:
33
+ interval: 100
34
+ max_new_tokens: 100
35
+ max_iters: 100
36
+ initial_validation: false
37
+ final_validation: true
38
+ evaluate_example: first
39
+ optimizer: AdamW
40
+ logger_name: csv
41
+ seed: 1337
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b78e63cb2edadfde3bc32f0b1a9a834d3a78560389360e4b2a5cd3775caed64
3
+ size 609286312
model_config.yaml ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ attention_logit_softcapping: null
2
+ attention_scores_scalar: null
3
+ attn_bias: false
4
+ bias: false
5
+ block_size: 2048
6
+ final_logit_softcapping: null
7
+ gelu_approximate: none
8
+ head_size: 64
9
+ hf_config:
10
+ name: MicroLlama
11
+ org: keeeeenw
12
+ intermediate_size: 5632
13
+ lm_head_bias: false
14
+ mlp_class_name: LLaMAMLP
15
+ n_embd: 1024
16
+ n_expert: 0
17
+ n_expert_per_token: 0
18
+ n_head: 16
19
+ n_layer: 12
20
+ n_query_groups: 4
21
+ name: micro-llama-300M
22
+ norm_class_name: RMSNorm
23
+ norm_eps: 1.0e-05
24
+ padded_vocab_size: 32000
25
+ padding_multiple: 64
26
+ parallel_residual: false
27
+ post_attention_norm: false
28
+ post_mlp_norm: false
29
+ rope_adjustments: null
30
+ rope_base: 10000
31
+ rope_condense_ratio: 1
32
+ rotary_percentage: 1.0
33
+ scale_embeddings: false
34
+ shared_attention_norm: false
35
+ sliding_window_layer_placing: null
36
+ sliding_window_size: null
37
+ vocab_size: 32000
prompt_style.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ class_path: litgpt.prompts.Alpaca
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": true,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": true,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": true,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": true,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ }
30
+ },
31
+ "bos_token": "<s>",
32
+ "clean_up_tokenization_spaces": false,
33
+ "eos_token": "</s>",
34
+ "legacy": true,
35
+ "model_max_length": 2048,
36
+ "pad_token": null,
37
+ "sp_model_kwargs": {},
38
+ "spaces_between_special_tokens": false,
39
+ "tokenizer_class": "LlamaTokenizer",
40
+ "unk_token": "<unk>",
41
+ "use_default_system_prompt": false
42
+ }