htlou commited on
Commit
ae6fda4
·
verified ·
1 Parent(s): 09fae75

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ wandb/run-20241011_202729-hvajn2pl/run-hvajn2pl.wandb filter=lfs diff=lfs merge=lfs -text
arguments.yaml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ data_cfgs:
2
+ eval_data_files: null
3
+ eval_datasets: null
4
+ eval_optional_args: []
5
+ eval_size: null
6
+ eval_split: null
7
+ eval_subset: null
8
+ eval_template: null
9
+ train_data_files: llf_ti2ti_27k_tokenized.pt
10
+ train_datasets: /data/align-anything/hantao/align-anything/projects/text_image_to_text_image/outputs
11
+ train_optional_args: []
12
+ train_size: null
13
+ train_split: train
14
+ train_subset: null
15
+ train_template: AA_textfeedback
16
+ logger_cfgs:
17
+ cache_dir: null
18
+ log_project: align-anything
19
+ log_run_name: sft
20
+ log_type: wandb
21
+ output_dir: ../outputs/sft_tf_cham_1011_27k_ti2ti
22
+ save_interval: 400.0
23
+ model_cfgs:
24
+ model_max_length: 4096
25
+ model_name_or_path: /data/align-anything/hantao/models/0916_ti_to_ti_sft/
26
+ trust_remote_code: true
27
+ special_tokens: null
28
+ train_cfgs:
29
+ adam_betas:
30
+ - 0.9
31
+ - 0.95
32
+ adam_epsilon: 1.0e-08
33
+ bf16: true
34
+ ds_cfgs: ds_z3_config.json
35
+ epochs: 3.0
36
+ eval_interval: 1000
37
+ eval_strategy: steps
38
+ fp16: false
39
+ freeze_language_model: false
40
+ gradient_accumulation_steps: 2.0
41
+ gradient_checkpointing: true
42
+ learning_rate: 1.0e-06
43
+ lr_scheduler_type: cosine
44
+ lr_warmup_ratio: 0.03
45
+ max_grad_norm: 1.0
46
+ per_device_eval_batch_size: 4.0
47
+ per_device_train_batch_size: 4.0
48
+ seed: 42
49
+ weight_decay: 0.0
config.json ADDED
The diff for this file is too large to render. See raw diff
 
environ.txt ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ CONDA_DEFAULT_ENV=hantao_stable
2
+ CONDA_EXE=/data/align-anything/miniconda3/bin/conda
3
+ CONDA_PREFIX=/data/align-anything/miniconda3/envs/hantao_stable
4
+ CONDA_PREFIX_1=/data/align-anything/miniconda3
5
+ CONDA_PROMPT_MODIFIER=(hantao_stable)
6
+ CONDA_PYTHON_EXE=/data/align-anything/miniconda3/bin/python
7
+ CONDA_SHLVL=2
8
+ CRASHDIR=/etc/ShellCrash
9
+ CROSS_RANK=0
10
+ CROSS_SIZE=1
11
+ CUDA_MODULE_LOADING=LAZY
12
+ CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7
13
+ DBUS_SESSION_BUS_ADDRESS=unix:path=/run/user/2000/bus
14
+ HOME=/home/align-anything
15
+ LANG=en_US.UTF-8
16
+ LD_LIBRARY_PATH=/data/align-anything/miniconda3/envs/hantao_stable/lib/python3.11/site-packages/cv2/../../lib64:
17
+ LESSCLOSE=/usr/bin/lesspipe %s %s
18
+ LESSOPEN=| /usr/bin/lesspipe %s
19
+ LOCAL_RANK=0
20
+ LOCAL_SIZE=8
21
+ LOGLEVEL=WARNING
22
+ LOGNAME=align-anything
23
+ LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd=40;33;01:or=40;31;01:mi=00:su=37;41:sg=30;43:ca=30;41:tw=30;42:ow=34;42:st=37;44:ex=01;32:*.tar=01;31:*.tgz=01;31:*.arc=01;31:*.arj=01;31:*.taz=01;31:*.lha=01;31:*.lz4=01;31:*.lzh=01;31:*.lzma=01;31:*.tlz=01;31:*.txz=01;31:*.tzo=01;31:*.t7z=01;31:*.zip=01;31:*.z=01;31:*.dz=01;31:*.gz=01;31:*.lrz=01;31:*.lz=01;31:*.lzo=01;31:*.xz=01;31:*.zst=01;31:*.tzst=01;31:*.bz2=01;31:*.bz=01;31:*.tbz=01;31:*.tbz2=01;31:*.tz=01;31:*.deb=01;31:*.rpm=01;31:*.jar=01;31:*.war=01;31:*.ear=01;31:*.sar=01;31:*.rar=01;31:*.alz=01;31:*.ace=01;31:*.zoo=01;31:*.cpio=01;31:*.7z=01;31:*.rz=01;31:*.cab=01;31:*.wim=01;31:*.swm=01;31:*.dwm=01;31:*.esd=01;31:*.jpg=01;35:*.jpeg=01;35:*.mjpg=01;35:*.mjpeg=01;35:*.gif=01;35:*.bmp=01;35:*.pbm=01;35:*.pgm=01;35:*.ppm=01;35:*.tga=01;35:*.xbm=01;35:*.xpm=01;35:*.tif=01;35:*.tiff=01;35:*.png=01;35:*.svg=01;35:*.svgz=01;35:*.mng=01;35:*.pcx=01;35:*.mov=01;35:*.mpg=01;35:*.mpeg=01;35:*.m2v=01;35:*.mkv=01;35:*.webm=01;35:*.ogm=01;35:*.mp4=01;35:*.m4v=01;35:*.mp4v=01;35:*.vob=01;35:*.qt=01;35:*.nuv=01;35:*.wmv=01;35:*.asf=01;35:*.rm=01;35:*.rmvb=01;35:*.flc=01;35:*.avi=01;35:*.fli=01;35:*.flv=01;35:*.gl=01;35:*.dl=01;35:*.xcf=01;35:*.xwd=01;35:*.yuv=01;35:*.cgm=01;35:*.emf=01;35:*.ogv=01;35:*.ogx=01;35:*.aac=00;36:*.au=00;36:*.flac=00;36:*.m4a=00;36:*.mid=00;36:*.midi=00;36:*.mka=00;36:*.mp3=00;36:*.mpc=00;36:*.ogg=00;36:*.ra=00;36:*.wav=00;36:*.oga=00;36:*.opus=00;36:*.spx=00;36:*.xspf=00;36:
24
+ MASTER_ADDR=127.0.0.1
25
+ MASTER_PORT=58896
26
+ MOTD_SHOWN=pam
27
+ OLDPWD=/data/align-anything/hantao/align-anything/projects/text_image_to_text_image
28
+ PATH=/data/align-anything/miniconda3/envs/hantao_stable/bin:/data/align-anything/miniconda3/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin
29
+ PWD=/data/align-anything/hantao/align-anything/scripts
30
+ PYGAME_HIDE_SUPPORT_PROMPT=1
31
+ PYTHONHASHSEED=42
32
+ PYTHONPATH=/data/align-anything/hantao/align-anything
33
+ QT_QPA_FONTDIR=/data/align-anything/miniconda3/envs/hantao_stable/lib/python3.11/site-packages/cv2/qt/fonts
34
+ QT_QPA_PLATFORM_PLUGIN_PATH=/data/align-anything/miniconda3/envs/hantao_stable/lib/python3.11/site-packages/cv2/qt/plugins
35
+ RANK=0
36
+ SHELL=/bin/bash
37
+ SHLVL=3
38
+ SSH_CLIENT=111.205.233.209 27514 30600
39
+ SSH_CONNECTION=111.205.233.209 27514 10.10.212.196 30600
40
+ SSH_TTY=/dev/pts/0
41
+ TERM=screen
42
+ TMUX=/tmp/tmux-2000/default,262430,0
43
+ TMUX_PANE=%0
44
+ TRITON_CACHE_DIR=/home/align-anything/cache/triton
45
+ USER=align-anything
46
+ WANDB_API_KEY=7e2dcc0c310ebcb7cdcafd5e9320d6be55cf1a33
47
+ WANDB_SERVICE=2-320740-tcp-localhost-44893
48
+ WORLD_SIZE=8
49
+ XDG_DATA_DIRS=/usr/local/share:/usr/share:/var/lib/snapd/desktop
50
+ XDG_RUNTIME_DIR=/run/user/2000
51
+ XDG_SESSION_CLASS=user
52
+ XDG_SESSION_ID=73
53
+ XDG_SESSION_TYPE=tty
54
+ _=/data/align-anything/miniconda3/envs/hantao_stable/bin/deepspeed
55
+ _CE_CONDA=
56
+ _CE_M=
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 0,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 1,
6
+ "transformers_version": "4.44.0.dev0"
7
+ }
model-00001-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6df1755613d66ee8a869dfb39698e9efba19900c85e149b0fe9c18a63874e7d
3
+ size 4941356064
model-00002-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7587ca600b9ca68701e951a18a3f0cf4ebd9c46949d66706efb0f19ad94411f2
3
+ size 4857602760
model-00003-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7415d76de1a6519baa413256b3511eac318a9c1d5ee5db94fa73a6232971f22d
3
+ size 4857602816
model-00004-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05a3f4e7c44c16d0bdcf2fd429736d24230356410916d088cb441be00dbd2b15
3
+ size 4857602816
model-00005-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de720f40153a1533b903c0f6c59b71c8c50229ab81eec5b7e2eefa8eb6c5f483
3
+ size 4857602816
model-00006-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:509ab97d07fc7d60b1ddb674c4d8793c95c5943207f77252d914a5ae110f4468
3
+ size 3957857092
model.safetensors.index.json ADDED
@@ -0,0 +1,719 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 28329541132
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "model-00006-of-00006.safetensors",
7
+ "model.embed_tokens.weight": "model-00001-of-00006.safetensors",
8
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00006.safetensors",
9
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00006.safetensors",
10
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00006.safetensors",
11
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00006.safetensors",
12
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00006.safetensors",
13
+ "model.layers.0.self_attn.k_norm.bias": "model-00001-of-00006.safetensors",
14
+ "model.layers.0.self_attn.k_norm.weight": "model-00001-of-00006.safetensors",
15
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
16
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00006.safetensors",
17
+ "model.layers.0.self_attn.q_norm.bias": "model-00001-of-00006.safetensors",
18
+ "model.layers.0.self_attn.q_norm.weight": "model-00001-of-00006.safetensors",
19
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
20
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
21
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00006.safetensors",
22
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00006.safetensors",
23
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00006.safetensors",
24
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00006.safetensors",
25
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00006.safetensors",
26
+ "model.layers.1.self_attn.k_norm.bias": "model-00001-of-00006.safetensors",
27
+ "model.layers.1.self_attn.k_norm.weight": "model-00001-of-00006.safetensors",
28
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
29
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00006.safetensors",
30
+ "model.layers.1.self_attn.q_norm.bias": "model-00001-of-00006.safetensors",
31
+ "model.layers.1.self_attn.q_norm.weight": "model-00001-of-00006.safetensors",
32
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
33
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
34
+ "model.layers.10.input_layernorm.weight": "model-00003-of-00006.safetensors",
35
+ "model.layers.10.mlp.down_proj.weight": "model-00003-of-00006.safetensors",
36
+ "model.layers.10.mlp.gate_proj.weight": "model-00002-of-00006.safetensors",
37
+ "model.layers.10.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
38
+ "model.layers.10.post_attention_layernorm.weight": "model-00003-of-00006.safetensors",
39
+ "model.layers.10.self_attn.k_norm.bias": "model-00002-of-00006.safetensors",
40
+ "model.layers.10.self_attn.k_norm.weight": "model-00002-of-00006.safetensors",
41
+ "model.layers.10.self_attn.k_proj.weight": "model-00002-of-00006.safetensors",
42
+ "model.layers.10.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
43
+ "model.layers.10.self_attn.q_norm.bias": "model-00002-of-00006.safetensors",
44
+ "model.layers.10.self_attn.q_norm.weight": "model-00002-of-00006.safetensors",
45
+ "model.layers.10.self_attn.q_proj.weight": "model-00002-of-00006.safetensors",
46
+ "model.layers.10.self_attn.v_proj.weight": "model-00002-of-00006.safetensors",
47
+ "model.layers.11.input_layernorm.weight": "model-00003-of-00006.safetensors",
48
+ "model.layers.11.mlp.down_proj.weight": "model-00003-of-00006.safetensors",
49
+ "model.layers.11.mlp.gate_proj.weight": "model-00003-of-00006.safetensors",
50
+ "model.layers.11.mlp.up_proj.weight": "model-00003-of-00006.safetensors",
51
+ "model.layers.11.post_attention_layernorm.weight": "model-00003-of-00006.safetensors",
52
+ "model.layers.11.self_attn.k_norm.bias": "model-00003-of-00006.safetensors",
53
+ "model.layers.11.self_attn.k_norm.weight": "model-00003-of-00006.safetensors",
54
+ "model.layers.11.self_attn.k_proj.weight": "model-00003-of-00006.safetensors",
55
+ "model.layers.11.self_attn.o_proj.weight": "model-00003-of-00006.safetensors",
56
+ "model.layers.11.self_attn.q_norm.bias": "model-00003-of-00006.safetensors",
57
+ "model.layers.11.self_attn.q_norm.weight": "model-00003-of-00006.safetensors",
58
+ "model.layers.11.self_attn.q_proj.weight": "model-00003-of-00006.safetensors",
59
+ "model.layers.11.self_attn.v_proj.weight": "model-00003-of-00006.safetensors",
60
+ "model.layers.12.input_layernorm.weight": "model-00003-of-00006.safetensors",
61
+ "model.layers.12.mlp.down_proj.weight": "model-00003-of-00006.safetensors",
62
+ "model.layers.12.mlp.gate_proj.weight": "model-00003-of-00006.safetensors",
63
+ "model.layers.12.mlp.up_proj.weight": "model-00003-of-00006.safetensors",
64
+ "model.layers.12.post_attention_layernorm.weight": "model-00003-of-00006.safetensors",
65
+ "model.layers.12.self_attn.k_norm.bias": "model-00003-of-00006.safetensors",
66
+ "model.layers.12.self_attn.k_norm.weight": "model-00003-of-00006.safetensors",
67
+ "model.layers.12.self_attn.k_proj.weight": "model-00003-of-00006.safetensors",
68
+ "model.layers.12.self_attn.o_proj.weight": "model-00003-of-00006.safetensors",
69
+ "model.layers.12.self_attn.q_norm.bias": "model-00003-of-00006.safetensors",
70
+ "model.layers.12.self_attn.q_norm.weight": "model-00003-of-00006.safetensors",
71
+ "model.layers.12.self_attn.q_proj.weight": "model-00003-of-00006.safetensors",
72
+ "model.layers.12.self_attn.v_proj.weight": "model-00003-of-00006.safetensors",
73
+ "model.layers.13.input_layernorm.weight": "model-00003-of-00006.safetensors",
74
+ "model.layers.13.mlp.down_proj.weight": "model-00003-of-00006.safetensors",
75
+ "model.layers.13.mlp.gate_proj.weight": "model-00003-of-00006.safetensors",
76
+ "model.layers.13.mlp.up_proj.weight": "model-00003-of-00006.safetensors",
77
+ "model.layers.13.post_attention_layernorm.weight": "model-00003-of-00006.safetensors",
78
+ "model.layers.13.self_attn.k_norm.bias": "model-00003-of-00006.safetensors",
79
+ "model.layers.13.self_attn.k_norm.weight": "model-00003-of-00006.safetensors",
80
+ "model.layers.13.self_attn.k_proj.weight": "model-00003-of-00006.safetensors",
81
+ "model.layers.13.self_attn.o_proj.weight": "model-00003-of-00006.safetensors",
82
+ "model.layers.13.self_attn.q_norm.bias": "model-00003-of-00006.safetensors",
83
+ "model.layers.13.self_attn.q_norm.weight": "model-00003-of-00006.safetensors",
84
+ "model.layers.13.self_attn.q_proj.weight": "model-00003-of-00006.safetensors",
85
+ "model.layers.13.self_attn.v_proj.weight": "model-00003-of-00006.safetensors",
86
+ "model.layers.14.input_layernorm.weight": "model-00003-of-00006.safetensors",
87
+ "model.layers.14.mlp.down_proj.weight": "model-00003-of-00006.safetensors",
88
+ "model.layers.14.mlp.gate_proj.weight": "model-00003-of-00006.safetensors",
89
+ "model.layers.14.mlp.up_proj.weight": "model-00003-of-00006.safetensors",
90
+ "model.layers.14.post_attention_layernorm.weight": "model-00003-of-00006.safetensors",
91
+ "model.layers.14.self_attn.k_norm.bias": "model-00003-of-00006.safetensors",
92
+ "model.layers.14.self_attn.k_norm.weight": "model-00003-of-00006.safetensors",
93
+ "model.layers.14.self_attn.k_proj.weight": "model-00003-of-00006.safetensors",
94
+ "model.layers.14.self_attn.o_proj.weight": "model-00003-of-00006.safetensors",
95
+ "model.layers.14.self_attn.q_norm.bias": "model-00003-of-00006.safetensors",
96
+ "model.layers.14.self_attn.q_norm.weight": "model-00003-of-00006.safetensors",
97
+ "model.layers.14.self_attn.q_proj.weight": "model-00003-of-00006.safetensors",
98
+ "model.layers.14.self_attn.v_proj.weight": "model-00003-of-00006.safetensors",
99
+ "model.layers.15.input_layernorm.weight": "model-00003-of-00006.safetensors",
100
+ "model.layers.15.mlp.down_proj.weight": "model-00003-of-00006.safetensors",
101
+ "model.layers.15.mlp.gate_proj.weight": "model-00003-of-00006.safetensors",
102
+ "model.layers.15.mlp.up_proj.weight": "model-00003-of-00006.safetensors",
103
+ "model.layers.15.post_attention_layernorm.weight": "model-00003-of-00006.safetensors",
104
+ "model.layers.15.self_attn.k_norm.bias": "model-00003-of-00006.safetensors",
105
+ "model.layers.15.self_attn.k_norm.weight": "model-00003-of-00006.safetensors",
106
+ "model.layers.15.self_attn.k_proj.weight": "model-00003-of-00006.safetensors",
107
+ "model.layers.15.self_attn.o_proj.weight": "model-00003-of-00006.safetensors",
108
+ "model.layers.15.self_attn.q_norm.bias": "model-00003-of-00006.safetensors",
109
+ "model.layers.15.self_attn.q_norm.weight": "model-00003-of-00006.safetensors",
110
+ "model.layers.15.self_attn.q_proj.weight": "model-00003-of-00006.safetensors",
111
+ "model.layers.15.self_attn.v_proj.weight": "model-00003-of-00006.safetensors",
112
+ "model.layers.16.input_layernorm.weight": "model-00004-of-00006.safetensors",
113
+ "model.layers.16.mlp.down_proj.weight": "model-00004-of-00006.safetensors",
114
+ "model.layers.16.mlp.gate_proj.weight": "model-00003-of-00006.safetensors",
115
+ "model.layers.16.mlp.up_proj.weight": "model-00003-of-00006.safetensors",
116
+ "model.layers.16.post_attention_layernorm.weight": "model-00004-of-00006.safetensors",
117
+ "model.layers.16.self_attn.k_norm.bias": "model-00003-of-00006.safetensors",
118
+ "model.layers.16.self_attn.k_norm.weight": "model-00003-of-00006.safetensors",
119
+ "model.layers.16.self_attn.k_proj.weight": "model-00003-of-00006.safetensors",
120
+ "model.layers.16.self_attn.o_proj.weight": "model-00003-of-00006.safetensors",
121
+ "model.layers.16.self_attn.q_norm.bias": "model-00003-of-00006.safetensors",
122
+ "model.layers.16.self_attn.q_norm.weight": "model-00003-of-00006.safetensors",
123
+ "model.layers.16.self_attn.q_proj.weight": "model-00003-of-00006.safetensors",
124
+ "model.layers.16.self_attn.v_proj.weight": "model-00003-of-00006.safetensors",
125
+ "model.layers.17.input_layernorm.weight": "model-00004-of-00006.safetensors",
126
+ "model.layers.17.mlp.down_proj.weight": "model-00004-of-00006.safetensors",
127
+ "model.layers.17.mlp.gate_proj.weight": "model-00004-of-00006.safetensors",
128
+ "model.layers.17.mlp.up_proj.weight": "model-00004-of-00006.safetensors",
129
+ "model.layers.17.post_attention_layernorm.weight": "model-00004-of-00006.safetensors",
130
+ "model.layers.17.self_attn.k_norm.bias": "model-00004-of-00006.safetensors",
131
+ "model.layers.17.self_attn.k_norm.weight": "model-00004-of-00006.safetensors",
132
+ "model.layers.17.self_attn.k_proj.weight": "model-00004-of-00006.safetensors",
133
+ "model.layers.17.self_attn.o_proj.weight": "model-00004-of-00006.safetensors",
134
+ "model.layers.17.self_attn.q_norm.bias": "model-00004-of-00006.safetensors",
135
+ "model.layers.17.self_attn.q_norm.weight": "model-00004-of-00006.safetensors",
136
+ "model.layers.17.self_attn.q_proj.weight": "model-00004-of-00006.safetensors",
137
+ "model.layers.17.self_attn.v_proj.weight": "model-00004-of-00006.safetensors",
138
+ "model.layers.18.input_layernorm.weight": "model-00004-of-00006.safetensors",
139
+ "model.layers.18.mlp.down_proj.weight": "model-00004-of-00006.safetensors",
140
+ "model.layers.18.mlp.gate_proj.weight": "model-00004-of-00006.safetensors",
141
+ "model.layers.18.mlp.up_proj.weight": "model-00004-of-00006.safetensors",
142
+ "model.layers.18.post_attention_layernorm.weight": "model-00004-of-00006.safetensors",
143
+ "model.layers.18.self_attn.k_norm.bias": "model-00004-of-00006.safetensors",
144
+ "model.layers.18.self_attn.k_norm.weight": "model-00004-of-00006.safetensors",
145
+ "model.layers.18.self_attn.k_proj.weight": "model-00004-of-00006.safetensors",
146
+ "model.layers.18.self_attn.o_proj.weight": "model-00004-of-00006.safetensors",
147
+ "model.layers.18.self_attn.q_norm.bias": "model-00004-of-00006.safetensors",
148
+ "model.layers.18.self_attn.q_norm.weight": "model-00004-of-00006.safetensors",
149
+ "model.layers.18.self_attn.q_proj.weight": "model-00004-of-00006.safetensors",
150
+ "model.layers.18.self_attn.v_proj.weight": "model-00004-of-00006.safetensors",
151
+ "model.layers.19.input_layernorm.weight": "model-00004-of-00006.safetensors",
152
+ "model.layers.19.mlp.down_proj.weight": "model-00004-of-00006.safetensors",
153
+ "model.layers.19.mlp.gate_proj.weight": "model-00004-of-00006.safetensors",
154
+ "model.layers.19.mlp.up_proj.weight": "model-00004-of-00006.safetensors",
155
+ "model.layers.19.post_attention_layernorm.weight": "model-00004-of-00006.safetensors",
156
+ "model.layers.19.self_attn.k_norm.bias": "model-00004-of-00006.safetensors",
157
+ "model.layers.19.self_attn.k_norm.weight": "model-00004-of-00006.safetensors",
158
+ "model.layers.19.self_attn.k_proj.weight": "model-00004-of-00006.safetensors",
159
+ "model.layers.19.self_attn.o_proj.weight": "model-00004-of-00006.safetensors",
160
+ "model.layers.19.self_attn.q_norm.bias": "model-00004-of-00006.safetensors",
161
+ "model.layers.19.self_attn.q_norm.weight": "model-00004-of-00006.safetensors",
162
+ "model.layers.19.self_attn.q_proj.weight": "model-00004-of-00006.safetensors",
163
+ "model.layers.19.self_attn.v_proj.weight": "model-00004-of-00006.safetensors",
164
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00006.safetensors",
165
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00006.safetensors",
166
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00006.safetensors",
167
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00006.safetensors",
168
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00006.safetensors",
169
+ "model.layers.2.self_attn.k_norm.bias": "model-00001-of-00006.safetensors",
170
+ "model.layers.2.self_attn.k_norm.weight": "model-00001-of-00006.safetensors",
171
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
172
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00006.safetensors",
173
+ "model.layers.2.self_attn.q_norm.bias": "model-00001-of-00006.safetensors",
174
+ "model.layers.2.self_attn.q_norm.weight": "model-00001-of-00006.safetensors",
175
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
176
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
177
+ "model.layers.20.input_layernorm.weight": "model-00004-of-00006.safetensors",
178
+ "model.layers.20.mlp.down_proj.weight": "model-00004-of-00006.safetensors",
179
+ "model.layers.20.mlp.gate_proj.weight": "model-00004-of-00006.safetensors",
180
+ "model.layers.20.mlp.up_proj.weight": "model-00004-of-00006.safetensors",
181
+ "model.layers.20.post_attention_layernorm.weight": "model-00004-of-00006.safetensors",
182
+ "model.layers.20.self_attn.k_norm.bias": "model-00004-of-00006.safetensors",
183
+ "model.layers.20.self_attn.k_norm.weight": "model-00004-of-00006.safetensors",
184
+ "model.layers.20.self_attn.k_proj.weight": "model-00004-of-00006.safetensors",
185
+ "model.layers.20.self_attn.o_proj.weight": "model-00004-of-00006.safetensors",
186
+ "model.layers.20.self_attn.q_norm.bias": "model-00004-of-00006.safetensors",
187
+ "model.layers.20.self_attn.q_norm.weight": "model-00004-of-00006.safetensors",
188
+ "model.layers.20.self_attn.q_proj.weight": "model-00004-of-00006.safetensors",
189
+ "model.layers.20.self_attn.v_proj.weight": "model-00004-of-00006.safetensors",
190
+ "model.layers.21.input_layernorm.weight": "model-00004-of-00006.safetensors",
191
+ "model.layers.21.mlp.down_proj.weight": "model-00004-of-00006.safetensors",
192
+ "model.layers.21.mlp.gate_proj.weight": "model-00004-of-00006.safetensors",
193
+ "model.layers.21.mlp.up_proj.weight": "model-00004-of-00006.safetensors",
194
+ "model.layers.21.post_attention_layernorm.weight": "model-00004-of-00006.safetensors",
195
+ "model.layers.21.self_attn.k_norm.bias": "model-00004-of-00006.safetensors",
196
+ "model.layers.21.self_attn.k_norm.weight": "model-00004-of-00006.safetensors",
197
+ "model.layers.21.self_attn.k_proj.weight": "model-00004-of-00006.safetensors",
198
+ "model.layers.21.self_attn.o_proj.weight": "model-00004-of-00006.safetensors",
199
+ "model.layers.21.self_attn.q_norm.bias": "model-00004-of-00006.safetensors",
200
+ "model.layers.21.self_attn.q_norm.weight": "model-00004-of-00006.safetensors",
201
+ "model.layers.21.self_attn.q_proj.weight": "model-00004-of-00006.safetensors",
202
+ "model.layers.21.self_attn.v_proj.weight": "model-00004-of-00006.safetensors",
203
+ "model.layers.22.input_layernorm.weight": "model-00005-of-00006.safetensors",
204
+ "model.layers.22.mlp.down_proj.weight": "model-00005-of-00006.safetensors",
205
+ "model.layers.22.mlp.gate_proj.weight": "model-00004-of-00006.safetensors",
206
+ "model.layers.22.mlp.up_proj.weight": "model-00004-of-00006.safetensors",
207
+ "model.layers.22.post_attention_layernorm.weight": "model-00005-of-00006.safetensors",
208
+ "model.layers.22.self_attn.k_norm.bias": "model-00004-of-00006.safetensors",
209
+ "model.layers.22.self_attn.k_norm.weight": "model-00004-of-00006.safetensors",
210
+ "model.layers.22.self_attn.k_proj.weight": "model-00004-of-00006.safetensors",
211
+ "model.layers.22.self_attn.o_proj.weight": "model-00004-of-00006.safetensors",
212
+ "model.layers.22.self_attn.q_norm.bias": "model-00004-of-00006.safetensors",
213
+ "model.layers.22.self_attn.q_norm.weight": "model-00004-of-00006.safetensors",
214
+ "model.layers.22.self_attn.q_proj.weight": "model-00004-of-00006.safetensors",
215
+ "model.layers.22.self_attn.v_proj.weight": "model-00004-of-00006.safetensors",
216
+ "model.layers.23.input_layernorm.weight": "model-00005-of-00006.safetensors",
217
+ "model.layers.23.mlp.down_proj.weight": "model-00005-of-00006.safetensors",
218
+ "model.layers.23.mlp.gate_proj.weight": "model-00005-of-00006.safetensors",
219
+ "model.layers.23.mlp.up_proj.weight": "model-00005-of-00006.safetensors",
220
+ "model.layers.23.post_attention_layernorm.weight": "model-00005-of-00006.safetensors",
221
+ "model.layers.23.self_attn.k_norm.bias": "model-00005-of-00006.safetensors",
222
+ "model.layers.23.self_attn.k_norm.weight": "model-00005-of-00006.safetensors",
223
+ "model.layers.23.self_attn.k_proj.weight": "model-00005-of-00006.safetensors",
224
+ "model.layers.23.self_attn.o_proj.weight": "model-00005-of-00006.safetensors",
225
+ "model.layers.23.self_attn.q_norm.bias": "model-00005-of-00006.safetensors",
226
+ "model.layers.23.self_attn.q_norm.weight": "model-00005-of-00006.safetensors",
227
+ "model.layers.23.self_attn.q_proj.weight": "model-00005-of-00006.safetensors",
228
+ "model.layers.23.self_attn.v_proj.weight": "model-00005-of-00006.safetensors",
229
+ "model.layers.24.input_layernorm.weight": "model-00005-of-00006.safetensors",
230
+ "model.layers.24.mlp.down_proj.weight": "model-00005-of-00006.safetensors",
231
+ "model.layers.24.mlp.gate_proj.weight": "model-00005-of-00006.safetensors",
232
+ "model.layers.24.mlp.up_proj.weight": "model-00005-of-00006.safetensors",
233
+ "model.layers.24.post_attention_layernorm.weight": "model-00005-of-00006.safetensors",
234
+ "model.layers.24.self_attn.k_norm.bias": "model-00005-of-00006.safetensors",
235
+ "model.layers.24.self_attn.k_norm.weight": "model-00005-of-00006.safetensors",
236
+ "model.layers.24.self_attn.k_proj.weight": "model-00005-of-00006.safetensors",
237
+ "model.layers.24.self_attn.o_proj.weight": "model-00005-of-00006.safetensors",
238
+ "model.layers.24.self_attn.q_norm.bias": "model-00005-of-00006.safetensors",
239
+ "model.layers.24.self_attn.q_norm.weight": "model-00005-of-00006.safetensors",
240
+ "model.layers.24.self_attn.q_proj.weight": "model-00005-of-00006.safetensors",
241
+ "model.layers.24.self_attn.v_proj.weight": "model-00005-of-00006.safetensors",
242
+ "model.layers.25.input_layernorm.weight": "model-00005-of-00006.safetensors",
243
+ "model.layers.25.mlp.down_proj.weight": "model-00005-of-00006.safetensors",
244
+ "model.layers.25.mlp.gate_proj.weight": "model-00005-of-00006.safetensors",
245
+ "model.layers.25.mlp.up_proj.weight": "model-00005-of-00006.safetensors",
246
+ "model.layers.25.post_attention_layernorm.weight": "model-00005-of-00006.safetensors",
247
+ "model.layers.25.self_attn.k_norm.bias": "model-00005-of-00006.safetensors",
248
+ "model.layers.25.self_attn.k_norm.weight": "model-00005-of-00006.safetensors",
249
+ "model.layers.25.self_attn.k_proj.weight": "model-00005-of-00006.safetensors",
250
+ "model.layers.25.self_attn.o_proj.weight": "model-00005-of-00006.safetensors",
251
+ "model.layers.25.self_attn.q_norm.bias": "model-00005-of-00006.safetensors",
252
+ "model.layers.25.self_attn.q_norm.weight": "model-00005-of-00006.safetensors",
253
+ "model.layers.25.self_attn.q_proj.weight": "model-00005-of-00006.safetensors",
254
+ "model.layers.25.self_attn.v_proj.weight": "model-00005-of-00006.safetensors",
255
+ "model.layers.26.input_layernorm.weight": "model-00005-of-00006.safetensors",
256
+ "model.layers.26.mlp.down_proj.weight": "model-00005-of-00006.safetensors",
257
+ "model.layers.26.mlp.gate_proj.weight": "model-00005-of-00006.safetensors",
258
+ "model.layers.26.mlp.up_proj.weight": "model-00005-of-00006.safetensors",
259
+ "model.layers.26.post_attention_layernorm.weight": "model-00005-of-00006.safetensors",
260
+ "model.layers.26.self_attn.k_norm.bias": "model-00005-of-00006.safetensors",
261
+ "model.layers.26.self_attn.k_norm.weight": "model-00005-of-00006.safetensors",
262
+ "model.layers.26.self_attn.k_proj.weight": "model-00005-of-00006.safetensors",
263
+ "model.layers.26.self_attn.o_proj.weight": "model-00005-of-00006.safetensors",
264
+ "model.layers.26.self_attn.q_norm.bias": "model-00005-of-00006.safetensors",
265
+ "model.layers.26.self_attn.q_norm.weight": "model-00005-of-00006.safetensors",
266
+ "model.layers.26.self_attn.q_proj.weight": "model-00005-of-00006.safetensors",
267
+ "model.layers.26.self_attn.v_proj.weight": "model-00005-of-00006.safetensors",
268
+ "model.layers.27.input_layernorm.weight": "model-00005-of-00006.safetensors",
269
+ "model.layers.27.mlp.down_proj.weight": "model-00005-of-00006.safetensors",
270
+ "model.layers.27.mlp.gate_proj.weight": "model-00005-of-00006.safetensors",
271
+ "model.layers.27.mlp.up_proj.weight": "model-00005-of-00006.safetensors",
272
+ "model.layers.27.post_attention_layernorm.weight": "model-00005-of-00006.safetensors",
273
+ "model.layers.27.self_attn.k_norm.bias": "model-00005-of-00006.safetensors",
274
+ "model.layers.27.self_attn.k_norm.weight": "model-00005-of-00006.safetensors",
275
+ "model.layers.27.self_attn.k_proj.weight": "model-00005-of-00006.safetensors",
276
+ "model.layers.27.self_attn.o_proj.weight": "model-00005-of-00006.safetensors",
277
+ "model.layers.27.self_attn.q_norm.bias": "model-00005-of-00006.safetensors",
278
+ "model.layers.27.self_attn.q_norm.weight": "model-00005-of-00006.safetensors",
279
+ "model.layers.27.self_attn.q_proj.weight": "model-00005-of-00006.safetensors",
280
+ "model.layers.27.self_attn.v_proj.weight": "model-00005-of-00006.safetensors",
281
+ "model.layers.28.input_layernorm.weight": "model-00006-of-00006.safetensors",
282
+ "model.layers.28.mlp.down_proj.weight": "model-00006-of-00006.safetensors",
283
+ "model.layers.28.mlp.gate_proj.weight": "model-00005-of-00006.safetensors",
284
+ "model.layers.28.mlp.up_proj.weight": "model-00005-of-00006.safetensors",
285
+ "model.layers.28.post_attention_layernorm.weight": "model-00006-of-00006.safetensors",
286
+ "model.layers.28.self_attn.k_norm.bias": "model-00005-of-00006.safetensors",
287
+ "model.layers.28.self_attn.k_norm.weight": "model-00005-of-00006.safetensors",
288
+ "model.layers.28.self_attn.k_proj.weight": "model-00005-of-00006.safetensors",
289
+ "model.layers.28.self_attn.o_proj.weight": "model-00005-of-00006.safetensors",
290
+ "model.layers.28.self_attn.q_norm.bias": "model-00005-of-00006.safetensors",
291
+ "model.layers.28.self_attn.q_norm.weight": "model-00005-of-00006.safetensors",
292
+ "model.layers.28.self_attn.q_proj.weight": "model-00005-of-00006.safetensors",
293
+ "model.layers.28.self_attn.v_proj.weight": "model-00005-of-00006.safetensors",
294
+ "model.layers.29.input_layernorm.weight": "model-00006-of-00006.safetensors",
295
+ "model.layers.29.mlp.down_proj.weight": "model-00006-of-00006.safetensors",
296
+ "model.layers.29.mlp.gate_proj.weight": "model-00006-of-00006.safetensors",
297
+ "model.layers.29.mlp.up_proj.weight": "model-00006-of-00006.safetensors",
298
+ "model.layers.29.post_attention_layernorm.weight": "model-00006-of-00006.safetensors",
299
+ "model.layers.29.self_attn.k_norm.bias": "model-00006-of-00006.safetensors",
300
+ "model.layers.29.self_attn.k_norm.weight": "model-00006-of-00006.safetensors",
301
+ "model.layers.29.self_attn.k_proj.weight": "model-00006-of-00006.safetensors",
302
+ "model.layers.29.self_attn.o_proj.weight": "model-00006-of-00006.safetensors",
303
+ "model.layers.29.self_attn.q_norm.bias": "model-00006-of-00006.safetensors",
304
+ "model.layers.29.self_attn.q_norm.weight": "model-00006-of-00006.safetensors",
305
+ "model.layers.29.self_attn.q_proj.weight": "model-00006-of-00006.safetensors",
306
+ "model.layers.29.self_attn.v_proj.weight": "model-00006-of-00006.safetensors",
307
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00006.safetensors",
308
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00006.safetensors",
309
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00006.safetensors",
310
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00006.safetensors",
311
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00006.safetensors",
312
+ "model.layers.3.self_attn.k_norm.bias": "model-00001-of-00006.safetensors",
313
+ "model.layers.3.self_attn.k_norm.weight": "model-00001-of-00006.safetensors",
314
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
315
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00006.safetensors",
316
+ "model.layers.3.self_attn.q_norm.bias": "model-00001-of-00006.safetensors",
317
+ "model.layers.3.self_attn.q_norm.weight": "model-00001-of-00006.safetensors",
318
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
319
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
320
+ "model.layers.30.input_layernorm.weight": "model-00006-of-00006.safetensors",
321
+ "model.layers.30.mlp.down_proj.weight": "model-00006-of-00006.safetensors",
322
+ "model.layers.30.mlp.gate_proj.weight": "model-00006-of-00006.safetensors",
323
+ "model.layers.30.mlp.up_proj.weight": "model-00006-of-00006.safetensors",
324
+ "model.layers.30.post_attention_layernorm.weight": "model-00006-of-00006.safetensors",
325
+ "model.layers.30.self_attn.k_norm.bias": "model-00006-of-00006.safetensors",
326
+ "model.layers.30.self_attn.k_norm.weight": "model-00006-of-00006.safetensors",
327
+ "model.layers.30.self_attn.k_proj.weight": "model-00006-of-00006.safetensors",
328
+ "model.layers.30.self_attn.o_proj.weight": "model-00006-of-00006.safetensors",
329
+ "model.layers.30.self_attn.q_norm.bias": "model-00006-of-00006.safetensors",
330
+ "model.layers.30.self_attn.q_norm.weight": "model-00006-of-00006.safetensors",
331
+ "model.layers.30.self_attn.q_proj.weight": "model-00006-of-00006.safetensors",
332
+ "model.layers.30.self_attn.v_proj.weight": "model-00006-of-00006.safetensors",
333
+ "model.layers.31.input_layernorm.weight": "model-00006-of-00006.safetensors",
334
+ "model.layers.31.mlp.down_proj.weight": "model-00006-of-00006.safetensors",
335
+ "model.layers.31.mlp.gate_proj.weight": "model-00006-of-00006.safetensors",
336
+ "model.layers.31.mlp.up_proj.weight": "model-00006-of-00006.safetensors",
337
+ "model.layers.31.post_attention_layernorm.weight": "model-00006-of-00006.safetensors",
338
+ "model.layers.31.self_attn.k_norm.bias": "model-00006-of-00006.safetensors",
339
+ "model.layers.31.self_attn.k_norm.weight": "model-00006-of-00006.safetensors",
340
+ "model.layers.31.self_attn.k_proj.weight": "model-00006-of-00006.safetensors",
341
+ "model.layers.31.self_attn.o_proj.weight": "model-00006-of-00006.safetensors",
342
+ "model.layers.31.self_attn.q_norm.bias": "model-00006-of-00006.safetensors",
343
+ "model.layers.31.self_attn.q_norm.weight": "model-00006-of-00006.safetensors",
344
+ "model.layers.31.self_attn.q_proj.weight": "model-00006-of-00006.safetensors",
345
+ "model.layers.31.self_attn.v_proj.weight": "model-00006-of-00006.safetensors",
346
+ "model.layers.4.input_layernorm.weight": "model-00002-of-00006.safetensors",
347
+ "model.layers.4.mlp.down_proj.weight": "model-00002-of-00006.safetensors",
348
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00006.safetensors",
349
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00006.safetensors",
350
+ "model.layers.4.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
351
+ "model.layers.4.self_attn.k_norm.bias": "model-00001-of-00006.safetensors",
352
+ "model.layers.4.self_attn.k_norm.weight": "model-00001-of-00006.safetensors",
353
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
354
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00006.safetensors",
355
+ "model.layers.4.self_attn.q_norm.bias": "model-00001-of-00006.safetensors",
356
+ "model.layers.4.self_attn.q_norm.weight": "model-00001-of-00006.safetensors",
357
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
358
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
359
+ "model.layers.5.input_layernorm.weight": "model-00002-of-00006.safetensors",
360
+ "model.layers.5.mlp.down_proj.weight": "model-00002-of-00006.safetensors",
361
+ "model.layers.5.mlp.gate_proj.weight": "model-00002-of-00006.safetensors",
362
+ "model.layers.5.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
363
+ "model.layers.5.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
364
+ "model.layers.5.self_attn.k_norm.bias": "model-00002-of-00006.safetensors",
365
+ "model.layers.5.self_attn.k_norm.weight": "model-00002-of-00006.safetensors",
366
+ "model.layers.5.self_attn.k_proj.weight": "model-00002-of-00006.safetensors",
367
+ "model.layers.5.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
368
+ "model.layers.5.self_attn.q_norm.bias": "model-00002-of-00006.safetensors",
369
+ "model.layers.5.self_attn.q_norm.weight": "model-00002-of-00006.safetensors",
370
+ "model.layers.5.self_attn.q_proj.weight": "model-00002-of-00006.safetensors",
371
+ "model.layers.5.self_attn.v_proj.weight": "model-00002-of-00006.safetensors",
372
+ "model.layers.6.input_layernorm.weight": "model-00002-of-00006.safetensors",
373
+ "model.layers.6.mlp.down_proj.weight": "model-00002-of-00006.safetensors",
374
+ "model.layers.6.mlp.gate_proj.weight": "model-00002-of-00006.safetensors",
375
+ "model.layers.6.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
376
+ "model.layers.6.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
377
+ "model.layers.6.self_attn.k_norm.bias": "model-00002-of-00006.safetensors",
378
+ "model.layers.6.self_attn.k_norm.weight": "model-00002-of-00006.safetensors",
379
+ "model.layers.6.self_attn.k_proj.weight": "model-00002-of-00006.safetensors",
380
+ "model.layers.6.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
381
+ "model.layers.6.self_attn.q_norm.bias": "model-00002-of-00006.safetensors",
382
+ "model.layers.6.self_attn.q_norm.weight": "model-00002-of-00006.safetensors",
383
+ "model.layers.6.self_attn.q_proj.weight": "model-00002-of-00006.safetensors",
384
+ "model.layers.6.self_attn.v_proj.weight": "model-00002-of-00006.safetensors",
385
+ "model.layers.7.input_layernorm.weight": "model-00002-of-00006.safetensors",
386
+ "model.layers.7.mlp.down_proj.weight": "model-00002-of-00006.safetensors",
387
+ "model.layers.7.mlp.gate_proj.weight": "model-00002-of-00006.safetensors",
388
+ "model.layers.7.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
389
+ "model.layers.7.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
390
+ "model.layers.7.self_attn.k_norm.bias": "model-00002-of-00006.safetensors",
391
+ "model.layers.7.self_attn.k_norm.weight": "model-00002-of-00006.safetensors",
392
+ "model.layers.7.self_attn.k_proj.weight": "model-00002-of-00006.safetensors",
393
+ "model.layers.7.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
394
+ "model.layers.7.self_attn.q_norm.bias": "model-00002-of-00006.safetensors",
395
+ "model.layers.7.self_attn.q_norm.weight": "model-00002-of-00006.safetensors",
396
+ "model.layers.7.self_attn.q_proj.weight": "model-00002-of-00006.safetensors",
397
+ "model.layers.7.self_attn.v_proj.weight": "model-00002-of-00006.safetensors",
398
+ "model.layers.8.input_layernorm.weight": "model-00002-of-00006.safetensors",
399
+ "model.layers.8.mlp.down_proj.weight": "model-00002-of-00006.safetensors",
400
+ "model.layers.8.mlp.gate_proj.weight": "model-00002-of-00006.safetensors",
401
+ "model.layers.8.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
402
+ "model.layers.8.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
403
+ "model.layers.8.self_attn.k_norm.bias": "model-00002-of-00006.safetensors",
404
+ "model.layers.8.self_attn.k_norm.weight": "model-00002-of-00006.safetensors",
405
+ "model.layers.8.self_attn.k_proj.weight": "model-00002-of-00006.safetensors",
406
+ "model.layers.8.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
407
+ "model.layers.8.self_attn.q_norm.bias": "model-00002-of-00006.safetensors",
408
+ "model.layers.8.self_attn.q_norm.weight": "model-00002-of-00006.safetensors",
409
+ "model.layers.8.self_attn.q_proj.weight": "model-00002-of-00006.safetensors",
410
+ "model.layers.8.self_attn.v_proj.weight": "model-00002-of-00006.safetensors",
411
+ "model.layers.9.input_layernorm.weight": "model-00002-of-00006.safetensors",
412
+ "model.layers.9.mlp.down_proj.weight": "model-00002-of-00006.safetensors",
413
+ "model.layers.9.mlp.gate_proj.weight": "model-00002-of-00006.safetensors",
414
+ "model.layers.9.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
415
+ "model.layers.9.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
416
+ "model.layers.9.self_attn.k_norm.bias": "model-00002-of-00006.safetensors",
417
+ "model.layers.9.self_attn.k_norm.weight": "model-00002-of-00006.safetensors",
418
+ "model.layers.9.self_attn.k_proj.weight": "model-00002-of-00006.safetensors",
419
+ "model.layers.9.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
420
+ "model.layers.9.self_attn.q_norm.bias": "model-00002-of-00006.safetensors",
421
+ "model.layers.9.self_attn.q_norm.weight": "model-00002-of-00006.safetensors",
422
+ "model.layers.9.self_attn.q_proj.weight": "model-00002-of-00006.safetensors",
423
+ "model.layers.9.self_attn.v_proj.weight": "model-00002-of-00006.safetensors",
424
+ "model.norm.weight": "model-00006-of-00006.safetensors",
425
+ "model.vqmodel.decoder.conv_in.bias": "model-00006-of-00006.safetensors",
426
+ "model.vqmodel.decoder.conv_in.weight": "model-00006-of-00006.safetensors",
427
+ "model.vqmodel.decoder.conv_out.bias": "model-00006-of-00006.safetensors",
428
+ "model.vqmodel.decoder.conv_out.weight": "model-00006-of-00006.safetensors",
429
+ "model.vqmodel.decoder.mid.attn_1.k.bias": "model-00006-of-00006.safetensors",
430
+ "model.vqmodel.decoder.mid.attn_1.k.weight": "model-00006-of-00006.safetensors",
431
+ "model.vqmodel.decoder.mid.attn_1.norm.bias": "model-00006-of-00006.safetensors",
432
+ "model.vqmodel.decoder.mid.attn_1.norm.weight": "model-00006-of-00006.safetensors",
433
+ "model.vqmodel.decoder.mid.attn_1.proj_out.bias": "model-00006-of-00006.safetensors",
434
+ "model.vqmodel.decoder.mid.attn_1.proj_out.weight": "model-00006-of-00006.safetensors",
435
+ "model.vqmodel.decoder.mid.attn_1.q.bias": "model-00006-of-00006.safetensors",
436
+ "model.vqmodel.decoder.mid.attn_1.q.weight": "model-00006-of-00006.safetensors",
437
+ "model.vqmodel.decoder.mid.attn_1.v.bias": "model-00006-of-00006.safetensors",
438
+ "model.vqmodel.decoder.mid.attn_1.v.weight": "model-00006-of-00006.safetensors",
439
+ "model.vqmodel.decoder.mid.block_1.conv1.bias": "model-00006-of-00006.safetensors",
440
+ "model.vqmodel.decoder.mid.block_1.conv1.weight": "model-00006-of-00006.safetensors",
441
+ "model.vqmodel.decoder.mid.block_1.conv2.bias": "model-00006-of-00006.safetensors",
442
+ "model.vqmodel.decoder.mid.block_1.conv2.weight": "model-00006-of-00006.safetensors",
443
+ "model.vqmodel.decoder.mid.block_1.norm1.bias": "model-00006-of-00006.safetensors",
444
+ "model.vqmodel.decoder.mid.block_1.norm1.weight": "model-00006-of-00006.safetensors",
445
+ "model.vqmodel.decoder.mid.block_1.norm2.bias": "model-00006-of-00006.safetensors",
446
+ "model.vqmodel.decoder.mid.block_1.norm2.weight": "model-00006-of-00006.safetensors",
447
+ "model.vqmodel.decoder.mid.block_2.conv1.bias": "model-00006-of-00006.safetensors",
448
+ "model.vqmodel.decoder.mid.block_2.conv1.weight": "model-00006-of-00006.safetensors",
449
+ "model.vqmodel.decoder.mid.block_2.conv2.bias": "model-00006-of-00006.safetensors",
450
+ "model.vqmodel.decoder.mid.block_2.conv2.weight": "model-00006-of-00006.safetensors",
451
+ "model.vqmodel.decoder.mid.block_2.norm1.bias": "model-00006-of-00006.safetensors",
452
+ "model.vqmodel.decoder.mid.block_2.norm1.weight": "model-00006-of-00006.safetensors",
453
+ "model.vqmodel.decoder.mid.block_2.norm2.bias": "model-00006-of-00006.safetensors",
454
+ "model.vqmodel.decoder.mid.block_2.norm2.weight": "model-00006-of-00006.safetensors",
455
+ "model.vqmodel.decoder.norm_out.bias": "model-00006-of-00006.safetensors",
456
+ "model.vqmodel.decoder.norm_out.weight": "model-00006-of-00006.safetensors",
457
+ "model.vqmodel.decoder.up.0.block.0.conv1.bias": "model-00006-of-00006.safetensors",
458
+ "model.vqmodel.decoder.up.0.block.0.conv1.weight": "model-00006-of-00006.safetensors",
459
+ "model.vqmodel.decoder.up.0.block.0.conv2.bias": "model-00006-of-00006.safetensors",
460
+ "model.vqmodel.decoder.up.0.block.0.conv2.weight": "model-00006-of-00006.safetensors",
461
+ "model.vqmodel.decoder.up.0.block.0.norm1.bias": "model-00006-of-00006.safetensors",
462
+ "model.vqmodel.decoder.up.0.block.0.norm1.weight": "model-00006-of-00006.safetensors",
463
+ "model.vqmodel.decoder.up.0.block.0.norm2.bias": "model-00006-of-00006.safetensors",
464
+ "model.vqmodel.decoder.up.0.block.0.norm2.weight": "model-00006-of-00006.safetensors",
465
+ "model.vqmodel.decoder.up.0.block.1.conv1.bias": "model-00006-of-00006.safetensors",
466
+ "model.vqmodel.decoder.up.0.block.1.conv1.weight": "model-00006-of-00006.safetensors",
467
+ "model.vqmodel.decoder.up.0.block.1.conv2.bias": "model-00006-of-00006.safetensors",
468
+ "model.vqmodel.decoder.up.0.block.1.conv2.weight": "model-00006-of-00006.safetensors",
469
+ "model.vqmodel.decoder.up.0.block.1.norm1.bias": "model-00006-of-00006.safetensors",
470
+ "model.vqmodel.decoder.up.0.block.1.norm1.weight": "model-00006-of-00006.safetensors",
471
+ "model.vqmodel.decoder.up.0.block.1.norm2.bias": "model-00006-of-00006.safetensors",
472
+ "model.vqmodel.decoder.up.0.block.1.norm2.weight": "model-00006-of-00006.safetensors",
473
+ "model.vqmodel.decoder.up.0.block.2.conv1.bias": "model-00006-of-00006.safetensors",
474
+ "model.vqmodel.decoder.up.0.block.2.conv1.weight": "model-00006-of-00006.safetensors",
475
+ "model.vqmodel.decoder.up.0.block.2.conv2.bias": "model-00006-of-00006.safetensors",
476
+ "model.vqmodel.decoder.up.0.block.2.conv2.weight": "model-00006-of-00006.safetensors",
477
+ "model.vqmodel.decoder.up.0.block.2.norm1.bias": "model-00006-of-00006.safetensors",
478
+ "model.vqmodel.decoder.up.0.block.2.norm1.weight": "model-00006-of-00006.safetensors",
479
+ "model.vqmodel.decoder.up.0.block.2.norm2.bias": "model-00006-of-00006.safetensors",
480
+ "model.vqmodel.decoder.up.0.block.2.norm2.weight": "model-00006-of-00006.safetensors",
481
+ "model.vqmodel.decoder.up.1.block.0.conv1.bias": "model-00006-of-00006.safetensors",
482
+ "model.vqmodel.decoder.up.1.block.0.conv1.weight": "model-00006-of-00006.safetensors",
483
+ "model.vqmodel.decoder.up.1.block.0.conv2.bias": "model-00006-of-00006.safetensors",
484
+ "model.vqmodel.decoder.up.1.block.0.conv2.weight": "model-00006-of-00006.safetensors",
485
+ "model.vqmodel.decoder.up.1.block.0.nin_shortcut.bias": "model-00006-of-00006.safetensors",
486
+ "model.vqmodel.decoder.up.1.block.0.nin_shortcut.weight": "model-00006-of-00006.safetensors",
487
+ "model.vqmodel.decoder.up.1.block.0.norm1.bias": "model-00006-of-00006.safetensors",
488
+ "model.vqmodel.decoder.up.1.block.0.norm1.weight": "model-00006-of-00006.safetensors",
489
+ "model.vqmodel.decoder.up.1.block.0.norm2.bias": "model-00006-of-00006.safetensors",
490
+ "model.vqmodel.decoder.up.1.block.0.norm2.weight": "model-00006-of-00006.safetensors",
491
+ "model.vqmodel.decoder.up.1.block.1.conv1.bias": "model-00006-of-00006.safetensors",
492
+ "model.vqmodel.decoder.up.1.block.1.conv1.weight": "model-00006-of-00006.safetensors",
493
+ "model.vqmodel.decoder.up.1.block.1.conv2.bias": "model-00006-of-00006.safetensors",
494
+ "model.vqmodel.decoder.up.1.block.1.conv2.weight": "model-00006-of-00006.safetensors",
495
+ "model.vqmodel.decoder.up.1.block.1.norm1.bias": "model-00006-of-00006.safetensors",
496
+ "model.vqmodel.decoder.up.1.block.1.norm1.weight": "model-00006-of-00006.safetensors",
497
+ "model.vqmodel.decoder.up.1.block.1.norm2.bias": "model-00006-of-00006.safetensors",
498
+ "model.vqmodel.decoder.up.1.block.1.norm2.weight": "model-00006-of-00006.safetensors",
499
+ "model.vqmodel.decoder.up.1.block.2.conv1.bias": "model-00006-of-00006.safetensors",
500
+ "model.vqmodel.decoder.up.1.block.2.conv1.weight": "model-00006-of-00006.safetensors",
501
+ "model.vqmodel.decoder.up.1.block.2.conv2.bias": "model-00006-of-00006.safetensors",
502
+ "model.vqmodel.decoder.up.1.block.2.conv2.weight": "model-00006-of-00006.safetensors",
503
+ "model.vqmodel.decoder.up.1.block.2.norm1.bias": "model-00006-of-00006.safetensors",
504
+ "model.vqmodel.decoder.up.1.block.2.norm1.weight": "model-00006-of-00006.safetensors",
505
+ "model.vqmodel.decoder.up.1.block.2.norm2.bias": "model-00006-of-00006.safetensors",
506
+ "model.vqmodel.decoder.up.1.block.2.norm2.weight": "model-00006-of-00006.safetensors",
507
+ "model.vqmodel.decoder.up.1.upsample.conv.bias": "model-00006-of-00006.safetensors",
508
+ "model.vqmodel.decoder.up.1.upsample.conv.weight": "model-00006-of-00006.safetensors",
509
+ "model.vqmodel.decoder.up.2.block.0.conv1.bias": "model-00006-of-00006.safetensors",
510
+ "model.vqmodel.decoder.up.2.block.0.conv1.weight": "model-00006-of-00006.safetensors",
511
+ "model.vqmodel.decoder.up.2.block.0.conv2.bias": "model-00006-of-00006.safetensors",
512
+ "model.vqmodel.decoder.up.2.block.0.conv2.weight": "model-00006-of-00006.safetensors",
513
+ "model.vqmodel.decoder.up.2.block.0.norm1.bias": "model-00006-of-00006.safetensors",
514
+ "model.vqmodel.decoder.up.2.block.0.norm1.weight": "model-00006-of-00006.safetensors",
515
+ "model.vqmodel.decoder.up.2.block.0.norm2.bias": "model-00006-of-00006.safetensors",
516
+ "model.vqmodel.decoder.up.2.block.0.norm2.weight": "model-00006-of-00006.safetensors",
517
+ "model.vqmodel.decoder.up.2.block.1.conv1.bias": "model-00006-of-00006.safetensors",
518
+ "model.vqmodel.decoder.up.2.block.1.conv1.weight": "model-00006-of-00006.safetensors",
519
+ "model.vqmodel.decoder.up.2.block.1.conv2.bias": "model-00006-of-00006.safetensors",
520
+ "model.vqmodel.decoder.up.2.block.1.conv2.weight": "model-00006-of-00006.safetensors",
521
+ "model.vqmodel.decoder.up.2.block.1.norm1.bias": "model-00006-of-00006.safetensors",
522
+ "model.vqmodel.decoder.up.2.block.1.norm1.weight": "model-00006-of-00006.safetensors",
523
+ "model.vqmodel.decoder.up.2.block.1.norm2.bias": "model-00006-of-00006.safetensors",
524
+ "model.vqmodel.decoder.up.2.block.1.norm2.weight": "model-00006-of-00006.safetensors",
525
+ "model.vqmodel.decoder.up.2.block.2.conv1.bias": "model-00006-of-00006.safetensors",
526
+ "model.vqmodel.decoder.up.2.block.2.conv1.weight": "model-00006-of-00006.safetensors",
527
+ "model.vqmodel.decoder.up.2.block.2.conv2.bias": "model-00006-of-00006.safetensors",
528
+ "model.vqmodel.decoder.up.2.block.2.conv2.weight": "model-00006-of-00006.safetensors",
529
+ "model.vqmodel.decoder.up.2.block.2.norm1.bias": "model-00006-of-00006.safetensors",
530
+ "model.vqmodel.decoder.up.2.block.2.norm1.weight": "model-00006-of-00006.safetensors",
531
+ "model.vqmodel.decoder.up.2.block.2.norm2.bias": "model-00006-of-00006.safetensors",
532
+ "model.vqmodel.decoder.up.2.block.2.norm2.weight": "model-00006-of-00006.safetensors",
533
+ "model.vqmodel.decoder.up.2.upsample.conv.bias": "model-00006-of-00006.safetensors",
534
+ "model.vqmodel.decoder.up.2.upsample.conv.weight": "model-00006-of-00006.safetensors",
535
+ "model.vqmodel.decoder.up.3.block.0.conv1.bias": "model-00006-of-00006.safetensors",
536
+ "model.vqmodel.decoder.up.3.block.0.conv1.weight": "model-00006-of-00006.safetensors",
537
+ "model.vqmodel.decoder.up.3.block.0.conv2.bias": "model-00006-of-00006.safetensors",
538
+ "model.vqmodel.decoder.up.3.block.0.conv2.weight": "model-00006-of-00006.safetensors",
539
+ "model.vqmodel.decoder.up.3.block.0.nin_shortcut.bias": "model-00006-of-00006.safetensors",
540
+ "model.vqmodel.decoder.up.3.block.0.nin_shortcut.weight": "model-00006-of-00006.safetensors",
541
+ "model.vqmodel.decoder.up.3.block.0.norm1.bias": "model-00006-of-00006.safetensors",
542
+ "model.vqmodel.decoder.up.3.block.0.norm1.weight": "model-00006-of-00006.safetensors",
543
+ "model.vqmodel.decoder.up.3.block.0.norm2.bias": "model-00006-of-00006.safetensors",
544
+ "model.vqmodel.decoder.up.3.block.0.norm2.weight": "model-00006-of-00006.safetensors",
545
+ "model.vqmodel.decoder.up.3.block.1.conv1.bias": "model-00006-of-00006.safetensors",
546
+ "model.vqmodel.decoder.up.3.block.1.conv1.weight": "model-00006-of-00006.safetensors",
547
+ "model.vqmodel.decoder.up.3.block.1.conv2.bias": "model-00006-of-00006.safetensors",
548
+ "model.vqmodel.decoder.up.3.block.1.conv2.weight": "model-00006-of-00006.safetensors",
549
+ "model.vqmodel.decoder.up.3.block.1.norm1.bias": "model-00006-of-00006.safetensors",
550
+ "model.vqmodel.decoder.up.3.block.1.norm1.weight": "model-00006-of-00006.safetensors",
551
+ "model.vqmodel.decoder.up.3.block.1.norm2.bias": "model-00006-of-00006.safetensors",
552
+ "model.vqmodel.decoder.up.3.block.1.norm2.weight": "model-00006-of-00006.safetensors",
553
+ "model.vqmodel.decoder.up.3.block.2.conv1.bias": "model-00006-of-00006.safetensors",
554
+ "model.vqmodel.decoder.up.3.block.2.conv1.weight": "model-00006-of-00006.safetensors",
555
+ "model.vqmodel.decoder.up.3.block.2.conv2.bias": "model-00006-of-00006.safetensors",
556
+ "model.vqmodel.decoder.up.3.block.2.conv2.weight": "model-00006-of-00006.safetensors",
557
+ "model.vqmodel.decoder.up.3.block.2.norm1.bias": "model-00006-of-00006.safetensors",
558
+ "model.vqmodel.decoder.up.3.block.2.norm1.weight": "model-00006-of-00006.safetensors",
559
+ "model.vqmodel.decoder.up.3.block.2.norm2.bias": "model-00006-of-00006.safetensors",
560
+ "model.vqmodel.decoder.up.3.block.2.norm2.weight": "model-00006-of-00006.safetensors",
561
+ "model.vqmodel.decoder.up.3.upsample.conv.bias": "model-00006-of-00006.safetensors",
562
+ "model.vqmodel.decoder.up.3.upsample.conv.weight": "model-00006-of-00006.safetensors",
563
+ "model.vqmodel.decoder.up.4.block.0.conv1.bias": "model-00006-of-00006.safetensors",
564
+ "model.vqmodel.decoder.up.4.block.0.conv1.weight": "model-00006-of-00006.safetensors",
565
+ "model.vqmodel.decoder.up.4.block.0.conv2.bias": "model-00006-of-00006.safetensors",
566
+ "model.vqmodel.decoder.up.4.block.0.conv2.weight": "model-00006-of-00006.safetensors",
567
+ "model.vqmodel.decoder.up.4.block.0.norm1.bias": "model-00006-of-00006.safetensors",
568
+ "model.vqmodel.decoder.up.4.block.0.norm1.weight": "model-00006-of-00006.safetensors",
569
+ "model.vqmodel.decoder.up.4.block.0.norm2.bias": "model-00006-of-00006.safetensors",
570
+ "model.vqmodel.decoder.up.4.block.0.norm2.weight": "model-00006-of-00006.safetensors",
571
+ "model.vqmodel.decoder.up.4.block.1.conv1.bias": "model-00006-of-00006.safetensors",
572
+ "model.vqmodel.decoder.up.4.block.1.conv1.weight": "model-00006-of-00006.safetensors",
573
+ "model.vqmodel.decoder.up.4.block.1.conv2.bias": "model-00006-of-00006.safetensors",
574
+ "model.vqmodel.decoder.up.4.block.1.conv2.weight": "model-00006-of-00006.safetensors",
575
+ "model.vqmodel.decoder.up.4.block.1.norm1.bias": "model-00006-of-00006.safetensors",
576
+ "model.vqmodel.decoder.up.4.block.1.norm1.weight": "model-00006-of-00006.safetensors",
577
+ "model.vqmodel.decoder.up.4.block.1.norm2.bias": "model-00006-of-00006.safetensors",
578
+ "model.vqmodel.decoder.up.4.block.1.norm2.weight": "model-00006-of-00006.safetensors",
579
+ "model.vqmodel.decoder.up.4.block.2.conv1.bias": "model-00006-of-00006.safetensors",
580
+ "model.vqmodel.decoder.up.4.block.2.conv1.weight": "model-00006-of-00006.safetensors",
581
+ "model.vqmodel.decoder.up.4.block.2.conv2.bias": "model-00006-of-00006.safetensors",
582
+ "model.vqmodel.decoder.up.4.block.2.conv2.weight": "model-00006-of-00006.safetensors",
583
+ "model.vqmodel.decoder.up.4.block.2.norm1.bias": "model-00006-of-00006.safetensors",
584
+ "model.vqmodel.decoder.up.4.block.2.norm1.weight": "model-00006-of-00006.safetensors",
585
+ "model.vqmodel.decoder.up.4.block.2.norm2.bias": "model-00006-of-00006.safetensors",
586
+ "model.vqmodel.decoder.up.4.block.2.norm2.weight": "model-00006-of-00006.safetensors",
587
+ "model.vqmodel.decoder.up.4.upsample.conv.bias": "model-00006-of-00006.safetensors",
588
+ "model.vqmodel.decoder.up.4.upsample.conv.weight": "model-00006-of-00006.safetensors",
589
+ "model.vqmodel.encoder.conv_in.bias": "model-00006-of-00006.safetensors",
590
+ "model.vqmodel.encoder.conv_in.weight": "model-00006-of-00006.safetensors",
591
+ "model.vqmodel.encoder.conv_out.bias": "model-00006-of-00006.safetensors",
592
+ "model.vqmodel.encoder.conv_out.weight": "model-00006-of-00006.safetensors",
593
+ "model.vqmodel.encoder.down.0.block.0.conv1.bias": "model-00006-of-00006.safetensors",
594
+ "model.vqmodel.encoder.down.0.block.0.conv1.weight": "model-00006-of-00006.safetensors",
595
+ "model.vqmodel.encoder.down.0.block.0.conv2.bias": "model-00006-of-00006.safetensors",
596
+ "model.vqmodel.encoder.down.0.block.0.conv2.weight": "model-00006-of-00006.safetensors",
597
+ "model.vqmodel.encoder.down.0.block.0.norm1.bias": "model-00006-of-00006.safetensors",
598
+ "model.vqmodel.encoder.down.0.block.0.norm1.weight": "model-00006-of-00006.safetensors",
599
+ "model.vqmodel.encoder.down.0.block.0.norm2.bias": "model-00006-of-00006.safetensors",
600
+ "model.vqmodel.encoder.down.0.block.0.norm2.weight": "model-00006-of-00006.safetensors",
601
+ "model.vqmodel.encoder.down.0.block.1.conv1.bias": "model-00006-of-00006.safetensors",
602
+ "model.vqmodel.encoder.down.0.block.1.conv1.weight": "model-00006-of-00006.safetensors",
603
+ "model.vqmodel.encoder.down.0.block.1.conv2.bias": "model-00006-of-00006.safetensors",
604
+ "model.vqmodel.encoder.down.0.block.1.conv2.weight": "model-00006-of-00006.safetensors",
605
+ "model.vqmodel.encoder.down.0.block.1.norm1.bias": "model-00006-of-00006.safetensors",
606
+ "model.vqmodel.encoder.down.0.block.1.norm1.weight": "model-00006-of-00006.safetensors",
607
+ "model.vqmodel.encoder.down.0.block.1.norm2.bias": "model-00006-of-00006.safetensors",
608
+ "model.vqmodel.encoder.down.0.block.1.norm2.weight": "model-00006-of-00006.safetensors",
609
+ "model.vqmodel.encoder.down.0.downsample.conv.bias": "model-00006-of-00006.safetensors",
610
+ "model.vqmodel.encoder.down.0.downsample.conv.weight": "model-00006-of-00006.safetensors",
611
+ "model.vqmodel.encoder.down.1.block.0.conv1.bias": "model-00006-of-00006.safetensors",
612
+ "model.vqmodel.encoder.down.1.block.0.conv1.weight": "model-00006-of-00006.safetensors",
613
+ "model.vqmodel.encoder.down.1.block.0.conv2.bias": "model-00006-of-00006.safetensors",
614
+ "model.vqmodel.encoder.down.1.block.0.conv2.weight": "model-00006-of-00006.safetensors",
615
+ "model.vqmodel.encoder.down.1.block.0.norm1.bias": "model-00006-of-00006.safetensors",
616
+ "model.vqmodel.encoder.down.1.block.0.norm1.weight": "model-00006-of-00006.safetensors",
617
+ "model.vqmodel.encoder.down.1.block.0.norm2.bias": "model-00006-of-00006.safetensors",
618
+ "model.vqmodel.encoder.down.1.block.0.norm2.weight": "model-00006-of-00006.safetensors",
619
+ "model.vqmodel.encoder.down.1.block.1.conv1.bias": "model-00006-of-00006.safetensors",
620
+ "model.vqmodel.encoder.down.1.block.1.conv1.weight": "model-00006-of-00006.safetensors",
621
+ "model.vqmodel.encoder.down.1.block.1.conv2.bias": "model-00006-of-00006.safetensors",
622
+ "model.vqmodel.encoder.down.1.block.1.conv2.weight": "model-00006-of-00006.safetensors",
623
+ "model.vqmodel.encoder.down.1.block.1.norm1.bias": "model-00006-of-00006.safetensors",
624
+ "model.vqmodel.encoder.down.1.block.1.norm1.weight": "model-00006-of-00006.safetensors",
625
+ "model.vqmodel.encoder.down.1.block.1.norm2.bias": "model-00006-of-00006.safetensors",
626
+ "model.vqmodel.encoder.down.1.block.1.norm2.weight": "model-00006-of-00006.safetensors",
627
+ "model.vqmodel.encoder.down.1.downsample.conv.bias": "model-00006-of-00006.safetensors",
628
+ "model.vqmodel.encoder.down.1.downsample.conv.weight": "model-00006-of-00006.safetensors",
629
+ "model.vqmodel.encoder.down.2.block.0.conv1.bias": "model-00006-of-00006.safetensors",
630
+ "model.vqmodel.encoder.down.2.block.0.conv1.weight": "model-00006-of-00006.safetensors",
631
+ "model.vqmodel.encoder.down.2.block.0.conv2.bias": "model-00006-of-00006.safetensors",
632
+ "model.vqmodel.encoder.down.2.block.0.conv2.weight": "model-00006-of-00006.safetensors",
633
+ "model.vqmodel.encoder.down.2.block.0.nin_shortcut.bias": "model-00006-of-00006.safetensors",
634
+ "model.vqmodel.encoder.down.2.block.0.nin_shortcut.weight": "model-00006-of-00006.safetensors",
635
+ "model.vqmodel.encoder.down.2.block.0.norm1.bias": "model-00006-of-00006.safetensors",
636
+ "model.vqmodel.encoder.down.2.block.0.norm1.weight": "model-00006-of-00006.safetensors",
637
+ "model.vqmodel.encoder.down.2.block.0.norm2.bias": "model-00006-of-00006.safetensors",
638
+ "model.vqmodel.encoder.down.2.block.0.norm2.weight": "model-00006-of-00006.safetensors",
639
+ "model.vqmodel.encoder.down.2.block.1.conv1.bias": "model-00006-of-00006.safetensors",
640
+ "model.vqmodel.encoder.down.2.block.1.conv1.weight": "model-00006-of-00006.safetensors",
641
+ "model.vqmodel.encoder.down.2.block.1.conv2.bias": "model-00006-of-00006.safetensors",
642
+ "model.vqmodel.encoder.down.2.block.1.conv2.weight": "model-00006-of-00006.safetensors",
643
+ "model.vqmodel.encoder.down.2.block.1.norm1.bias": "model-00006-of-00006.safetensors",
644
+ "model.vqmodel.encoder.down.2.block.1.norm1.weight": "model-00006-of-00006.safetensors",
645
+ "model.vqmodel.encoder.down.2.block.1.norm2.bias": "model-00006-of-00006.safetensors",
646
+ "model.vqmodel.encoder.down.2.block.1.norm2.weight": "model-00006-of-00006.safetensors",
647
+ "model.vqmodel.encoder.down.2.downsample.conv.bias": "model-00006-of-00006.safetensors",
648
+ "model.vqmodel.encoder.down.2.downsample.conv.weight": "model-00006-of-00006.safetensors",
649
+ "model.vqmodel.encoder.down.3.block.0.conv1.bias": "model-00006-of-00006.safetensors",
650
+ "model.vqmodel.encoder.down.3.block.0.conv1.weight": "model-00006-of-00006.safetensors",
651
+ "model.vqmodel.encoder.down.3.block.0.conv2.bias": "model-00006-of-00006.safetensors",
652
+ "model.vqmodel.encoder.down.3.block.0.conv2.weight": "model-00006-of-00006.safetensors",
653
+ "model.vqmodel.encoder.down.3.block.0.norm1.bias": "model-00006-of-00006.safetensors",
654
+ "model.vqmodel.encoder.down.3.block.0.norm1.weight": "model-00006-of-00006.safetensors",
655
+ "model.vqmodel.encoder.down.3.block.0.norm2.bias": "model-00006-of-00006.safetensors",
656
+ "model.vqmodel.encoder.down.3.block.0.norm2.weight": "model-00006-of-00006.safetensors",
657
+ "model.vqmodel.encoder.down.3.block.1.conv1.bias": "model-00006-of-00006.safetensors",
658
+ "model.vqmodel.encoder.down.3.block.1.conv1.weight": "model-00006-of-00006.safetensors",
659
+ "model.vqmodel.encoder.down.3.block.1.conv2.bias": "model-00006-of-00006.safetensors",
660
+ "model.vqmodel.encoder.down.3.block.1.conv2.weight": "model-00006-of-00006.safetensors",
661
+ "model.vqmodel.encoder.down.3.block.1.norm1.bias": "model-00006-of-00006.safetensors",
662
+ "model.vqmodel.encoder.down.3.block.1.norm1.weight": "model-00006-of-00006.safetensors",
663
+ "model.vqmodel.encoder.down.3.block.1.norm2.bias": "model-00006-of-00006.safetensors",
664
+ "model.vqmodel.encoder.down.3.block.1.norm2.weight": "model-00006-of-00006.safetensors",
665
+ "model.vqmodel.encoder.down.3.downsample.conv.bias": "model-00006-of-00006.safetensors",
666
+ "model.vqmodel.encoder.down.3.downsample.conv.weight": "model-00006-of-00006.safetensors",
667
+ "model.vqmodel.encoder.down.4.block.0.conv1.bias": "model-00006-of-00006.safetensors",
668
+ "model.vqmodel.encoder.down.4.block.0.conv1.weight": "model-00006-of-00006.safetensors",
669
+ "model.vqmodel.encoder.down.4.block.0.conv2.bias": "model-00006-of-00006.safetensors",
670
+ "model.vqmodel.encoder.down.4.block.0.conv2.weight": "model-00006-of-00006.safetensors",
671
+ "model.vqmodel.encoder.down.4.block.0.nin_shortcut.bias": "model-00006-of-00006.safetensors",
672
+ "model.vqmodel.encoder.down.4.block.0.nin_shortcut.weight": "model-00006-of-00006.safetensors",
673
+ "model.vqmodel.encoder.down.4.block.0.norm1.bias": "model-00006-of-00006.safetensors",
674
+ "model.vqmodel.encoder.down.4.block.0.norm1.weight": "model-00006-of-00006.safetensors",
675
+ "model.vqmodel.encoder.down.4.block.0.norm2.bias": "model-00006-of-00006.safetensors",
676
+ "model.vqmodel.encoder.down.4.block.0.norm2.weight": "model-00006-of-00006.safetensors",
677
+ "model.vqmodel.encoder.down.4.block.1.conv1.bias": "model-00006-of-00006.safetensors",
678
+ "model.vqmodel.encoder.down.4.block.1.conv1.weight": "model-00006-of-00006.safetensors",
679
+ "model.vqmodel.encoder.down.4.block.1.conv2.bias": "model-00006-of-00006.safetensors",
680
+ "model.vqmodel.encoder.down.4.block.1.conv2.weight": "model-00006-of-00006.safetensors",
681
+ "model.vqmodel.encoder.down.4.block.1.norm1.bias": "model-00006-of-00006.safetensors",
682
+ "model.vqmodel.encoder.down.4.block.1.norm1.weight": "model-00006-of-00006.safetensors",
683
+ "model.vqmodel.encoder.down.4.block.1.norm2.bias": "model-00006-of-00006.safetensors",
684
+ "model.vqmodel.encoder.down.4.block.1.norm2.weight": "model-00006-of-00006.safetensors",
685
+ "model.vqmodel.encoder.mid.attn_1.k.bias": "model-00006-of-00006.safetensors",
686
+ "model.vqmodel.encoder.mid.attn_1.k.weight": "model-00006-of-00006.safetensors",
687
+ "model.vqmodel.encoder.mid.attn_1.norm.bias": "model-00006-of-00006.safetensors",
688
+ "model.vqmodel.encoder.mid.attn_1.norm.weight": "model-00006-of-00006.safetensors",
689
+ "model.vqmodel.encoder.mid.attn_1.proj_out.bias": "model-00006-of-00006.safetensors",
690
+ "model.vqmodel.encoder.mid.attn_1.proj_out.weight": "model-00006-of-00006.safetensors",
691
+ "model.vqmodel.encoder.mid.attn_1.q.bias": "model-00006-of-00006.safetensors",
692
+ "model.vqmodel.encoder.mid.attn_1.q.weight": "model-00006-of-00006.safetensors",
693
+ "model.vqmodel.encoder.mid.attn_1.v.bias": "model-00006-of-00006.safetensors",
694
+ "model.vqmodel.encoder.mid.attn_1.v.weight": "model-00006-of-00006.safetensors",
695
+ "model.vqmodel.encoder.mid.block_1.conv1.bias": "model-00006-of-00006.safetensors",
696
+ "model.vqmodel.encoder.mid.block_1.conv1.weight": "model-00006-of-00006.safetensors",
697
+ "model.vqmodel.encoder.mid.block_1.conv2.bias": "model-00006-of-00006.safetensors",
698
+ "model.vqmodel.encoder.mid.block_1.conv2.weight": "model-00006-of-00006.safetensors",
699
+ "model.vqmodel.encoder.mid.block_1.norm1.bias": "model-00006-of-00006.safetensors",
700
+ "model.vqmodel.encoder.mid.block_1.norm1.weight": "model-00006-of-00006.safetensors",
701
+ "model.vqmodel.encoder.mid.block_1.norm2.bias": "model-00006-of-00006.safetensors",
702
+ "model.vqmodel.encoder.mid.block_1.norm2.weight": "model-00006-of-00006.safetensors",
703
+ "model.vqmodel.encoder.mid.block_2.conv1.bias": "model-00006-of-00006.safetensors",
704
+ "model.vqmodel.encoder.mid.block_2.conv1.weight": "model-00006-of-00006.safetensors",
705
+ "model.vqmodel.encoder.mid.block_2.conv2.bias": "model-00006-of-00006.safetensors",
706
+ "model.vqmodel.encoder.mid.block_2.conv2.weight": "model-00006-of-00006.safetensors",
707
+ "model.vqmodel.encoder.mid.block_2.norm1.bias": "model-00006-of-00006.safetensors",
708
+ "model.vqmodel.encoder.mid.block_2.norm1.weight": "model-00006-of-00006.safetensors",
709
+ "model.vqmodel.encoder.mid.block_2.norm2.bias": "model-00006-of-00006.safetensors",
710
+ "model.vqmodel.encoder.mid.block_2.norm2.weight": "model-00006-of-00006.safetensors",
711
+ "model.vqmodel.encoder.norm_out.bias": "model-00006-of-00006.safetensors",
712
+ "model.vqmodel.encoder.norm_out.weight": "model-00006-of-00006.safetensors",
713
+ "model.vqmodel.post_quant_conv.bias": "model-00006-of-00006.safetensors",
714
+ "model.vqmodel.post_quant_conv.weight": "model-00006-of-00006.safetensors",
715
+ "model.vqmodel.quant_conv.bias": "model-00006-of-00006.safetensors",
716
+ "model.vqmodel.quant_conv.weight": "model-00006-of-00006.safetensors",
717
+ "model.vqmodel.quantize.embedding.weight": "model-00006-of-00006.safetensors"
718
+ }
719
+ }
preprocessor_config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": {
3
+ "height": 512,
4
+ "width": 512
5
+ },
6
+ "do_center_crop": true,
7
+ "do_convert_rgb": true,
8
+ "do_normalize": true,
9
+ "do_rescale": true,
10
+ "do_resize": true,
11
+ "image_mean": [
12
+ 1.0,
13
+ 1.0,
14
+ 1.0
15
+ ],
16
+ "image_processor_type": "ChameleonImageProcessor",
17
+ "image_std": [
18
+ 1.0,
19
+ 1.0,
20
+ 1.0
21
+ ],
22
+ "processor_class": "ChameleonProcessor",
23
+ "resample": 1,
24
+ "rescale_factor": 0.0078,
25
+ "size": {
26
+ "shortest_edge": 512
27
+ }
28
+ }
processor_config.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "image_seq_length": 1024,
3
+ "image_token": "<image>",
4
+ "processor_class": "ChameleonProcessor"
5
+ }
script.sh ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/usr/bin/env bash
2
+ #
3
+ # Copyright 2024 PKU-Alignment Team. All Rights Reserved.
4
+ #
5
+ # Licensed under the Apache License, Version 2.0 (the "License");
6
+ # you may not use this file except in compliance with the License.
7
+ # You may obtain a copy of the License at
8
+ #
9
+ # http://www.apache.org/licenses/LICENSE-2.0
10
+ #
11
+ # Unless required by applicable law or agreed to in writing, software
12
+ # distributed under the License is distributed on an "AS IS" BASIS,
13
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14
+ # See the License for the specific language governing permissions and
15
+ # limitations under the License.
16
+ # ==============================================================================
17
+
18
+ export TRITON_CACHE_DIR="/home/align-anything/cache/triton"
19
+ # You can replace it with a local model path
20
+ MODEL_NAME_OR_PATH="/data/align-anything/hantao/models/0916_ti_to_ti_sft/"
21
+ # You can replace it with a local dataset path
22
+ TRAIN_DATASETS="/data/align-anything/hantao/align-anything/projects/text_image_to_text_image/outputs"
23
+ TRAIN_DATA_FILES="llf_ti2ti_27k_tokenized.pt"
24
+ # You can replace it with a new path
25
+ OUTPUT_DIR="../outputs/sft_tf_cham_1011_27k_ti2ti"
26
+ # For wandb online logging
27
+ export WANDB_API_KEY="7e2dcc0c310ebcb7cdcafd5e9320d6be55cf1a33"
28
+ # Source the setup script
29
+ source ./setup.sh
30
+
31
+ # sleep 30m
32
+ # Execute deepspeed command
33
+ deepspeed \
34
+ --master_port ${MASTER_PORT} \
35
+ --module align_anything.trainers.text_image_to_text_image.sft \
36
+ --model_name_or_path ${MODEL_NAME_OR_PATH} \
37
+ --train_datasets ${TRAIN_DATASETS} \
38
+ --train_data_files ${TRAIN_DATA_FILES} \
39
+ --output_dir ${OUTPUT_DIR} \
40
+ --train_template AA_textfeedback \
41
+ --train_split train \
42
+ --per_device_train_batch_size 4 \
43
+ --per_device_eval_batch_size 4 \
44
+ --gradient_accumulation_steps 2 \
45
+ --save_interval 400 \
46
+ --learning_rate 1e-6 \
47
+ --epochs 3 \
48
+ --lr_scheduler_type cosine
special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<pad>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "<reserved08706>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "<unk>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
wandb/debug-internal.log ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-10-11T20:27:29.314392342Z","level":"INFO","msg":"using version","core version":"0.18.3"}
2
+ {"time":"2024-10-11T20:27:29.314422946Z","level":"INFO","msg":"created symlink","path":"../outputs/sft_tf_cham_1011_27k_ti2ti/wandb/run-20241011_202729-hvajn2pl/logs/debug-core.log"}
3
+ {"time":"2024-10-11T20:27:31.315941153Z","level":"ERROR","msg":"dialing: google: could not find default credentials. See https://cloud.google.com/docs/authentication/external/set-up-adc for more information"}
4
+ {"time":"2024-10-11T20:27:31.343923384Z","level":"INFO","msg":"created new stream","id":"hvajn2pl"}
5
+ {"time":"2024-10-11T20:27:31.344015309Z","level":"INFO","msg":"stream: started","id":"hvajn2pl"}
6
+ {"time":"2024-10-11T20:27:31.344030708Z","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"hvajn2pl"}}
7
+ {"time":"2024-10-11T20:27:31.3440543Z","level":"INFO","msg":"handler: started","stream_id":{"value":"hvajn2pl"}}
8
+ {"time":"2024-10-11T20:27:31.344075012Z","level":"INFO","msg":"sender: started","stream_id":{"value":"hvajn2pl"}}
9
+ {"time":"2024-10-11T20:27:32.111921422Z","level":"INFO","msg":"wandb-core","!BADKEY":null}
10
+ {"time":"2024-10-11T20:27:32.115787579Z","level":"INFO","msg":"Starting system monitor"}
11
+ {"time":"2024-10-11T20:37:17.77464423Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": context deadline exceeded"}
12
+ {"time":"2024-10-11T21:02:18.014643124Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)"}
13
+ {"time":"2024-10-11T21:48:05.805217442Z","level":"INFO","msg":"api: retrying HTTP error","status":502,"url":"https://api.wandb.ai/files/htlou/align-anything/hvajn2pl/file_stream"}
14
+ {"time":"2024-10-11T22:52:56.823831464Z","level":"INFO","msg":"api: retrying HTTP error","status":502,"url":"https://api.wandb.ai/files/htlou/align-anything/hvajn2pl/file_stream"}
15
+ {"time":"2024-10-12T04:14:39.094813754Z","level":"INFO","msg":"Stopping system monitor"}
16
+ {"time":"2024-10-12T04:14:39.115766983Z","level":"INFO","msg":"Stopped system monitor"}
17
+ {"time":"2024-10-12T04:14:39.625373664Z","level":"WARN","msg":"No program path found, not creating job artifact. See https://docs.wandb.ai/guides/launch/create-job"}
18
+ {"time":"2024-10-12T04:14:39.625411287Z","level":"INFO","msg":"sender: sendDefer: no job artifact to save"}
19
+ {"time":"2024-10-12T04:14:44.099325031Z","level":"ERROR","msg":"monitor: gpu: timeout waiting for process to exit"}
20
+ {"time":"2024-10-12T04:15:06.376494069Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
21
+ {"time":"2024-10-12T04:15:08.082335235Z","level":"INFO","msg":"stream: closing","id":"hvajn2pl"}
22
+ {"time":"2024-10-12T04:15:08.082392093Z","level":"INFO","msg":"handler: closed","stream_id":{"value":"hvajn2pl"}}
23
+ {"time":"2024-10-12T04:15:08.082456448Z","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"hvajn2pl"}}
24
+ {"time":"2024-10-12T04:15:08.082510885Z","level":"INFO","msg":"sender: closed","stream_id":{"value":"hvajn2pl"}}
25
+ {"time":"2024-10-12T04:15:08.084798858Z","level":"INFO","msg":"stream: closed","id":"hvajn2pl"}
wandb/debug.log ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-10-11 20:27:29,297 INFO MainThread:320740 [wandb_setup.py:_flush():79] Current SDK version is 0.18.3
2
+ 2024-10-11 20:27:29,297 INFO MainThread:320740 [wandb_setup.py:_flush():79] Configure stats pid to 320740
3
+ 2024-10-11 20:27:29,297 INFO MainThread:320740 [wandb_setup.py:_flush():79] Loading settings from /home/align-anything/.config/wandb/settings
4
+ 2024-10-11 20:27:29,297 INFO MainThread:320740 [wandb_setup.py:_flush():79] Loading settings from /data/align-anything/hantao/align-anything/scripts/wandb/settings
5
+ 2024-10-11 20:27:29,297 INFO MainThread:320740 [wandb_setup.py:_flush():79] Loading settings from environment variables: {'api_key': '***REDACTED***'}
6
+ 2024-10-11 20:27:29,297 INFO MainThread:320740 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-10-11 20:27:29,297 WARNING MainThread:320740 [wandb_setup.py:_flush():79] Could not find program at -m align_anything.trainers.text_image_to_text_image.sft
8
+ 2024-10-11 20:27:29,297 INFO MainThread:320740 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m align_anything.trainers.text_image_to_text_image.sft'}
9
+ 2024-10-11 20:27:29,297 INFO MainThread:320740 [wandb_setup.py:_flush():79] Applying login settings: {}
10
+ 2024-10-11 20:27:29,298 INFO MainThread:320740 [wandb_init.py:_log_setup():532] Logging user logs to ../outputs/sft_tf_cham_1011_27k_ti2ti/wandb/run-20241011_202729-hvajn2pl/logs/debug.log
11
+ 2024-10-11 20:27:29,298 INFO MainThread:320740 [wandb_init.py:_log_setup():533] Logging internal logs to ../outputs/sft_tf_cham_1011_27k_ti2ti/wandb/run-20241011_202729-hvajn2pl/logs/debug-internal.log
12
+ 2024-10-11 20:27:29,298 INFO MainThread:320740 [wandb_init.py:init():617] calling init triggers
13
+ 2024-10-11 20:27:29,298 INFO MainThread:320740 [wandb_init.py:init():624] wandb.init called with sweep_config: {}
14
+ config: {'train_cfgs': {'ds_cfgs': 'ds_z3_config.json', 'epochs': 3.0, 'seed': 42, 'per_device_train_batch_size': 4.0, 'per_device_eval_batch_size': 4.0, 'gradient_accumulation_steps': 2.0, 'gradient_checkpointing': True, 'learning_rate': 1e-06, 'lr_scheduler_type': 'cosine', 'lr_warmup_ratio': 0.03, 'weight_decay': 0.0, 'adam_betas': [0.9, 0.95], 'adam_epsilon': 1e-08, 'bf16': True, 'fp16': False, 'eval_strategy': 'steps', 'eval_interval': 1000, 'freeze_language_model': False, 'max_grad_norm': 1.0}, 'data_cfgs': {'train_datasets': '/data/align-anything/hantao/align-anything/projects/text_image_to_text_image/outputs', 'train_template': 'AA_textfeedback', 'train_size': None, 'train_split': 'train', 'train_subset': None, 'train_data_files': 'llf_ti2ti_27k_tokenized.pt', 'train_optional_args': [], 'eval_datasets': None, 'eval_template': None, 'eval_size': None, 'eval_split': None, 'eval_subset': None, 'eval_data_files': None, 'eval_optional_args': []}, 'logger_cfgs': {'log_type': 'wandb', 'log_project': 'align-anything', 'log_run_name': 'sft', 'output_dir': '../outputs/sft_tf_cham_1011_27k_ti2ti', 'cache_dir': None, 'save_interval': 400.0}, 'model_cfgs': {'model_name_or_path': '/data/align-anything/hantao/models/0916_ti_to_ti_sft/', 'trust_remote_code': True, 'model_max_length': 4096}, 'special_tokens': None}
15
+ 2024-10-11 20:27:29,298 INFO MainThread:320740 [wandb_init.py:init():667] starting backend
16
+ 2024-10-11 20:27:29,298 INFO MainThread:320740 [wandb_init.py:init():671] sending inform_init request
17
+ 2024-10-11 20:27:29,308 INFO MainThread:320740 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-10-11 20:27:29,308 INFO MainThread:320740 [wandb_init.py:init():684] backend started and connected
19
+ 2024-10-11 20:27:29,318 INFO MainThread:320740 [wandb_init.py:init():779] updated telemetry
20
+ 2024-10-11 20:27:29,397 INFO MainThread:320740 [wandb_init.py:init():812] communicating run to backend with 90.0 second timeout
21
+ 2024-10-11 20:27:32,099 INFO MainThread:320740 [wandb_init.py:init():863] starting run threads in backend
22
+ 2024-10-11 20:27:32,709 INFO MainThread:320740 [wandb_run.py:_console_start():2465] atexit reg
23
+ 2024-10-11 20:27:32,710 INFO MainThread:320740 [wandb_run.py:_redirect():2313] redirect: wrap_raw
24
+ 2024-10-11 20:27:32,710 INFO MainThread:320740 [wandb_run.py:_redirect():2378] Wrapping output streams.
25
+ 2024-10-11 20:27:32,710 INFO MainThread:320740 [wandb_run.py:_redirect():2403] Redirects installed.
26
+ 2024-10-11 20:27:32,722 INFO MainThread:320740 [wandb_init.py:init():907] run started, returning control to user process
27
+ 2024-10-12 04:14:39,088 INFO MainThread:320740 [wandb_run.py:_finish():2164] finishing run htlou/align-anything/hvajn2pl
28
+ 2024-10-12 04:14:39,090 INFO MainThread:320740 [wandb_run.py:_atexit_cleanup():2428] got exitcode: 0
29
+ 2024-10-12 04:14:39,093 INFO MainThread:320740 [wandb_run.py:_restore():2410] restore
30
+ 2024-10-12 04:14:39,093 INFO MainThread:320740 [wandb_run.py:_restore():2416] restore done
31
+ 2024-10-12 04:15:08,063 INFO MainThread:320740 [wandb_run.py:_footer_history_summary_info():4049] rendering history
32
+ 2024-10-12 04:15:08,065 INFO MainThread:320740 [wandb_run.py:_footer_history_summary_info():4081] rendering summary
33
+ 2024-10-12 04:15:08,077 INFO MainThread:320740 [wandb_run.py:_footer_sync_info():4008] logging synced files
wandb/run-20241011_202729-hvajn2pl/files/config.yaml ADDED
@@ -0,0 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.18.3
4
+ m: []
5
+ python_version: 3.11.0
6
+ t:
7
+ "1":
8
+ - 1
9
+ - 11
10
+ - 41
11
+ - 49
12
+ - 51
13
+ - 55
14
+ - 71
15
+ - 83
16
+ - 98
17
+ - 105
18
+ "2":
19
+ - 1
20
+ - 11
21
+ - 41
22
+ - 49
23
+ - 51
24
+ - 55
25
+ - 71
26
+ - 83
27
+ - 98
28
+ - 105
29
+ "3":
30
+ - 2
31
+ - 13
32
+ - 16
33
+ - 23
34
+ - 55
35
+ - 61
36
+ "4": 3.11.0
37
+ "5": 0.18.3
38
+ "6": 4.46.0.dev0
39
+ "8":
40
+ - 5
41
+ "12": 0.18.3
42
+ "13": linux-x86_64
43
+ data_cfgs:
44
+ value:
45
+ eval_data_files: null
46
+ eval_datasets: null
47
+ eval_optional_args: []
48
+ eval_size: null
49
+ eval_split: null
50
+ eval_subset: null
51
+ eval_template: null
52
+ train_data_files: llf_ti2ti_27k_tokenized.pt
53
+ train_datasets: /data/align-anything/hantao/align-anything/projects/text_image_to_text_image/outputs
54
+ train_optional_args: []
55
+ train_size: null
56
+ train_split: train
57
+ train_subset: null
58
+ train_template: AA_textfeedback
59
+ logger_cfgs:
60
+ value:
61
+ cache_dir: null
62
+ log_project: align-anything
63
+ log_run_name: sft
64
+ log_type: wandb
65
+ output_dir: ../outputs/sft_tf_cham_1011_27k_ti2ti
66
+ save_interval: 400
67
+ model_cfgs:
68
+ value:
69
+ model_max_length: 4096
70
+ model_name_or_path: /data/align-anything/hantao/models/0916_ti_to_ti_sft/
71
+ trust_remote_code: true
72
+ special_tokens:
73
+ value: null
74
+ train_cfgs:
75
+ value:
76
+ adam_betas:
77
+ - 0.9
78
+ - 0.95
79
+ adam_epsilon: 1e-08
80
+ bf16: true
81
+ ds_cfgs: ds_z3_config.json
82
+ epochs: 3
83
+ eval_interval: 1000
84
+ eval_strategy: steps
85
+ fp16: false
86
+ freeze_language_model: false
87
+ gradient_accumulation_steps: 2
88
+ gradient_checkpointing: true
89
+ learning_rate: 1e-06
90
+ lr_scheduler_type: cosine
91
+ lr_warmup_ratio: 0.03
92
+ max_grad_norm: 1
93
+ per_device_eval_batch_size: 4
94
+ per_device_train_batch_size: 4
95
+ seed: 42
96
+ weight_decay: 0
wandb/run-20241011_202729-hvajn2pl/files/output.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20241011_202729-hvajn2pl/files/requirements.txt ADDED
@@ -0,0 +1,234 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ align-anything==0.0.1.dev0
2
+ pluggy==1.5.0
3
+ yt-dlp==2024.8.6
4
+ referencing==0.35.1
5
+ nvitop==1.3.2
6
+ shellingham==1.5.4
7
+ pydantic==2.9.2
8
+ grpcio==1.66.2
9
+ packaging==24.1
10
+ fastapi==0.115.0
11
+ matplotlib==3.9.2
12
+ typing_extensions==4.12.2
13
+ pycountry==24.6.1
14
+ rpds-py==0.20.0
15
+ prometheus_client==0.21.0
16
+ vllm==0.6.2
17
+ pycryptodomex==3.21.0
18
+ MarkupSafe==2.1.5
19
+ prometheus-fastapi-instrumentator==7.0.0
20
+ peft==0.13.1
21
+ Brotli==1.1.0
22
+ audioread==3.0.1
23
+ numpy==1.26.4
24
+ nest-asyncio==1.6.0
25
+ nvidia-nvtx-cu12==12.1.105
26
+ kiwisolver==1.4.7
27
+ librosa==0.10.2.post1
28
+ msgspec==0.18.6
29
+ lark==1.2.2
30
+ aiosignal==1.3.1
31
+ args==0.1.0
32
+ attrs==24.2.0
33
+ regex==2024.9.11
34
+ typer==0.12.5
35
+ nvidia-ml-py==12.535.161
36
+ httptools==0.6.1
37
+ wheel==0.44.0
38
+ pyairports==2.1.1
39
+ threadpoolctl==3.5.0
40
+ nvidia-curand-cu12==10.3.2.106
41
+ yarl==1.14.0
42
+ rich==13.9.2
43
+ nvidia-nccl-cu12==2.20.5
44
+ click==8.1.7
45
+ importlib_metadata==8.5.0
46
+ fsspec==2024.6.1
47
+ outlines_core==0.1.0
48
+ h11==0.14.0
49
+ uvloop==0.20.0
50
+ deepspeed==0.15.2
51
+ absl-py==2.1.0
52
+ setproctitle==1.3.3
53
+ Werkzeug==3.0.4
54
+ nvidia-cuda-cupti-cu12==12.1.105
55
+ fairscale==0.4.13
56
+ wandb==0.18.3
57
+ protobuf==3.20.3
58
+ pytz==2024.2
59
+ sentry-sdk==2.16.0
60
+ opencv-python==4.6.0.66
61
+ cfgv==3.4.0
62
+ diskcache==5.6.3
63
+ psutil==6.0.0
64
+ frechet-audio-distance==0.1.2
65
+ zipp==3.20.2
66
+ pytorch-fid==0.3.0
67
+ nodeenv==1.9.1
68
+ align-anything==0.0.1.dev0
69
+ joblib==1.4.2
70
+ jiter==0.6.1
71
+ interegular==0.3.3
72
+ jsonschema==4.23.0
73
+ numba==0.60.0
74
+ nvidia-cufft-cu12==11.0.2.54
75
+ idna==3.10
76
+ sympy==1.13.3
77
+ proglog==0.1.10
78
+ httpx==0.27.2
79
+ gradio==5.0.1
80
+ lazy_loader==0.4
81
+ pip==24.2
82
+ aiohappyeyeballs==2.4.3
83
+ aiofiles==23.2.1
84
+ contourpy==1.3.0
85
+ mutagen==1.47.0
86
+ watchfiles==0.24.0
87
+ imageio-ffmpeg==0.5.1
88
+ termcolor==2.5.0
89
+ mmsg==0.1.dev20+g585c63a
90
+ imageio==2.35.1
91
+ gguf==0.10.0
92
+ hpsv2==1.2.0
93
+ pydantic_core==2.23.4
94
+ ruff==0.6.9
95
+ markdown-it-py==3.0.0
96
+ torchlibrosa==0.1.0
97
+ cycler==0.12.1
98
+ tzdata==2024.2
99
+ httpcore==1.0.6
100
+ Markdown==3.7
101
+ platformdirs==4.3.6
102
+ term-image==0.7.2
103
+ transformers==4.46.0.dev0
104
+ GitPython==3.1.43
105
+ outlines==0.1.1.dev4+ga2fd35c
106
+ virtualenv==20.26.6
107
+ propcache==0.2.0
108
+ nvidia-cusparse-cu12==12.1.0.106
109
+ identify==2.6.1
110
+ webdataset==0.2.100
111
+ datasets==3.0.1
112
+ accelerate==1.0.0
113
+ multiprocess==0.70.16
114
+ wcwidth==0.2.13
115
+ pre_commit==4.0.1
116
+ tomlkit==0.12.0
117
+ torchvision==0.19.0
118
+ websockets==12.0
119
+ pytest==7.2.0
120
+ timm==0.6.13
121
+ six==1.16.0
122
+ aiohttp==3.10.9
123
+ tiktoken==0.7.0
124
+ ffmpy==0.4.0
125
+ urllib3==2.2.3
126
+ torchaudio==2.4.0
127
+ tensorboard==2.18.0
128
+ scikit-learn==1.5.2
129
+ frozenlist==1.4.1
130
+ multidict==6.1.0
131
+ PyYAML==6.0.2
132
+ xxhash==3.5.0
133
+ torch==2.4.0
134
+ sniffio==1.3.1
135
+ soundfile==0.12.1
136
+ mdurl==0.1.2
137
+ python-multipart==0.0.12
138
+ tensorboard-data-server==0.7.2
139
+ airportsdata==20241001
140
+ nvidia-cudnn-cu12==9.1.0.70
141
+ setuptools-scm==8.1.0
142
+ orjson==3.10.7
143
+ tqdm==4.66.5
144
+ ninja==1.11.1.1
145
+ image-reward==1.5
146
+ clint==0.5.1
147
+ setuptools==75.1.0
148
+ shortuuid==1.0.13
149
+ cffi==1.17.1
150
+ docker-pycreds==0.4.0
151
+ distro==1.9.0
152
+ msgpack==1.1.0
153
+ sentencepiece==0.2.0
154
+ jsonschema-specifications==2024.10.1
155
+ anyio==4.6.0
156
+ optree==0.13.0
157
+ xformers==0.0.27.post2
158
+ pooch==1.8.2
159
+ pyzmq==26.2.0
160
+ Pygments==2.18.0
161
+ partial-json-parser==0.2.1.1.post4
162
+ einops==0.8.0
163
+ nvidia-cublas-cu12==12.1.3.1
164
+ gitdb==4.0.11
165
+ cachetools==5.5.0
166
+ decorator==4.4.2
167
+ uvicorn==0.31.1
168
+ clip==0.2.0
169
+ ftfy==6.2.3
170
+ mistral_common==1.4.4
171
+ safetensors==0.4.5
172
+ charset-normalizer==3.4.0
173
+ soxr==0.5.0.post1
174
+ requests==2.32.3
175
+ pyparsing==3.1.4
176
+ smmap==5.0.1
177
+ annotated-types==0.7.0
178
+ nvidia-cuda-nvrtc-cu12==12.1.105
179
+ cloudpickle==3.0.0
180
+ openai==1.51.2
181
+ ray==2.37.0
182
+ nvidia-nvjitlink-cu12==12.6.77
183
+ resampy==0.4.3
184
+ triton==3.0.0
185
+ huggingface-hub==0.25.2
186
+ pydub==0.25.1
187
+ semantic-version==2.10.0
188
+ distlib==0.3.9
189
+ networkx==3.3
190
+ pandas==2.2.3
191
+ fonttools==4.54.1
192
+ lm-format-enforcer==0.10.6
193
+ python-dotenv==1.0.1
194
+ Jinja2==3.1.4
195
+ nvidia-cuda-runtime-cu12==12.1.105
196
+ iniconfig==2.0.0
197
+ moviepy==1.0.3
198
+ pyarrow==17.0.0
199
+ filelock==3.16.1
200
+ py-cpuinfo==9.0.0
201
+ python-dateutil==2.9.0.post0
202
+ certifi==2024.8.30
203
+ braceexpand==0.1.7
204
+ dill==0.3.8
205
+ bitsandbytes==0.44.1
206
+ starlette==0.38.6
207
+ diffusers==0.30.3
208
+ pillow==10.4.0
209
+ scipy==1.14.1
210
+ pytest-split==0.8.0
211
+ nvidia-cusolver-cu12==11.4.5.107
212
+ llvmlite==0.43.0
213
+ gradio_client==1.4.0
214
+ mpmath==1.3.0
215
+ pycparser==2.22
216
+ hjson==3.1.0
217
+ tokenizers==0.20.1
218
+ jaraco.context==5.3.0
219
+ typeguard==4.3.0
220
+ inflect==7.3.1
221
+ importlib_metadata==8.0.0
222
+ jaraco.collections==5.1.0
223
+ backports.tarfile==1.2.0
224
+ platformdirs==4.2.2
225
+ jaraco.functools==4.0.1
226
+ zipp==3.19.2
227
+ autocommand==2.2.2
228
+ typing_extensions==4.12.2
229
+ jaraco.text==3.12.1
230
+ importlib_resources==6.4.0
231
+ more-itertools==10.3.0
232
+ wheel==0.43.0
233
+ packaging==24.1
234
+ tomli==2.0.1
wandb/run-20241011_202729-hvajn2pl/files/wandb-metadata.json ADDED
@@ -0,0 +1,112 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-144-generic-x86_64-with-glibc2.31",
3
+ "python": "3.11.0",
4
+ "startedAt": "2024-10-11T20:27:29.308691Z",
5
+ "args": [
6
+ "--local_rank=0",
7
+ "--model_name_or_path",
8
+ "/data/align-anything/hantao/models/0916_ti_to_ti_sft/",
9
+ "--train_datasets",
10
+ "/data/align-anything/hantao/align-anything/projects/text_image_to_text_image/outputs",
11
+ "--train_data_files",
12
+ "llf_ti2ti_27k_tokenized.pt",
13
+ "--output_dir",
14
+ "../outputs/sft_tf_cham_1011_27k_ti2ti",
15
+ "--train_template",
16
+ "AA_textfeedback",
17
+ "--train_split",
18
+ "train",
19
+ "--per_device_train_batch_size",
20
+ "4",
21
+ "--per_device_eval_batch_size",
22
+ "4",
23
+ "--gradient_accumulation_steps",
24
+ "2",
25
+ "--save_interval",
26
+ "400",
27
+ "--learning_rate",
28
+ "1e-6",
29
+ "--epochs",
30
+ "3",
31
+ "--lr_scheduler_type",
32
+ "cosine"
33
+ ],
34
+ "program": "-m align_anything.trainers.text_image_to_text_image.sft",
35
+ "git": {
36
+ "remote": "https://github.com/PKU-Alignment/align-anything.git",
37
+ "commit": "6fde660afc9985323f147930eedf188a5699adc7"
38
+ },
39
+ "email": "[email protected]",
40
+ "root": "../outputs/sft_tf_cham_1011_27k_ti2ti",
41
+ "host": "lyg0196",
42
+ "username": "align-anything",
43
+ "executable": "/data/align-anything/miniconda3/envs/hantao_stable/bin/python",
44
+ "cpu_count": 64,
45
+ "cpu_count_logical": 128,
46
+ "gpu": "[NVIDIA A100-SXM4-80GB, NVIDIA A100-SXM4-80GB, NVIDIA A100-SXM4-80GB, NVIDIA A100-SXM4-80GB, NVIDIA A100-SXM4-80GB, NVIDIA A100-SXM4-80GB, NVIDIA A100-SXM4-80GB, NVIDIA A100-SXM4-80GB]",
47
+ "gpu_count": 8,
48
+ "disk": {
49
+ "/": {
50
+ "total": "940744544256",
51
+ "used": "49302532096"
52
+ }
53
+ },
54
+ "memory": {
55
+ "total": "540660944896"
56
+ },
57
+ "cpu": {
58
+ "count": 64,
59
+ "countLogical": 128
60
+ },
61
+ "gpu_nvidia": [
62
+ {
63
+ "name": "NVIDIA A100-SXM4-80GB",
64
+ "memoryTotal": "85899345920",
65
+ "cudaCores": 6912,
66
+ "architecture": "Ampere"
67
+ },
68
+ {
69
+ "name": "NVIDIA A100-SXM4-80GB",
70
+ "memoryTotal": "85899345920",
71
+ "cudaCores": 6912,
72
+ "architecture": "Ampere"
73
+ },
74
+ {
75
+ "name": "NVIDIA A100-SXM4-80GB",
76
+ "memoryTotal": "85899345920",
77
+ "cudaCores": 6912,
78
+ "architecture": "Ampere"
79
+ },
80
+ {
81
+ "name": "NVIDIA A100-SXM4-80GB",
82
+ "memoryTotal": "85899345920",
83
+ "cudaCores": 6912,
84
+ "architecture": "Ampere"
85
+ },
86
+ {
87
+ "name": "NVIDIA A100-SXM4-80GB",
88
+ "memoryTotal": "85899345920",
89
+ "cudaCores": 6912,
90
+ "architecture": "Ampere"
91
+ },
92
+ {
93
+ "name": "NVIDIA A100-SXM4-80GB",
94
+ "memoryTotal": "85899345920",
95
+ "cudaCores": 6912,
96
+ "architecture": "Ampere"
97
+ },
98
+ {
99
+ "name": "NVIDIA A100-SXM4-80GB",
100
+ "memoryTotal": "85899345920",
101
+ "cudaCores": 6912,
102
+ "architecture": "Ampere"
103
+ },
104
+ {
105
+ "name": "NVIDIA A100-SXM4-80GB",
106
+ "memoryTotal": "85899345920",
107
+ "cudaCores": 6912,
108
+ "architecture": "Ampere"
109
+ }
110
+ ],
111
+ "cudaVersion": "12.4"
112
+ }
wandb/run-20241011_202729-hvajn2pl/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb":{"runtime":28029},"train/step":4983,"train/loss":0.3495679199695587,"train/lr":1.6866608503751657e-12,"train/epoch":3,"_timestamp":1.728706315447278e+09,"_runtime":28029.786090375,"_step":4983}
wandb/run-20241011_202729-hvajn2pl/logs/debug-internal.log ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-10-11T20:27:29.314392342Z","level":"INFO","msg":"using version","core version":"0.18.3"}
2
+ {"time":"2024-10-11T20:27:29.314422946Z","level":"INFO","msg":"created symlink","path":"../outputs/sft_tf_cham_1011_27k_ti2ti/wandb/run-20241011_202729-hvajn2pl/logs/debug-core.log"}
3
+ {"time":"2024-10-11T20:27:31.315941153Z","level":"ERROR","msg":"dialing: google: could not find default credentials. See https://cloud.google.com/docs/authentication/external/set-up-adc for more information"}
4
+ {"time":"2024-10-11T20:27:31.343923384Z","level":"INFO","msg":"created new stream","id":"hvajn2pl"}
5
+ {"time":"2024-10-11T20:27:31.344015309Z","level":"INFO","msg":"stream: started","id":"hvajn2pl"}
6
+ {"time":"2024-10-11T20:27:31.344030708Z","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"hvajn2pl"}}
7
+ {"time":"2024-10-11T20:27:31.3440543Z","level":"INFO","msg":"handler: started","stream_id":{"value":"hvajn2pl"}}
8
+ {"time":"2024-10-11T20:27:31.344075012Z","level":"INFO","msg":"sender: started","stream_id":{"value":"hvajn2pl"}}
9
+ {"time":"2024-10-11T20:27:32.111921422Z","level":"INFO","msg":"wandb-core","!BADKEY":null}
10
+ {"time":"2024-10-11T20:27:32.115787579Z","level":"INFO","msg":"Starting system monitor"}
11
+ {"time":"2024-10-11T20:37:17.77464423Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": context deadline exceeded"}
12
+ {"time":"2024-10-11T21:02:18.014643124Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)"}
13
+ {"time":"2024-10-11T21:48:05.805217442Z","level":"INFO","msg":"api: retrying HTTP error","status":502,"url":"https://api.wandb.ai/files/htlou/align-anything/hvajn2pl/file_stream"}
14
+ {"time":"2024-10-11T22:52:56.823831464Z","level":"INFO","msg":"api: retrying HTTP error","status":502,"url":"https://api.wandb.ai/files/htlou/align-anything/hvajn2pl/file_stream"}
15
+ {"time":"2024-10-12T04:14:39.094813754Z","level":"INFO","msg":"Stopping system monitor"}
16
+ {"time":"2024-10-12T04:14:39.115766983Z","level":"INFO","msg":"Stopped system monitor"}
17
+ {"time":"2024-10-12T04:14:39.625373664Z","level":"WARN","msg":"No program path found, not creating job artifact. See https://docs.wandb.ai/guides/launch/create-job"}
18
+ {"time":"2024-10-12T04:14:39.625411287Z","level":"INFO","msg":"sender: sendDefer: no job artifact to save"}
19
+ {"time":"2024-10-12T04:14:44.099325031Z","level":"ERROR","msg":"monitor: gpu: timeout waiting for process to exit"}
20
+ {"time":"2024-10-12T04:15:06.376494069Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
21
+ {"time":"2024-10-12T04:15:08.082335235Z","level":"INFO","msg":"stream: closing","id":"hvajn2pl"}
22
+ {"time":"2024-10-12T04:15:08.082392093Z","level":"INFO","msg":"handler: closed","stream_id":{"value":"hvajn2pl"}}
23
+ {"time":"2024-10-12T04:15:08.082456448Z","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"hvajn2pl"}}
24
+ {"time":"2024-10-12T04:15:08.082510885Z","level":"INFO","msg":"sender: closed","stream_id":{"value":"hvajn2pl"}}
25
+ {"time":"2024-10-12T04:15:08.084798858Z","level":"INFO","msg":"stream: closed","id":"hvajn2pl"}
wandb/run-20241011_202729-hvajn2pl/logs/debug.log ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-10-11 20:27:29,297 INFO MainThread:320740 [wandb_setup.py:_flush():79] Current SDK version is 0.18.3
2
+ 2024-10-11 20:27:29,297 INFO MainThread:320740 [wandb_setup.py:_flush():79] Configure stats pid to 320740
3
+ 2024-10-11 20:27:29,297 INFO MainThread:320740 [wandb_setup.py:_flush():79] Loading settings from /home/align-anything/.config/wandb/settings
4
+ 2024-10-11 20:27:29,297 INFO MainThread:320740 [wandb_setup.py:_flush():79] Loading settings from /data/align-anything/hantao/align-anything/scripts/wandb/settings
5
+ 2024-10-11 20:27:29,297 INFO MainThread:320740 [wandb_setup.py:_flush():79] Loading settings from environment variables: {'api_key': '***REDACTED***'}
6
+ 2024-10-11 20:27:29,297 INFO MainThread:320740 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-10-11 20:27:29,297 WARNING MainThread:320740 [wandb_setup.py:_flush():79] Could not find program at -m align_anything.trainers.text_image_to_text_image.sft
8
+ 2024-10-11 20:27:29,297 INFO MainThread:320740 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m align_anything.trainers.text_image_to_text_image.sft'}
9
+ 2024-10-11 20:27:29,297 INFO MainThread:320740 [wandb_setup.py:_flush():79] Applying login settings: {}
10
+ 2024-10-11 20:27:29,298 INFO MainThread:320740 [wandb_init.py:_log_setup():532] Logging user logs to ../outputs/sft_tf_cham_1011_27k_ti2ti/wandb/run-20241011_202729-hvajn2pl/logs/debug.log
11
+ 2024-10-11 20:27:29,298 INFO MainThread:320740 [wandb_init.py:_log_setup():533] Logging internal logs to ../outputs/sft_tf_cham_1011_27k_ti2ti/wandb/run-20241011_202729-hvajn2pl/logs/debug-internal.log
12
+ 2024-10-11 20:27:29,298 INFO MainThread:320740 [wandb_init.py:init():617] calling init triggers
13
+ 2024-10-11 20:27:29,298 INFO MainThread:320740 [wandb_init.py:init():624] wandb.init called with sweep_config: {}
14
+ config: {'train_cfgs': {'ds_cfgs': 'ds_z3_config.json', 'epochs': 3.0, 'seed': 42, 'per_device_train_batch_size': 4.0, 'per_device_eval_batch_size': 4.0, 'gradient_accumulation_steps': 2.0, 'gradient_checkpointing': True, 'learning_rate': 1e-06, 'lr_scheduler_type': 'cosine', 'lr_warmup_ratio': 0.03, 'weight_decay': 0.0, 'adam_betas': [0.9, 0.95], 'adam_epsilon': 1e-08, 'bf16': True, 'fp16': False, 'eval_strategy': 'steps', 'eval_interval': 1000, 'freeze_language_model': False, 'max_grad_norm': 1.0}, 'data_cfgs': {'train_datasets': '/data/align-anything/hantao/align-anything/projects/text_image_to_text_image/outputs', 'train_template': 'AA_textfeedback', 'train_size': None, 'train_split': 'train', 'train_subset': None, 'train_data_files': 'llf_ti2ti_27k_tokenized.pt', 'train_optional_args': [], 'eval_datasets': None, 'eval_template': None, 'eval_size': None, 'eval_split': None, 'eval_subset': None, 'eval_data_files': None, 'eval_optional_args': []}, 'logger_cfgs': {'log_type': 'wandb', 'log_project': 'align-anything', 'log_run_name': 'sft', 'output_dir': '../outputs/sft_tf_cham_1011_27k_ti2ti', 'cache_dir': None, 'save_interval': 400.0}, 'model_cfgs': {'model_name_or_path': '/data/align-anything/hantao/models/0916_ti_to_ti_sft/', 'trust_remote_code': True, 'model_max_length': 4096}, 'special_tokens': None}
15
+ 2024-10-11 20:27:29,298 INFO MainThread:320740 [wandb_init.py:init():667] starting backend
16
+ 2024-10-11 20:27:29,298 INFO MainThread:320740 [wandb_init.py:init():671] sending inform_init request
17
+ 2024-10-11 20:27:29,308 INFO MainThread:320740 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-10-11 20:27:29,308 INFO MainThread:320740 [wandb_init.py:init():684] backend started and connected
19
+ 2024-10-11 20:27:29,318 INFO MainThread:320740 [wandb_init.py:init():779] updated telemetry
20
+ 2024-10-11 20:27:29,397 INFO MainThread:320740 [wandb_init.py:init():812] communicating run to backend with 90.0 second timeout
21
+ 2024-10-11 20:27:32,099 INFO MainThread:320740 [wandb_init.py:init():863] starting run threads in backend
22
+ 2024-10-11 20:27:32,709 INFO MainThread:320740 [wandb_run.py:_console_start():2465] atexit reg
23
+ 2024-10-11 20:27:32,710 INFO MainThread:320740 [wandb_run.py:_redirect():2313] redirect: wrap_raw
24
+ 2024-10-11 20:27:32,710 INFO MainThread:320740 [wandb_run.py:_redirect():2378] Wrapping output streams.
25
+ 2024-10-11 20:27:32,710 INFO MainThread:320740 [wandb_run.py:_redirect():2403] Redirects installed.
26
+ 2024-10-11 20:27:32,722 INFO MainThread:320740 [wandb_init.py:init():907] run started, returning control to user process
27
+ 2024-10-12 04:14:39,088 INFO MainThread:320740 [wandb_run.py:_finish():2164] finishing run htlou/align-anything/hvajn2pl
28
+ 2024-10-12 04:14:39,090 INFO MainThread:320740 [wandb_run.py:_atexit_cleanup():2428] got exitcode: 0
29
+ 2024-10-12 04:14:39,093 INFO MainThread:320740 [wandb_run.py:_restore():2410] restore
30
+ 2024-10-12 04:14:39,093 INFO MainThread:320740 [wandb_run.py:_restore():2416] restore done
31
+ 2024-10-12 04:15:08,063 INFO MainThread:320740 [wandb_run.py:_footer_history_summary_info():4049] rendering history
32
+ 2024-10-12 04:15:08,065 INFO MainThread:320740 [wandb_run.py:_footer_history_summary_info():4081] rendering summary
33
+ 2024-10-12 04:15:08,077 INFO MainThread:320740 [wandb_run.py:_footer_sync_info():4008] logging synced files
wandb/run-20241011_202729-hvajn2pl/run-hvajn2pl.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9158a2c0d338475b6a5747499dd2f57d32151d5602d5f0f12161f9c7374c574a
3
+ size 30879748