Training in progress, epoch 1
Browse files- config.json +21 -0
- model.safetensors +3 -0
- runs/Mar09_16-55-13_torch-flash-large222-0-2/events.out.tfevents.1741539314.torch-flash-large222-0-2.334.0 +3 -0
- runs/Mar09_16-56-59_torch-flash-large222-0-2/events.out.tfevents.1741539419.torch-flash-large222-0-2.546.0 +3 -0
- runs/Mar09_16-58-05_torch-flash-large222-0-2/events.out.tfevents.1741539486.torch-flash-large222-0-2.788.0 +3 -0
- runs/Mar09_16-58-25_torch-flash-large222-0-2/events.out.tfevents.1741539506.torch-flash-large222-0-2.999.0 +3 -0
- runs/Mar09_16-58-53_torch-flash-large222-0-2/events.out.tfevents.1741539533.torch-flash-large222-0-2.1210.0 +3 -0
- training_args.bin +3 -0
config.json
ADDED
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "tattabio/gLM2_650M",
|
3 |
+
"architectures": [
|
4 |
+
"gLM2ForMaskedLM"
|
5 |
+
],
|
6 |
+
"auto_map": {
|
7 |
+
"AutoConfig": "configuration_glm2.gLM2Config",
|
8 |
+
"AutoModel": "modeling_glm2.gLM2Model",
|
9 |
+
"AutoModelForMaskedLM": "modeling_glm2.gLM2ForMaskedLM"
|
10 |
+
},
|
11 |
+
"depth": 33,
|
12 |
+
"dim": 1280,
|
13 |
+
"ffn_dim_multiplier": null,
|
14 |
+
"heads": 20,
|
15 |
+
"model_type": "gLM2",
|
16 |
+
"norm_eps": 1e-05,
|
17 |
+
"swiglu_multiple_of": 256,
|
18 |
+
"torch_dtype": "float32",
|
19 |
+
"transformers_version": "4.49.0",
|
20 |
+
"vocab_size": 37
|
21 |
+
}
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e9710ceb6e6839dcee08d16f242c5bba9ec7262d7b88c7e640aceb3b663ff6f3
|
3 |
+
size 2682482800
|
runs/Mar09_16-55-13_torch-flash-large222-0-2/events.out.tfevents.1741539314.torch-flash-large222-0-2.334.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4ce16ca9da70ffbd166ebe0e80cc0c99bd3330e823886e77cd8fcc610a7858d2
|
3 |
+
size 4236
|
runs/Mar09_16-56-59_torch-flash-large222-0-2/events.out.tfevents.1741539419.torch-flash-large222-0-2.546.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5178ceb2319fe53b4d5eef728b598ae64f81429473f07a9c757e06ad190842a0
|
3 |
+
size 4869
|
runs/Mar09_16-58-05_torch-flash-large222-0-2/events.out.tfevents.1741539486.torch-flash-large222-0-2.788.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4038020e82904edb2d532111b35f18935f93b2166906c562177ab416dad6bc78
|
3 |
+
size 4869
|
runs/Mar09_16-58-25_torch-flash-large222-0-2/events.out.tfevents.1741539506.torch-flash-large222-0-2.999.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d4d6bf66047a6dc983530cb9cab1dd553f90df79570581f63e4009c917915dc7
|
3 |
+
size 4869
|
runs/Mar09_16-58-53_torch-flash-large222-0-2/events.out.tfevents.1741539533.torch-flash-large222-0-2.1210.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:77b8b25650d2ff1db824e7d83cd00dbeb32e3b473bbe05ad749a2465567c8f2c
|
3 |
+
size 5349
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:145c2fa7d2449400b8b78f9c3bc76f59c339909b1e11ef066d215b797c8472c8
|
3 |
+
size 5368
|