Ivan Chelombiev
commited on
Commit
·
a36365b
1
Parent(s):
b2c44fb
Model save
Browse files- config.json +4 -3
- training_args.bin +1 -1
config.json
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
-
"
|
5 |
],
|
6 |
"attention_probs_dropout_prob": 0.0,
|
7 |
"classifier_dropout": null,
|
@@ -15,8 +15,9 @@
|
|
15 |
"initializer_range": 0.02,
|
16 |
"intermediate_size": 3072,
|
17 |
"layer_norm_eps": 0.001,
|
|
|
18 |
"max_position_embeddings": 512,
|
19 |
-
"model_type": "
|
20 |
"num_attention_heads": 12,
|
21 |
"num_hidden_layers": 12,
|
22 |
"pad_token_id": 0,
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "run9-phase2",
|
3 |
"architectures": [
|
4 |
+
"PipelinedGroupBertForPreTraining"
|
5 |
],
|
6 |
"attention_probs_dropout_prob": 0.0,
|
7 |
"classifier_dropout": null,
|
|
|
15 |
"initializer_range": 0.02,
|
16 |
"intermediate_size": 3072,
|
17 |
"layer_norm_eps": 0.001,
|
18 |
+
"max_num_masked_tokens": 96,
|
19 |
"max_position_embeddings": 512,
|
20 |
+
"model_type": "groupbert",
|
21 |
"num_attention_heads": 12,
|
22 |
"num_hidden_layers": 12,
|
23 |
"pad_token_id": 0,
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 2927
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5e9dacdcba4a49066c7b62de11e237b5e1160ec5c2b1b8e090e1416d50d6a051
|
3 |
size 2927
|