g8a9 commited on
Commit
eb7cfd1
·
1 Parent(s): 4a62011

Training in progress, step 10

Browse files
config.json CHANGED
@@ -35,7 +35,7 @@
35
  }
36
  },
37
  "torch_dtype": "float32",
38
- "transformers_version": "4.27.4",
39
  "use_cache": true,
40
  "vocab_size": 50259
41
  }
 
35
  }
36
  },
37
  "torch_dtype": "float32",
38
+ "transformers_version": "4.28.0",
39
  "use_cache": true,
40
  "vocab_size": 50259
41
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c2a22eeadb8ace6405fba562d163c0e85561161ca62ab7fba2af545504dd785b
3
  size 1444575065
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19e85bb3e941147e4f318301b98a18e3087db51ff22001e3f3baa64dfeb63d2e
3
  size 1444575065
tokenizer_config.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "add_prefix_space": false,
3
  "bos_token": "<|endoftext|>",
 
4
  "eos_token": "<|endoftext|>",
5
  "model_max_length": 1024,
6
- "special_tokens_map_file": null,
7
  "tokenizer_class": "GPT2Tokenizer",
8
  "unk_token": "<|endoftext|>"
9
  }
 
1
  {
2
  "add_prefix_space": false,
3
  "bos_token": "<|endoftext|>",
4
+ "clean_up_tokenization_spaces": true,
5
  "eos_token": "<|endoftext|>",
6
  "model_max_length": 1024,
 
7
  "tokenizer_class": "GPT2Tokenizer",
8
  "unk_token": "<|endoftext|>"
9
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c4beaf22f03249e85340bc9c7028ded9ceb77fbd23dd1fb5782eca8c6b19abba
3
- size 3631
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c30c8b48a9ef3ad78ceb74f8fe21b907fb41abd675c5a65a1c4d03dabf696807
3
+ size 3695