Quazim0t0 commited on
Commit
567be63
·
verified ·
1 Parent(s): 6129279

(Trained with Unsloth)

Browse files
Files changed (2) hide show
  1. config.json +2 -2
  2. generation_config.json +1 -1
config.json CHANGED
@@ -60,8 +60,8 @@
60
  "rope_theta": 1000000,
61
  "sliding_window": null,
62
  "tie_word_embeddings": true,
63
- "torch_dtype": "float16",
64
- "transformers_version": "4.53.2",
65
  "unsloth_fixed": true,
66
  "unsloth_version": "2025.7.8",
67
  "use_cache": true,
 
60
  "rope_theta": 1000000,
61
  "sliding_window": null,
62
  "tie_word_embeddings": true,
63
+ "torch_dtype": "bfloat16",
64
+ "transformers_version": "4.53.3",
65
  "unsloth_fixed": true,
66
  "unsloth_version": "2025.7.8",
67
  "use_cache": true,
generation_config.json CHANGED
@@ -3,5 +3,5 @@
3
  "eos_token_id": 151643,
4
  "max_length": 32768,
5
  "pad_token_id": 0,
6
- "transformers_version": "4.53.2"
7
  }
 
3
  "eos_token_id": 151643,
4
  "max_length": 32768,
5
  "pad_token_id": 0,
6
+ "transformers_version": "4.53.3"
7
  }