Merge branch 'OpenAccess-AI-Collective:main' into logging_enhancement
Browse files- README.md +5 -0
- src/axolotl/utils/trainer.py +4 -0
README.md
CHANGED
@@ -305,6 +305,8 @@ base_model_ignore_patterns:
|
|
305 |
# if the base_model repo on hf hub doesn't include configuration .json files,
|
306 |
# you can set that here, or leave this empty to default to base_model
|
307 |
base_model_config: ./llama-7b-hf
|
|
|
|
|
308 |
# Optional tokenizer configuration override in case you want to use a different tokenizer
|
309 |
# than the one defined in the base model
|
310 |
tokenizer_config:
|
@@ -411,6 +413,9 @@ logging_steps:
|
|
411 |
save_steps:
|
412 |
eval_steps:
|
413 |
|
|
|
|
|
|
|
414 |
# whether to mask out or include the human's prompt from the training labels
|
415 |
train_on_inputs: false
|
416 |
# don't use this, leads to wonky training (according to someone on the internet)
|
|
|
305 |
# if the base_model repo on hf hub doesn't include configuration .json files,
|
306 |
# you can set that here, or leave this empty to default to base_model
|
307 |
base_model_config: ./llama-7b-hf
|
308 |
+
# you can specify to choose a specific model revision from huggingface hub
|
309 |
+
model_revision:
|
310 |
# Optional tokenizer configuration override in case you want to use a different tokenizer
|
311 |
# than the one defined in the base model
|
312 |
tokenizer_config:
|
|
|
413 |
save_steps:
|
414 |
eval_steps:
|
415 |
|
416 |
+
# save model as safetensors (require safetensors package)
|
417 |
+
save_safetensors:
|
418 |
+
|
419 |
# whether to mask out or include the human's prompt from the training labels
|
420 |
train_on_inputs: false
|
421 |
# don't use this, leads to wonky training (according to someone on the internet)
|
src/axolotl/utils/trainer.py
CHANGED
@@ -183,6 +183,10 @@ def setup_trainer(cfg, train_dataset, eval_dataset, model, tokenizer):
|
|
183 |
if cfg.hub_model_id:
|
184 |
training_arguments_kwargs["hub_model_id"] = cfg.hub_model_id
|
185 |
training_arguments_kwargs["push_to_hub"] = True
|
|
|
|
|
|
|
|
|
186 |
|
187 |
training_args = AxolotlTrainingArguments(
|
188 |
per_device_train_batch_size=cfg.micro_batch_size,
|
|
|
183 |
if cfg.hub_model_id:
|
184 |
training_arguments_kwargs["hub_model_id"] = cfg.hub_model_id
|
185 |
training_arguments_kwargs["push_to_hub"] = True
|
186 |
+
training_arguments_kwargs["hub_private_repo"] = True
|
187 |
+
|
188 |
+
if cfg.save_safetensors:
|
189 |
+
training_arguments_kwargs["save_safetensors"] = cfg.save_safetensors
|
190 |
|
191 |
training_args = AxolotlTrainingArguments(
|
192 |
per_device_train_batch_size=cfg.micro_batch_size,
|