The Objective Dad commited on
Commit
83237b8
·
unverified ·
2 Parent(s): 9234b75 168a7a0

Merge branch 'OpenAccess-AI-Collective:main' into logging_enhancement

Browse files
Files changed (2) hide show
  1. README.md +5 -0
  2. src/axolotl/utils/trainer.py +4 -0
README.md CHANGED
@@ -305,6 +305,8 @@ base_model_ignore_patterns:
305
  # if the base_model repo on hf hub doesn't include configuration .json files,
306
  # you can set that here, or leave this empty to default to base_model
307
  base_model_config: ./llama-7b-hf
 
 
308
  # Optional tokenizer configuration override in case you want to use a different tokenizer
309
  # than the one defined in the base model
310
  tokenizer_config:
@@ -411,6 +413,9 @@ logging_steps:
411
  save_steps:
412
  eval_steps:
413
 
 
 
 
414
  # whether to mask out or include the human's prompt from the training labels
415
  train_on_inputs: false
416
  # don't use this, leads to wonky training (according to someone on the internet)
 
305
  # if the base_model repo on hf hub doesn't include configuration .json files,
306
  # you can set that here, or leave this empty to default to base_model
307
  base_model_config: ./llama-7b-hf
308
+ # you can specify to choose a specific model revision from huggingface hub
309
+ model_revision:
310
  # Optional tokenizer configuration override in case you want to use a different tokenizer
311
  # than the one defined in the base model
312
  tokenizer_config:
 
413
  save_steps:
414
  eval_steps:
415
 
416
+ # save model as safetensors (require safetensors package)
417
+ save_safetensors:
418
+
419
  # whether to mask out or include the human's prompt from the training labels
420
  train_on_inputs: false
421
  # don't use this, leads to wonky training (according to someone on the internet)
src/axolotl/utils/trainer.py CHANGED
@@ -183,6 +183,10 @@ def setup_trainer(cfg, train_dataset, eval_dataset, model, tokenizer):
183
  if cfg.hub_model_id:
184
  training_arguments_kwargs["hub_model_id"] = cfg.hub_model_id
185
  training_arguments_kwargs["push_to_hub"] = True
 
 
 
 
186
 
187
  training_args = AxolotlTrainingArguments(
188
  per_device_train_batch_size=cfg.micro_batch_size,
 
183
  if cfg.hub_model_id:
184
  training_arguments_kwargs["hub_model_id"] = cfg.hub_model_id
185
  training_arguments_kwargs["push_to_hub"] = True
186
+ training_arguments_kwargs["hub_private_repo"] = True
187
+
188
+ if cfg.save_safetensors:
189
+ training_arguments_kwargs["save_safetensors"] = cfg.save_safetensors
190
 
191
  training_args = AxolotlTrainingArguments(
192
  per_device_train_batch_size=cfg.micro_batch_size,