Built with Axolotl

bda794e5-89a6-4eaa-862e-e688558b05e4

This model is a fine-tuned version of fxmarty/really-tiny-falcon-testing on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 10.9657

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.000211
  • train_batch_size: 4
  • eval_batch_size: 4
  • seed: 42
  • gradient_accumulation_steps: 2
  • total_train_batch_size: 8
  • optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_steps: 50
  • training_steps: 452

Training results

Training Loss Epoch Step Validation Loss
No log 0.0022 1 11.0913
22.0436 0.1107 50 11.0186
21.9701 0.2215 100 10.9933
21.9368 0.3322 150 10.9824
21.971 0.4430 200 10.9762
21.9452 0.5537 250 10.9720
21.9155 0.6645 300 10.9685
21.931 0.7752 350 10.9672
21.9188 0.8859 400 10.9662
21.9059 0.9967 450 10.9657

Framework versions

  • PEFT 0.13.2
  • Transformers 4.46.0
  • Pytorch 2.5.0+cu124
  • Datasets 3.0.1
  • Tokenizers 0.20.1
Downloads last month
0
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no pipeline_tag.

Model tree for lesso11/bda794e5-89a6-4eaa-862e-e688558b05e4

Adapter
(262)
this model