cartesinus's picture
update model card README.md
30aa00b
|
raw
history blame
2.08 kB
metadata
license: mit
tags:
  - generated_from_trainer
metrics:
  - bleu
model-index:
  - name: iva_mt_wslot-m2m100_418M-0.1.0
    results: []

iva_mt_wslot-m2m100_418M-0.1.0

This model is a fine-tuned version of facebook/m2m100_418M on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.0176
  • Bleu: 61.6249
  • Gen Len: 21.157

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 4
  • eval_batch_size: 4
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss Bleu Gen Len
0.0178 1.0 5091 0.0171 57.4439 21.1396
0.013 2.0 10182 0.0159 58.886 21.2285
0.0091 3.0 15273 0.0157 60.159 21.1222
0.0073 4.0 20364 0.0159 60.5893 21.1212
0.0054 5.0 25455 0.0161 60.6484 21.0679
0.004 6.0 30546 0.0166 61.5283 21.0875
0.0031 7.0 35637 0.0169 61.0439 21.1562
0.0024 8.0 40728 0.0172 61.9427 21.2203
0.0018 9.0 45819 0.0175 61.7325 21.1478
0.0014 10.0 50910 0.0176 61.6249 21.157

Framework versions

  • Transformers 4.26.1
  • Pytorch 1.13.1+cu116
  • Datasets 2.10.1
  • Tokenizers 0.13.2