Bert-base-cased Fine Tuned Glue Mrpc Demo
This checkpoint was initialized from the pre-trained checkpoint bert-base-cased and subsequently fine-tuned on GLUE task: mrpc using this notebook. Training was conducted for 3 epochs, using a linear decaying learning rate of 2e-05, and a total batch size of 32.
The model has a final training loss of 0.103 and a accuracy of 0.831.
- Downloads last month
- 11
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.