End of training
Browse files- README.md +28 -26
- emissions.csv +1 -1
- metrics.json +6 -6
- model.safetensors +1 -1
README.md
CHANGED
|
@@ -4,6 +4,8 @@ license: mit
|
|
| 4 |
base_model: roberta-base
|
| 5 |
tags:
|
| 6 |
- generated_from_trainer
|
|
|
|
|
|
|
| 7 |
model-index:
|
| 8 |
- name: cwe-parent-vulnerability-classification-roberta-base
|
| 9 |
results: []
|
|
@@ -16,9 +18,9 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 16 |
|
| 17 |
This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on an unknown dataset.
|
| 18 |
It achieves the following results on the evaluation set:
|
| 19 |
-
- Loss: 0.
|
| 20 |
-
-
|
| 21 |
-
-
|
| 22 |
|
| 23 |
## Model description
|
| 24 |
|
|
@@ -47,33 +49,33 @@ The following hyperparameters were used during training:
|
|
| 47 |
|
| 48 |
### Training results
|
| 49 |
|
| 50 |
-
| Training Loss | Epoch | Step | Validation Loss |
|
| 51 |
-
|
| 52 |
-
|
|
| 53 |
-
| 1.
|
| 54 |
-
| 1.
|
| 55 |
-
| 1.
|
| 56 |
-
| 1.
|
| 57 |
-
| 1.
|
| 58 |
-
|
|
| 59 |
-
|
|
| 60 |
-
| 0.
|
| 61 |
-
|
|
| 62 |
-
|
|
| 63 |
-
| 0.
|
| 64 |
-
|
|
| 65 |
-
| 0.
|
| 66 |
-
| 0.
|
| 67 |
-
| 0.
|
| 68 |
-
| 0.
|
| 69 |
-
| 0.
|
| 70 |
-
| 0.
|
| 71 |
-
| 0.
|
| 72 |
|
| 73 |
|
| 74 |
### Framework versions
|
| 75 |
|
| 76 |
-
- Transformers 4.55.
|
| 77 |
- Pytorch 2.7.1+cu126
|
| 78 |
- Datasets 4.0.0
|
| 79 |
- Tokenizers 0.21.2
|
|
|
|
| 4 |
base_model: roberta-base
|
| 5 |
tags:
|
| 6 |
- generated_from_trainer
|
| 7 |
+
metrics:
|
| 8 |
+
- accuracy
|
| 9 |
model-index:
|
| 10 |
- name: cwe-parent-vulnerability-classification-roberta-base
|
| 11 |
results: []
|
|
|
|
| 18 |
|
| 19 |
This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on an unknown dataset.
|
| 20 |
It achieves the following results on the evaluation set:
|
| 21 |
+
- Loss: 0.6890
|
| 22 |
+
- Accuracy: 0.8427
|
| 23 |
+
- F1 Macro: 0.3821
|
| 24 |
|
| 25 |
## Model description
|
| 26 |
|
|
|
|
| 49 |
|
| 50 |
### Training results
|
| 51 |
|
| 52 |
+
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 Macro |
|
| 53 |
+
|:-------------:|:-----:|:----:|:---------------:|:--------:|:--------:|
|
| 54 |
+
| 2.8193 | 1.0 | 25 | 2.1550 | 0.4270 | 0.0460 |
|
| 55 |
+
| 1.9743 | 2.0 | 50 | 1.9856 | 0.4270 | 0.0460 |
|
| 56 |
+
| 1.7803 | 3.0 | 75 | 1.9237 | 0.4270 | 0.0460 |
|
| 57 |
+
| 1.5493 | 4.0 | 100 | 1.4930 | 0.6292 | 0.1120 |
|
| 58 |
+
| 1.2607 | 5.0 | 125 | 1.3301 | 0.6742 | 0.1229 |
|
| 59 |
+
| 1.1316 | 6.0 | 150 | 1.2706 | 0.6742 | 0.1229 |
|
| 60 |
+
| 1.0911 | 7.0 | 175 | 1.1787 | 0.6742 | 0.1229 |
|
| 61 |
+
| 0.9869 | 8.0 | 200 | 1.0819 | 0.7416 | 0.2359 |
|
| 62 |
+
| 0.8985 | 9.0 | 225 | 1.0360 | 0.7640 | 0.2587 |
|
| 63 |
+
| 0.7404 | 10.0 | 250 | 0.9223 | 0.7753 | 0.2498 |
|
| 64 |
+
| 0.8315 | 11.0 | 275 | 0.8992 | 0.7753 | 0.2441 |
|
| 65 |
+
| 0.6514 | 12.0 | 300 | 0.8443 | 0.7753 | 0.2395 |
|
| 66 |
+
| 0.6427 | 13.0 | 325 | 0.8207 | 0.7753 | 0.2395 |
|
| 67 |
+
| 0.5719 | 14.0 | 350 | 0.7653 | 0.8090 | 0.2887 |
|
| 68 |
+
| 0.5153 | 15.0 | 375 | 0.7396 | 0.8202 | 0.3028 |
|
| 69 |
+
| 0.4863 | 16.0 | 400 | 0.7580 | 0.8202 | 0.3376 |
|
| 70 |
+
| 0.4833 | 17.0 | 425 | 0.7177 | 0.8315 | 0.3806 |
|
| 71 |
+
| 0.4948 | 18.0 | 450 | 0.6983 | 0.8427 | 0.3821 |
|
| 72 |
+
| 0.4546 | 19.0 | 475 | 0.6890 | 0.8427 | 0.3821 |
|
| 73 |
+
| 0.4445 | 20.0 | 500 | 0.6911 | 0.8427 | 0.3821 |
|
| 74 |
|
| 75 |
|
| 76 |
### Framework versions
|
| 77 |
|
| 78 |
+
- Transformers 4.55.4
|
| 79 |
- Pytorch 2.7.1+cu126
|
| 80 |
- Datasets 4.0.0
|
| 81 |
- Tokenizers 0.21.2
|
emissions.csv
CHANGED
|
@@ -1,2 +1,2 @@
|
|
| 1 |
timestamp,project_name,run_id,experiment_id,duration,emissions,emissions_rate,cpu_power,gpu_power,ram_power,cpu_energy,gpu_energy,ram_energy,energy_consumed,country_name,country_iso_code,region,cloud_provider,cloud_region,os,python_version,codecarbon_version,cpu_count,cpu_model,gpu_count,gpu_model,longitude,latitude,ram_total_size,tracking_mode,on_cloud,pue
|
| 2 |
-
2025-08-
|
|
|
|
| 1 |
timestamp,project_name,run_id,experiment_id,duration,emissions,emissions_rate,cpu_power,gpu_power,ram_power,cpu_energy,gpu_energy,ram_energy,energy_consumed,country_name,country_iso_code,region,cloud_provider,cloud_region,os,python_version,codecarbon_version,cpu_count,cpu_model,gpu_count,gpu_model,longitude,latitude,ram_total_size,tracking_mode,on_cloud,pue
|
| 2 |
+
2025-08-28T16:05:20,codecarbon,e2a8bcaf-0f58-4bb5-a248-bf70c0976bdd,5b0fa12a-3dd7-45bb-9766-cc326314d9f1,182.2731417969335,0.003007971348785084,1.6502548423377697e-05,42.5,225.04609279119165,94.34468507766725,0.0021501255845857996,0.021653612322872107,0.004772032769656413,0.028575770677114313,Luxembourg,LUX,luxembourg,,,Linux-6.8.0-71-generic-x86_64-with-glibc2.39,3.12.3,2.8.4,64,AMD EPYC 9124 16-Core Processor,2,2 x NVIDIA L40S,6.1294,49.6113,251.5858268737793,machine,N,1.0
|
metrics.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
-
"eval_loss": 0.
|
| 3 |
-
"
|
| 4 |
-
"
|
| 5 |
-
"eval_runtime": 0.
|
| 6 |
-
"eval_samples_per_second":
|
| 7 |
-
"eval_steps_per_second":
|
| 8 |
"epoch": 20.0
|
| 9 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"eval_loss": 0.6889647245407104,
|
| 3 |
+
"eval_accuracy": 0.8426966292134831,
|
| 4 |
+
"eval_f1_macro": 0.3820724249295678,
|
| 5 |
+
"eval_runtime": 0.3068,
|
| 6 |
+
"eval_samples_per_second": 290.118,
|
| 7 |
+
"eval_steps_per_second": 9.779,
|
| 8 |
"epoch": 20.0
|
| 9 |
}
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 498686648
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:40cb69ba70445c2b0f6c62d1c8d6567e93d3105fbd0204f204d5af4fdd4e102f
|
| 3 |
size 498686648
|