mdeberta-semeval25_thresh07_fold1
This model is a fine-tuned version of microsoft/mdeberta-v3-base on the None dataset. It achieves the following results on the evaluation set:
- Loss: 8.6241
- Precision Samples: 0.1204
- Recall Samples: 0.5520
- F1 Samples: 0.1866
- Precision Macro: 0.8184
- Recall Macro: 0.3704
- F1 Macro: 0.2547
- Precision Micro: 0.1111
- Recall Micro: 0.4660
- F1 Micro: 0.1794
- Precision Weighted: 0.5223
- Recall Weighted: 0.4660
- F1 Weighted: 0.1091
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- num_epochs: 10
Training results
| Training Loss | Epoch | Step | Validation Loss | Precision Samples | Recall Samples | F1 Samples | Precision Macro | Recall Macro | F1 Macro | Precision Micro | Recall Micro | F1 Micro | Precision Weighted | Recall Weighted | F1 Weighted |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| 10.7091 | 1.0 | 19 | 9.6339 | 0.2123 | 0.1986 | 0.1986 | 0.9911 | 0.2333 | 0.2259 | 0.2014 | 0.0895 | 0.1239 | 0.9285 | 0.0895 | 0.0300 |
| 10.3202 | 2.0 | 38 | 9.2779 | 0.1495 | 0.2401 | 0.1695 | 0.9702 | 0.2444 | 0.2282 | 0.1508 | 0.1420 | 0.1463 | 0.8572 | 0.1420 | 0.0406 |
| 9.7575 | 3.0 | 57 | 9.1258 | 0.1484 | 0.3153 | 0.1854 | 0.9337 | 0.2634 | 0.2371 | 0.1474 | 0.2130 | 0.1742 | 0.7249 | 0.2130 | 0.0727 |
| 10.0673 | 4.0 | 76 | 9.0135 | 0.1121 | 0.4228 | 0.1630 | 0.9006 | 0.2995 | 0.2406 | 0.1101 | 0.3302 | 0.1651 | 0.6447 | 0.3302 | 0.0812 |
| 9.7122 | 5.0 | 95 | 8.9266 | 0.1032 | 0.4806 | 0.1584 | 0.8482 | 0.3301 | 0.2474 | 0.1014 | 0.3920 | 0.1612 | 0.5456 | 0.3920 | 0.0933 |
| 9.755 | 6.0 | 114 | 8.8301 | 0.1091 | 0.5002 | 0.1670 | 0.8389 | 0.3407 | 0.2505 | 0.1037 | 0.4167 | 0.1661 | 0.5468 | 0.4167 | 0.1010 |
| 9.3217 | 7.0 | 133 | 8.7446 | 0.1146 | 0.5249 | 0.1766 | 0.8395 | 0.3500 | 0.2522 | 0.1085 | 0.4383 | 0.1739 | 0.5486 | 0.4383 | 0.1057 |
| 9.4961 | 8.0 | 152 | 8.6814 | 0.1185 | 0.5383 | 0.1825 | 0.8299 | 0.3619 | 0.2550 | 0.1114 | 0.4537 | 0.1789 | 0.5313 | 0.4537 | 0.1083 |
| 9.4404 | 9.0 | 171 | 8.6369 | 0.1179 | 0.5406 | 0.1818 | 0.8180 | 0.3641 | 0.2539 | 0.1092 | 0.4568 | 0.1763 | 0.5213 | 0.4568 | 0.1073 |
| 9.1011 | 10.0 | 190 | 8.6241 | 0.1204 | 0.5520 | 0.1866 | 0.8184 | 0.3704 | 0.2547 | 0.1111 | 0.4660 | 0.1794 | 0.5223 | 0.4660 | 0.1091 |
Framework versions
- Transformers 4.46.0
- Pytorch 2.3.1
- Datasets 2.21.0
- Tokenizers 0.20.1
- Downloads last month
- 7
Model tree for g-assismoraes/mdeberta-semeval25_thresh07_fold1
Base model
microsoft/mdeberta-v3-base