g-assismoraes commited on
Commit
fb42f38
·
verified ·
1 Parent(s): 3b16473

End of training

Browse files
README.md ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: mit
4
+ base_model: microsoft/mdeberta-v3-base
5
+ tags:
6
+ - generated_from_trainer
7
+ model-index:
8
+ - name: mdeberta-semeval25_narratives09_fold1
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ # mdeberta-semeval25_narratives09_fold1
16
+
17
+ This model is a fine-tuned version of [microsoft/mdeberta-v3-base](https://huggingface.co/microsoft/mdeberta-v3-base) on the None dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 4.1440
20
+ - Precision Samples: 0.3489
21
+ - Recall Samples: 0.7666
22
+ - F1 Samples: 0.4484
23
+ - Precision Macro: 0.6713
24
+ - Recall Macro: 0.4701
25
+ - F1 Macro: 0.2642
26
+ - Precision Micro: 0.3133
27
+ - Recall Micro: 0.7518
28
+ - F1 Micro: 0.4423
29
+ - Precision Weighted: 0.4454
30
+ - Recall Weighted: 0.7518
31
+ - F1 Weighted: 0.3929
32
+
33
+ ## Model description
34
+
35
+ More information needed
36
+
37
+ ## Intended uses & limitations
38
+
39
+ More information needed
40
+
41
+ ## Training and evaluation data
42
+
43
+ More information needed
44
+
45
+ ## Training procedure
46
+
47
+ ### Training hyperparameters
48
+
49
+ The following hyperparameters were used during training:
50
+ - learning_rate: 2e-05
51
+ - train_batch_size: 32
52
+ - eval_batch_size: 32
53
+ - seed: 42
54
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
55
+ - lr_scheduler_type: linear
56
+ - num_epochs: 10
57
+
58
+ ### Training results
59
+
60
+ | Training Loss | Epoch | Step | Validation Loss | Precision Samples | Recall Samples | F1 Samples | Precision Macro | Recall Macro | F1 Macro | Precision Micro | Recall Micro | F1 Micro | Precision Weighted | Recall Weighted | F1 Weighted |
61
+ |:-------------:|:-----:|:----:|:---------------:|:-----------------:|:--------------:|:----------:|:---------------:|:------------:|:--------:|:---------------:|:------------:|:--------:|:------------------:|:---------------:|:-----------:|
62
+ | 5.3976 | 1.0 | 19 | 5.3094 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0476 | 0.0476 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 |
63
+ | 5.0729 | 2.0 | 38 | 5.0051 | 0.2991 | 0.4812 | 0.3465 | 0.8683 | 0.2245 | 0.1355 | 0.3056 | 0.4496 | 0.3639 | 0.6682 | 0.4496 | 0.2244 |
64
+ | 4.799 | 3.0 | 57 | 4.7268 | 0.3634 | 0.5035 | 0.3759 | 0.8348 | 0.2364 | 0.1574 | 0.3291 | 0.4640 | 0.3851 | 0.6206 | 0.4640 | 0.2617 |
65
+ | 4.4077 | 4.0 | 76 | 4.5072 | 0.3846 | 0.6225 | 0.4435 | 0.7933 | 0.3190 | 0.2043 | 0.3383 | 0.5755 | 0.4261 | 0.5591 | 0.5755 | 0.3232 |
66
+ | 4.1905 | 5.0 | 95 | 4.3919 | 0.4006 | 0.6444 | 0.4575 | 0.7484 | 0.3320 | 0.2140 | 0.3395 | 0.5935 | 0.4319 | 0.5242 | 0.5935 | 0.3411 |
67
+ | 4.1939 | 6.0 | 114 | 4.2724 | 0.3817 | 0.7296 | 0.4634 | 0.7094 | 0.4205 | 0.2478 | 0.3229 | 0.7050 | 0.4429 | 0.4663 | 0.7050 | 0.3791 |
68
+ | 3.9286 | 7.0 | 133 | 4.2600 | 0.3753 | 0.7336 | 0.4620 | 0.6853 | 0.4257 | 0.2568 | 0.3311 | 0.7050 | 0.4506 | 0.4556 | 0.7050 | 0.3882 |
69
+ | 3.8896 | 8.0 | 152 | 4.1871 | 0.3528 | 0.7581 | 0.4505 | 0.6713 | 0.4559 | 0.2625 | 0.3188 | 0.7374 | 0.4452 | 0.4462 | 0.7374 | 0.3929 |
70
+ | 3.993 | 9.0 | 171 | 4.1598 | 0.3525 | 0.7629 | 0.4503 | 0.6712 | 0.4645 | 0.2639 | 0.3170 | 0.7446 | 0.4447 | 0.4443 | 0.7446 | 0.3920 |
71
+ | 4.1424 | 10.0 | 190 | 4.1440 | 0.3489 | 0.7666 | 0.4484 | 0.6713 | 0.4701 | 0.2642 | 0.3133 | 0.7518 | 0.4423 | 0.4454 | 0.7518 | 0.3929 |
72
+
73
+
74
+ ### Framework versions
75
+
76
+ - Transformers 4.46.0
77
+ - Pytorch 2.3.1
78
+ - Datasets 2.21.0
79
+ - Tokenizers 0.20.1
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2092065d86ad07ad6d91b3136c82bd0088775e94a95cf984af0943d6290ecd6c
3
  size 1115326652
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a173501f565035d68faf205823262ce3f93b4c9812eb5778b5263882f1a561b9
3
  size 1115326652
runs/Oct27_16-37-16_icuff-Z790-UD/events.out.tfevents.1730057837.icuff-Z790-UD.525411.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:db6f486c7ce68e450e2c263a45d3addb8f6eb39b1d11c98623ae9f2e29ed8697
3
- size 18135
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de6220b2eb740f49fac86e2c53cee7fc81a538a2eba07f873484e4cb17ac91bc
3
+ size 19865