parlange commited on
Commit
6c89b58
·
verified ·
1 Parent(s): 1cb115d

Upload ViT model from experiment s2

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +2 -0
  2. README.md +165 -0
  3. config.json +76 -0
  4. confusion_matrices/ViT_Confusion_Matrix_a.png +0 -0
  5. confusion_matrices/ViT_Confusion_Matrix_b.png +0 -0
  6. confusion_matrices/ViT_Confusion_Matrix_c.png +0 -0
  7. confusion_matrices/ViT_Confusion_Matrix_d.png +0 -0
  8. confusion_matrices/ViT_Confusion_Matrix_e.png +0 -0
  9. confusion_matrices/ViT_Confusion_Matrix_f.png +0 -0
  10. confusion_matrices/ViT_Confusion_Matrix_g.png +0 -0
  11. confusion_matrices/ViT_Confusion_Matrix_h.png +0 -0
  12. confusion_matrices/ViT_Confusion_Matrix_i.png +0 -0
  13. confusion_matrices/ViT_Confusion_Matrix_j.png +0 -0
  14. confusion_matrices/ViT_Confusion_Matrix_k.png +0 -0
  15. confusion_matrices/ViT_Confusion_Matrix_l.png +0 -0
  16. evaluation_results.csv +133 -0
  17. model.safetensors +3 -0
  18. pytorch_model.bin +3 -0
  19. roc_confusion_matrix/ViT_roc_confusion_matrix_a.png +0 -0
  20. roc_confusion_matrix/ViT_roc_confusion_matrix_b.png +0 -0
  21. roc_confusion_matrix/ViT_roc_confusion_matrix_c.png +0 -0
  22. roc_confusion_matrix/ViT_roc_confusion_matrix_d.png +0 -0
  23. roc_confusion_matrix/ViT_roc_confusion_matrix_e.png +0 -0
  24. roc_confusion_matrix/ViT_roc_confusion_matrix_f.png +0 -0
  25. roc_confusion_matrix/ViT_roc_confusion_matrix_g.png +0 -0
  26. roc_confusion_matrix/ViT_roc_confusion_matrix_h.png +0 -0
  27. roc_confusion_matrix/ViT_roc_confusion_matrix_i.png +0 -0
  28. roc_confusion_matrix/ViT_roc_confusion_matrix_j.png +0 -0
  29. roc_confusion_matrix/ViT_roc_confusion_matrix_k.png +0 -0
  30. roc_confusion_matrix/ViT_roc_confusion_matrix_l.png +0 -0
  31. roc_curves/ViT_ROC_a.png +0 -0
  32. roc_curves/ViT_ROC_b.png +0 -0
  33. roc_curves/ViT_ROC_c.png +0 -0
  34. roc_curves/ViT_ROC_d.png +0 -0
  35. roc_curves/ViT_ROC_e.png +0 -0
  36. roc_curves/ViT_ROC_f.png +0 -0
  37. roc_curves/ViT_ROC_g.png +0 -0
  38. roc_curves/ViT_ROC_h.png +0 -0
  39. roc_curves/ViT_ROC_i.png +0 -0
  40. roc_curves/ViT_ROC_j.png +0 -0
  41. roc_curves/ViT_ROC_k.png +0 -0
  42. roc_curves/ViT_ROC_l.png +0 -0
  43. training_curves/ViT_accuracy.png +0 -0
  44. training_curves/ViT_auc.png +0 -0
  45. training_curves/ViT_combined_metrics.png +3 -0
  46. training_curves/ViT_f1.png +0 -0
  47. training_curves/ViT_loss.png +0 -0
  48. training_curves/ViT_metrics.csv +56 -0
  49. training_metrics.csv +56 -0
  50. training_notebook_s2.ipynb +3 -0
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ training_curves/ViT_combined_metrics.png filter=lfs diff=lfs merge=lfs -text
37
+ training_notebook_s2.ipynb filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,165 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - vision-transformer
5
+ - image-classification
6
+ - pytorch
7
+ - timm
8
+ - vit
9
+ - gravitational-lensing
10
+ - strong-lensing
11
+ - astronomy
12
+ - astrophysics
13
+ datasets:
14
+ - parlange/gravit-c21
15
+ metrics:
16
+ - accuracy
17
+ - auc
18
+ - f1
19
+ paper:
20
+ - title: "GraViT: A Gravitational Lens Discovery Toolkit with Vision Transformers"
21
+ url: "https://arxiv.org/abs/2509.00226"
22
+ authors: "Parlange et al."
23
+ model-index:
24
+ - name: ViT-s2
25
+ results:
26
+ - task:
27
+ type: image-classification
28
+ name: Strong Gravitational Lens Discovery
29
+ dataset:
30
+ type: common-test-sample
31
+ name: Common Test Sample (More et al. 2024)
32
+ metrics:
33
+ - type: accuracy
34
+ value: 0.8558
35
+ name: Average Accuracy
36
+ - type: auc
37
+ value: 0.8814
38
+ name: Average AUC-ROC
39
+ - type: f1
40
+ value: 0.5950
41
+ name: Average F1-Score
42
+ ---
43
+
44
+ # 🌌 vit-gravit-s2
45
+
46
+ 🔭 This model is part of **GraViT**: Transfer Learning with Vision Transformers and MLP-Mixer for Strong Gravitational Lens Discovery
47
+
48
+ 🔗 **GitHub Repository**: [https://github.com/parlange/gravit](https://github.com/parlange/gravit)
49
+
50
+ ## 🛰️ Model Details
51
+
52
+ - **🤖 Model Type**: ViT
53
+ - **🧪 Experiment**: S2 - C21-half-18660
54
+ - **🌌 Dataset**: C21
55
+ - **🪐 Fine-tuning Strategy**: half
56
+
57
+ - **🎲 Random Seed**: 18660
58
+
59
+ ## 💻 Quick Start
60
+
61
+ ```python
62
+ import torch
63
+ import timm
64
+
65
+ # Load the model directly from the Hub
66
+ model = timm.create_model(
67
+ 'hf-hub:parlange/vit-gravit-s2',
68
+ pretrained=True
69
+ )
70
+ model.eval()
71
+
72
+ # Example inference
73
+ dummy_input = torch.randn(1, 3, 224, 224)
74
+ with torch.no_grad():
75
+ output = model(dummy_input)
76
+ predictions = torch.softmax(output, dim=1)
77
+ print(f"Lens probability: {predictions[0][1]:.4f}")
78
+ ```
79
+
80
+ ## ⚡️ Training Configuration
81
+
82
+ **Training Dataset:** C21 (Cañameras et al. 2021)
83
+ **Fine-tuning Strategy:** half
84
+
85
+
86
+ | 🔧 Parameter | 📝 Value |
87
+ |--------------|----------|
88
+ | Batch Size | 192 |
89
+ | Learning Rate | AdamW with ReduceLROnPlateau |
90
+ | Epochs | 100 |
91
+ | Patience | 10 |
92
+ | Optimizer | AdamW |
93
+ | Scheduler | ReduceLROnPlateau |
94
+ | Image Size | 224x224 |
95
+ | Fine Tune Mode | half |
96
+ | Stochastic Depth Probability | 0.1 |
97
+
98
+
99
+ ## 📈 Training Curves
100
+
101
+ ![Combined Training Metrics](https://huggingface.co/parlange/vit-gravit-s2/resolve/main/training_curves/ViT_combined_metrics.png)
102
+
103
+
104
+ ## 🏁 Final Epoch Training Metrics
105
+
106
+ | Metric | Training | Validation |
107
+ |:---------:|:-----------:|:-------------:|
108
+ | 📉 Loss | 0.0160 | 0.0710 |
109
+ | 🎯 Accuracy | 0.9939 | 0.9820 |
110
+ | 📊 AUC-ROC | 0.9998 | 0.9984 |
111
+ | ⚖️ F1 Score | 0.9939 | 0.9819 |
112
+
113
+
114
+ ## ☑️ Evaluation Results
115
+
116
+ ### ROC Curves and Confusion Matrices
117
+
118
+ Performance across all test datasets (a through l) in the Common Test Sample (More et al. 2024):
119
+
120
+ ![ROC + Confusion Matrix - Dataset A](https://huggingface.co/parlange/vit-gravit-s2/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_a.png)
121
+ ![ROC + Confusion Matrix - Dataset B](https://huggingface.co/parlange/vit-gravit-s2/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_b.png)
122
+ ![ROC + Confusion Matrix - Dataset C](https://huggingface.co/parlange/vit-gravit-s2/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_c.png)
123
+ ![ROC + Confusion Matrix - Dataset D](https://huggingface.co/parlange/vit-gravit-s2/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_d.png)
124
+ ![ROC + Confusion Matrix - Dataset E](https://huggingface.co/parlange/vit-gravit-s2/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_e.png)
125
+ ![ROC + Confusion Matrix - Dataset F](https://huggingface.co/parlange/vit-gravit-s2/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_f.png)
126
+ ![ROC + Confusion Matrix - Dataset G](https://huggingface.co/parlange/vit-gravit-s2/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_g.png)
127
+ ![ROC + Confusion Matrix - Dataset H](https://huggingface.co/parlange/vit-gravit-s2/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_h.png)
128
+ ![ROC + Confusion Matrix - Dataset I](https://huggingface.co/parlange/vit-gravit-s2/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_i.png)
129
+ ![ROC + Confusion Matrix - Dataset J](https://huggingface.co/parlange/vit-gravit-s2/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_j.png)
130
+ ![ROC + Confusion Matrix - Dataset K](https://huggingface.co/parlange/vit-gravit-s2/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_k.png)
131
+ ![ROC + Confusion Matrix - Dataset L](https://huggingface.co/parlange/vit-gravit-s2/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_l.png)
132
+
133
+ ### 📋 Performance Summary
134
+
135
+ Average performance across 12 test datasets from the Common Test Sample (More et al. 2024):
136
+
137
+ | Metric | Value |
138
+ |-----------|----------|
139
+ | 🎯 Average Accuracy | 0.8558 |
140
+ | 📈 Average AUC-ROC | 0.8814 |
141
+ | ⚖️ Average F1-Score | 0.5950 |
142
+
143
+
144
+ ## 📘 Citation
145
+
146
+ If you use this model in your research, please cite:
147
+
148
+ ```bibtex
149
+ @misc{parlange2025gravit,
150
+ title={GraViT: Transfer Learning with Vision Transformers and MLP-Mixer for Strong Gravitational Lens Discovery},
151
+ author={René Parlange and Juan C. Cuevas-Tello and Octavio Valenzuela and Omar de J. Cabrera-Rosas and Tomás Verdugo and Anupreeta More and Anton T. Jaelani},
152
+ year={2025},
153
+ eprint={2509.00226},
154
+ archivePrefix={arXiv},
155
+ primaryClass={cs.CV},
156
+ url={https://arxiv.org/abs/2509.00226},
157
+ }
158
+ ```
159
+
160
+ ---
161
+
162
+
163
+ ## Model Card Contact
164
+
165
+ For questions about this model, please contact the author through: https://github.com/parlange/
config.json ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architecture": "vit_base_patch16_224",
3
+ "num_classes": 2,
4
+ "num_features": 768,
5
+ "global_pool": "token",
6
+ "crop_pct": 0.875,
7
+ "interpolation": "bicubic",
8
+ "mean": [
9
+ 0.485,
10
+ 0.456,
11
+ 0.406
12
+ ],
13
+ "std": [
14
+ 0.229,
15
+ 0.224,
16
+ 0.225
17
+ ],
18
+ "first_conv": "patch_embed.proj",
19
+ "classifier": "head",
20
+ "input_size": [
21
+ 3,
22
+ 224,
23
+ 224
24
+ ],
25
+ "pool_size": [
26
+ 7,
27
+ 7
28
+ ],
29
+ "pretrained_cfg": {
30
+ "tag": "gravit_s2",
31
+ "custom_load": false,
32
+ "input_size": [
33
+ 3,
34
+ 224,
35
+ 224
36
+ ],
37
+ "fixed_input_size": true,
38
+ "interpolation": "bicubic",
39
+ "crop_pct": 0.875,
40
+ "crop_mode": "center",
41
+ "mean": [
42
+ 0.485,
43
+ 0.456,
44
+ 0.406
45
+ ],
46
+ "std": [
47
+ 0.229,
48
+ 0.224,
49
+ 0.225
50
+ ],
51
+ "num_classes": 2,
52
+ "pool_size": [
53
+ 7,
54
+ 7
55
+ ],
56
+ "first_conv": "patch_embed.proj",
57
+ "classifier": "head"
58
+ },
59
+ "model_name": "vit_gravit_s2",
60
+ "experiment": "s2",
61
+ "training_strategy": "half",
62
+ "dataset": "C21",
63
+ "hyperparameters": {
64
+ "batch_size": "192",
65
+ "learning_rate": "AdamW with ReduceLROnPlateau",
66
+ "epochs": "100",
67
+ "patience": "10",
68
+ "optimizer": "AdamW",
69
+ "scheduler": "ReduceLROnPlateau",
70
+ "image_size": "224x224",
71
+ "fine_tune_mode": "half",
72
+ "stochastic_depth_probability": "0.1"
73
+ },
74
+ "hf_hub_id": "parlange/vit-gravit-s2",
75
+ "license": "apache-2.0"
76
+ }
confusion_matrices/ViT_Confusion_Matrix_a.png ADDED
confusion_matrices/ViT_Confusion_Matrix_b.png ADDED
confusion_matrices/ViT_Confusion_Matrix_c.png ADDED
confusion_matrices/ViT_Confusion_Matrix_d.png ADDED
confusion_matrices/ViT_Confusion_Matrix_e.png ADDED
confusion_matrices/ViT_Confusion_Matrix_f.png ADDED
confusion_matrices/ViT_Confusion_Matrix_g.png ADDED
confusion_matrices/ViT_Confusion_Matrix_h.png ADDED
confusion_matrices/ViT_Confusion_Matrix_i.png ADDED
confusion_matrices/ViT_Confusion_Matrix_j.png ADDED
confusion_matrices/ViT_Confusion_Matrix_k.png ADDED
confusion_matrices/ViT_Confusion_Matrix_l.png ADDED
evaluation_results.csv ADDED
@@ -0,0 +1,133 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Model,Dataset,Loss,Accuracy,AUCROC,F1
2
+ ViT,a,0.26081677189425306,0.9402703552342031,0.948634438305709,0.602510460251046
3
+ ViT,b,0.22714076965991073,0.9490726186733731,0.954792817679558,0.64
4
+ ViT,c,0.5409317808831798,0.8802263439170073,0.9180930018416207,0.4304932735426009
5
+ ViT,d,0.2093577366257868,0.9440427538509902,0.9635985267034991,0.6180257510729614
6
+ ViT,e,0.626060025425041,0.8693743139407245,0.9101869371073942,0.7076167076167076
7
+ ViT,f,0.28432284784108347,0.9316861590891488,0.9442099621115129,0.24615384615384617
8
+ ViT,g,0.0920504999384284,0.9766666666666667,0.9990343333333334,0.9770867430441899
9
+ ViT,h,0.2584120511338115,0.9401666666666667,0.9969072222222223,0.9432769789856218
10
+ ViT,i,0.08262255262583494,0.974,0.9991285555555557,0.9745347698334965
11
+ ViT,j,4.850409872978926,0.5285,0.5466985,0.17304881613563286
12
+ ViT,k,4.840981937706471,0.5258333333333334,0.6174096111111111,0.17224323537969158
13
+ ViT,l,1.707986980357526,0.8095817249220031,0.7775850352542502,0.6554396708448952
14
+ MLP-Mixer,a,0.3746531411129437,0.8808550770198051,0.9587697974217311,0.4669479606188467
15
+ MLP-Mixer,b,0.4263077380896289,0.8786545111600126,0.9555377532228361,0.4623955431754875
16
+ MLP-Mixer,c,0.7193417321657543,0.8101226029550456,0.9323139963167587,0.3547008547008547
17
+ MLP-Mixer,d,0.06999178051348941,0.9739075762338887,0.9913977900552486,0.8
18
+ MLP-Mixer,e,0.5372455838482937,0.8397365532381997,0.938428063271021,0.694560669456067
19
+ MLP-Mixer,f,0.40886121419995697,0.8808767717450237,0.9582961898851193,0.17754010695187167
20
+ MLP-Mixer,g,0.220160967502743,0.936,0.9966087777777778,0.9396036489462095
21
+ MLP-Mixer,h,0.3755178214646876,0.8996666666666666,0.9936971666666665,0.9084549878345499
22
+ MLP-Mixer,i,0.03125413155928254,0.9865,0.9996044444444444,0.9866226259289843
23
+ MLP-Mixer,j,4.689226661682129,0.45866666666666667,0.35093755555555556,0.07040641099026904
24
+ MLP-Mixer,k,4.50031984564662,0.5091666666666667,0.5708096111111112,0.0770918207458477
25
+ MLP-Mixer,l,1.7004726856614425,0.7658505631642959,0.7001407272343229,0.5967213114754099
26
+ CvT,a,0.7275567101967556,0.6862621817038667,0.8178130755064457,0.223950233281493
27
+ CvT,b,0.8674648388381725,0.6309336686576549,0.7877642725598526,0.19699042407660738
28
+ CvT,c,0.8444347186157517,0.6369066331342346,0.7903591160220995,0.1995841995841996
29
+ CvT,d,0.06786359480738227,0.9761081420936812,0.984121546961326,0.7912087912087912
30
+ CvT,e,1.079856214224965,0.544456641053787,0.714894422159994,0.4096728307254623
31
+ CvT,f,0.6667535994093697,0.7157462628766168,0.8375527856501153,0.07276402223345124
32
+ CvT,g,0.4902210609912872,0.7958333333333333,0.9368645,0.8262164846077458
33
+ CvT,h,0.4780112521648407,0.799,0.9404210555555556,0.8284495021337127
34
+ CvT,i,0.06629910692572594,0.9788333333333333,0.998182,0.9786590488993446
35
+ CvT,j,3.285769058704376,0.31566666666666665,0.11980044444444445,0.014875239923224568
36
+ CvT,k,2.8618470991551876,0.49866666666666665,0.5598825555555555,0.020195439739413682
37
+ CvT,l,1.3706902678378083,0.6442811062344667,0.5954304635509091,0.4785675529028757
38
+ Swin,a,0.17904385026568925,0.9320968248978309,0.9377541436464087,0.5573770491803278
39
+ Swin,b,0.16908444278234958,0.9302106255894372,0.9438968692449355,0.5506072874493927
40
+ Swin,c,0.3163154192856744,0.8805407104684062,0.9053370165745855,0.4171779141104294
41
+ Swin,d,0.05074489927894287,0.9833385727758567,0.9865469613259669,0.8369230769230769
42
+ Swin,e,0.4723492567023645,0.7859495060373216,0.8619995459017633,0.582441113490364
43
+ Swin,f,0.17172798980294332,0.9313763457516846,0.9387167824732112,0.23488773747841105
44
+ Swin,g,0.07050645374506712,0.9703333333333334,0.9997747777777777,0.971178756476684
45
+ Swin,h,0.14856340130418538,0.944,0.9992891111111111,0.9469529523208083
46
+ Swin,i,0.007766767971217632,0.9985,0.9999895555555555,0.9985017479607124
47
+ Swin,j,2.8674492346346376,0.485,0.43114961111111116,0.05330882352941176
48
+ Swin,k,2.804709552191198,0.5131666666666667,0.6342635555555556,0.056219709208400644
49
+ Swin,l,1.0054080692499547,0.7990587488763153,0.7133887608594007,0.6290511518937915
50
+ CaiT,a,0.4621231952897036,0.821754165356806,0.9297688766114179,0.36363636363636365
51
+ CaiT,b,0.258825070179322,0.9122917321596982,0.9613351749539596,0.5373134328358209
52
+ CaiT,c,0.8353647122552507,0.685004715498271,0.8978766114180479,0.24434389140271492
53
+ CaiT,d,0.031897671438732525,0.9902546369066332,0.9952596685082873,0.9126760563380282
54
+ CaiT,e,0.4524560969171618,0.8463227222832053,0.9379777491864074,0.6982758620689655
55
+ CaiT,f,0.4045827676295477,0.8495081713267756,0.9455966026222479,0.14292015880017644
56
+ CaiT,g,0.12979762570466846,0.9553333333333334,0.9989632222222222,0.9571337172104927
57
+ CaiT,h,0.43545972697343677,0.8348333333333333,0.9963943888888888,0.857921146953405
58
+ CaiT,i,0.009488287813030183,0.9966666666666667,0.9999612222222222,0.9966688874083944
59
+ CaiT,j,3.0854302213191986,0.4766666666666667,0.32313933333333333,0.07100591715976332
60
+ CaiT,k,2.9651209139451384,0.518,0.598769,0.07662835249042145
61
+ CaiT,l,1.2156231775795154,0.7445402146898631,0.6535540584867622,0.5754459970120397
62
+ DeiT,a,0.21973140770554805,0.9163784973278843,0.9421279926335173,0.5333333333333333
63
+ DeiT,b,0.18026995446669986,0.9349261238604213,0.9484125230202579,0.5949119373776908
64
+ DeiT,c,0.24070323096322999,0.9154353976736875,0.9384106813996317,0.5305410122164049
65
+ DeiT,d,0.045521379231525945,0.9877397044954417,0.9928232044198896,0.8862973760932945
66
+ DeiT,e,0.6630487046833227,0.7760702524698134,0.8773329296904565,0.5984251968503937
67
+ DeiT,f,0.1809546241546691,0.9326930524359074,0.9509643553098133,0.2591645353793691
68
+ DeiT,g,0.08131901465170085,0.9685,0.9991698888888889,0.9693530079455165
69
+ DeiT,h,0.11335872827284038,0.9581666666666667,0.9988661111111111,0.9597046074811366
70
+ DeiT,i,0.009879813833162188,0.9965,0.9999472222222223,0.9964994165694282
71
+ DeiT,j,3.730827052116394,0.49016666666666664,0.3371602222222222,0.07218683651804671
72
+ DeiT,k,3.6593878800719977,0.5181666666666667,0.7405100555555556,0.07606263982102908
73
+ DeiT,l,1.2843114430264382,0.8011210406641637,0.708836413391112,0.634179554518043
74
+ DeiT3,a,0.10351562584202569,0.9666771455517132,0.967061694290976,0.7309644670050761
75
+ DeiT3,b,0.10786095303265189,0.9651053127947187,0.9586629834254143,0.7218045112781954
76
+ DeiT3,c,0.12137072830127493,0.9622760138321282,0.9632596685082873,0.7058823529411765
77
+ DeiT3,d,0.05938508253192066,0.9852247720842502,0.9764714548802946,0.8597014925373134
78
+ DeiT3,e,0.2852305471569987,0.9198682766190999,0.9255581624158026,0.7977839335180056
79
+ DeiT3,f,0.06806116103929771,0.9760669196808922,0.9640239483015283,0.4824120603015075
80
+ DeiT3,g,0.03457943443208933,0.9858333333333333,0.9997122222222222,0.9859805376876134
81
+ DeiT3,h,0.04174186686426401,0.9843333333333333,0.99965,0.9845191040843215
82
+ DeiT3,i,0.00887914503365755,0.9965,0.9999582222222222,0.9964994165694282
83
+ DeiT3,j,3.737933710604906,0.519,0.5266332222222222,0.11526670754138565
84
+ DeiT3,k,3.71223342192173,0.5296666666666666,0.6497483333333334,0.11757348342714197
85
+ DeiT3,l,1.2237240775665559,0.8343821056527947,0.8068982843173076,0.6795580110497238
86
+ Twins_SVT,a,0.42038975148709157,0.8192392329456146,0.8869475138121546,0.33983926521239954
87
+ Twins_SVT,b,0.4005945102762254,0.8091795033008488,0.8857955801104973,0.327796234772979
88
+ Twins_SVT,c,0.4493913765093771,0.8054071046840616,0.87690423572744,0.32349726775956283
89
+ Twins_SVT,d,0.10364232180450968,0.9761081420936812,0.9843406998158379,0.7956989247311828
90
+ Twins_SVT,e,0.5988346448190649,0.70801317233809,0.8065163096949973,0.5266903914590747
91
+ Twins_SVT,f,0.3545602475236788,0.8442413445898846,0.9026489390789582,0.1283051582141309
92
+ Twins_SVT,g,0.2474187276363373,0.8956666666666667,0.9830134444444444,0.9040171726464274
93
+ Twins_SVT,h,0.273289222240448,0.8936666666666667,0.9851210555555556,0.9023569023569024
94
+ Twins_SVT,i,0.0899845923781395,0.9841666666666666,0.9985422222222222,0.9841428809881488
95
+ Twins_SVT,j,2.6232716159820555,0.4155,0.1803668888888889,0.03680307607800055
96
+ Twins_SVT,k,2.4658374714255333,0.504,0.3982547777777778,0.043086816720257236
97
+ Twins_SVT,l,1.027051387313659,0.7358151340489663,0.5966640795291248,0.5587352057940294
98
+ Twins_PCPVT,a,0.430052495628586,0.8000628733102798,0.8719401473296501,0.3041575492341357
99
+ Twins_PCPVT,b,0.3465015141235437,0.8560201194592896,0.9034383057090241,0.37771739130434784
100
+ Twins_PCPVT,c,0.4657778265396779,0.7752279157497642,0.857121546961326,0.2799597180261833
101
+ Twins_PCPVT,d,0.2879813692922751,0.9056900345803206,0.9308342541436465,0.4809688581314879
102
+ Twins_PCPVT,e,0.4774637080311906,0.7716794731064764,0.8551956406569287,0.5720164609053497
103
+ Twins_PCPVT,f,0.3843120121954025,0.8335527844473705,0.8887894780242434,0.11454470539761022
104
+ Twins_PCPVT,g,0.25515789008140566,0.9081666666666667,0.9717176666666667,0.9122751154274797
105
+ Twins_PCPVT,h,0.31839424538612365,0.8653333333333333,0.9604102222222224,0.8764148057509942
106
+ Twins_PCPVT,i,0.2241324300765991,0.9345,0.9829576666666666,0.9358157765801078
107
+ Twins_PCPVT,j,1.4245353891849517,0.49833333333333335,0.46813805555555554,0.21245421245421245
108
+ Twins_PCPVT,k,1.3935099244117737,0.5246666666666666,0.4594932222222222,0.22161572052401746
109
+ Twins_PCPVT,l,0.6877527087257096,0.7420548886891227,0.697346002300591,0.5830056419900838
110
+ PiT,a,0.45982032410006446,0.7887456774599183,0.924538674033149,0.3225806451612903
111
+ PiT,b,0.4223166468832261,0.8010059729644766,0.9304309392265193,0.3357817418677859
112
+ PiT,c,0.6344863635794392,0.7136120716755737,0.8947255985267035,0.25995125913891143
113
+ PiT,d,0.03659489378333092,0.9845960389814524,0.9966464088397791,0.8672086720867209
114
+ PiT,e,0.6467190528151494,0.7200878155872668,0.8809127374555362,0.5565217391304348
115
+ PiT,f,0.41419300931409125,0.8113236774843157,0.93339286411791,0.11611030478955008
116
+ PiT,g,0.22269588689506054,0.8966666666666666,0.9961491111111112,0.9061175045427013
117
+ PiT,h,0.3351811931580305,0.8503333333333334,0.9926552222222224,0.8695146759662888
118
+ PiT,i,0.018199062839150428,0.994,0.999871,0.9940199335548173
119
+ PiT,j,4.0110640263557436,0.4066666666666667,0.19481644444444443,0.028384279475982533
120
+ PiT,k,3.806567204385996,0.504,0.6367970000000001,0.033766233766233764
121
+ PiT,l,1.4887937803340863,0.7148749405108138,0.6321520235401981,0.5430508474576271
122
+ Ensemble,a,,0.9481295190191764,0.9716399631675874,0.656964656964657
123
+ Ensemble,b,,0.945300220056586,0.973926335174954,0.6448979591836734
124
+ Ensemble,c,,0.9022320025149324,0.9555580110497237,0.5039872408293461
125
+ Ensemble,d,,0.9905690034580321,0.9966666666666667,0.9132947976878613
126
+ Ensemble,e,,0.8737650933040615,0.9450768182850223,0.7331786542923434
127
+ Ensemble,f,,0.9455503059406708,0.9727634725471218,0.310107948969578
128
+ Ensemble,g,,0.9748333333333333,0.9997657777777778,0.9754511461550968
129
+ Ensemble,h,,0.952,0.9993787777777778,0.9541984732824428
130
+ Ensemble,i,,0.9988333333333334,0.999993888888889,0.9988346928583319
131
+ Ensemble,j,,0.4875,0.25661944444444446,0.04710257204834211
132
+ Ensemble,k,,0.5115,0.4922346666666667,0.04930262731106066
133
+ Ensemble,l,,0.8082068637301042,0.6421131491191425,0.6407132243684993
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce476482d4d0f9929d75db7f8312f2b6371bcea67a65e12167023292160907e1
3
+ size 343214864
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe0e7f22050c2d021328fb1298f921ac6caad55c4f068fc39fd1d23f44d89983
3
+ size 343259038
roc_confusion_matrix/ViT_roc_confusion_matrix_a.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_b.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_c.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_d.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_e.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_f.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_g.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_h.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_i.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_j.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_k.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_l.png ADDED
roc_curves/ViT_ROC_a.png ADDED
roc_curves/ViT_ROC_b.png ADDED
roc_curves/ViT_ROC_c.png ADDED
roc_curves/ViT_ROC_d.png ADDED
roc_curves/ViT_ROC_e.png ADDED
roc_curves/ViT_ROC_f.png ADDED
roc_curves/ViT_ROC_g.png ADDED
roc_curves/ViT_ROC_h.png ADDED
roc_curves/ViT_ROC_i.png ADDED
roc_curves/ViT_ROC_j.png ADDED
roc_curves/ViT_ROC_k.png ADDED
roc_curves/ViT_ROC_l.png ADDED
training_curves/ViT_accuracy.png ADDED
training_curves/ViT_auc.png ADDED
training_curves/ViT_combined_metrics.png ADDED

Git LFS Details

  • SHA256: a5dc2af2ab0931ccf282af1b53f3a2a355497ea00ebb2f82ddced0b9d58f287e
  • Pointer size: 131 Bytes
  • Size of remote file: 155 kB
training_curves/ViT_f1.png ADDED
training_curves/ViT_loss.png ADDED
training_curves/ViT_metrics.csv ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,train_loss,val_loss,train_accuracy,val_accuracy,train_auc,val_auc,train_f1,val_f1
2
+ 1,0.2992559493115094,0.20876842200756074,0.8865487674169347,0.919,0.9521798566667701,0.980348,0.8863599763809115,0.9218900675024108
3
+ 2,0.12608231064590994,0.1582107105255127,0.9475884244372991,0.937,0.9902657514914032,0.9894020000000001,0.947512477861858,0.938894277400582
4
+ 3,0.10436653757401984,0.12591355288028716,0.9578510182207931,0.954,0.9932662259373755,0.9909999999999999,0.9578927644082769,0.9536290322580645
5
+ 4,0.09696241902289283,0.11985185146331787,0.9608788853161844,0.952,0.9942321571553459,0.991154,0.9608285039708092,0.9524752475247524
6
+ 5,0.0849455009031909,0.1156887508481741,0.9664790996784566,0.957,0.9955087571468451,0.9942759999999999,0.9664764048556957,0.9579667644183774
7
+ 6,0.07620173852255874,0.09895226681232452,0.970016077170418,0.966,0.996354714131942,0.99409,0.9699975869372872,0.9657258064516129
8
+ 7,0.07403722440290873,0.12703941571712493,0.9697481243301179,0.949,0.9965995118835506,0.995868,0.9697278455557046,0.9511025886864813
9
+ 8,0.06730379444993194,0.10513309073448181,0.9730171489817792,0.956,0.9971600100633093,0.9936699999999999,0.9730294346091009,0.9566929133858267
10
+ 9,0.06610564867879033,0.07069086188450456,0.9747320471596999,0.975,0.9972713986046924,0.997084,0.9747476099938409,0.9747729566094854
11
+ 10,0.0625152113479098,0.12434254777431487,0.9758574490889603,0.957,0.9974816941971697,0.9933960000000001,0.9758386742109356,0.9577187807276303
12
+ 11,0.06051965755473379,0.09611465513706208,0.9763933547695606,0.971,0.9976965044934514,0.9939479999999999,0.9763851287962044,0.971201588877855
13
+ 12,0.05805372083829147,0.07930001637339593,0.9771436227224009,0.969,0.9978529768900009,0.9960419999999999,0.9771466845277964,0.9690927218344965
14
+ 13,0.05441189735672673,0.07749747556447983,0.980064308681672,0.971,0.9981060429827373,0.9961779999999999,0.9800696490758103,0.971028971028971
15
+ 14,0.05495417037098354,0.0988565045595169,0.9785369774919614,0.967,0.9980853563341985,0.9964459999999999,0.9785248934287782,0.967551622418879
16
+ 15,0.05151973974642455,0.06946688985824585,0.9796891747052519,0.974,0.99831483798187,0.9978020000000001,0.9796859087741866,0.9743589743589743
17
+ 16,0.05302145266456236,0.09198650085926056,0.980117899249732,0.972,0.9982026036515108,0.996364,0.9801104379992495,0.9715447154471545
18
+ 17,0.05111499030286764,0.1082226443439722,0.9811897106109325,0.973,0.9983336621714921,0.99519,0.9811866859623734,0.972972972972973
19
+ 18,0.05241213196535202,0.1071859257221222,0.9806002143622723,0.965,0.9982049974209898,0.995126,0.9806002143622723,0.9656526005888125
20
+ 19,0.04990142801734604,0.07836620014905929,0.9808949624866024,0.97,0.9983857636340034,0.9972000000000001,0.9808964981378775,0.9703557312252964
21
+ 20,0.046595271926528965,0.0676819824874401,0.9821811361200429,0.972,0.9986023229472171,0.99761,0.9821816135687682,0.9722222222222222
22
+ 21,0.04662280432951795,0.07094095158576966,0.9825562700964631,0.973,0.9985773829422313,0.997676,0.982560475769509,0.972972972972973
23
+ 22,0.04306263393120559,0.0786460086107254,0.9832797427652733,0.972,0.9988035704644169,0.997056,0.9832878033103005,0.9719438877755511
24
+ 23,0.044755603147856295,0.11225838279724121,0.9831189710610932,0.959,0.9987064081223314,0.995554,0.9831253013339047,0.9601554907677357
25
+ 24,0.043202851166487505,0.08695128572359681,0.9837084673097535,0.966,0.998731348127317,0.996696,0.9837145765254192,0.9661354581673307
26
+ 25,0.044267977352502645,0.07391600479185581,0.9836012861736334,0.976,0.9986831194880118,0.9976659999999999,0.9836091917081793,0.9762376237623762
27
+ 26,0.039225034822688895,0.09814529120922089,0.984914255091104,0.968,0.9989991846536832,0.9954779999999999,0.9849162759544541,0.9678714859437751
28
+ 27,0.029010193963861543,0.07500301361083984,0.9891747052518757,0.975,0.9994347056654364,0.997734,0.9891747052518757,0.9749247743229689
29
+ 28,0.02800916622713256,0.07024864779412747,0.989871382636656,0.977,0.9994683189563568,0.99807,0.9898757231626313,0.9769307923771314
30
+ 29,0.025017252657861956,0.06947895266115665,0.9904876741693462,0.975,0.9995947076872884,0.998088,0.9904925145290447,0.9749247743229689
31
+ 30,0.02682399880739941,0.06291172051429748,0.9898177920685959,0.977,0.999542028963031,0.998342,0.9898188832922517,0.977022977022977
32
+ 31,0.0245672710880733,0.07222659438848496,0.9910503751339764,0.976,0.999612209344403,0.998126,0.9910522931847406,0.9760479041916168
33
+ 32,0.022775541941584404,0.06732542610354722,0.9907824222936763,0.978,0.9996702040462314,0.9983400000000001,0.9907878528198811,0.9777327935222672
34
+ 33,0.02264225509532778,0.0671888653114438,0.9913719185423365,0.977,0.9996683013800289,0.9982840000000001,0.991372380901345,0.9767911200807265
35
+ 34,0.022598851041514965,0.06741214002668858,0.9916666666666667,0.981,0.9996392731556631,0.997224,0.99166956847829,0.981018981018981
36
+ 35,0.02185062739003893,0.062260420559905466,0.9919078242229368,0.982,0.9996926023763655,0.998424,0.99190652302085,0.9818913480885312
37
+ 36,0.021898518254758845,0.08399707531929015,0.9919346195069668,0.978,0.9996654480987122,0.9980199999999999,0.9919322415502961,0.9776422764227642
38
+ 37,0.022006946835079953,0.06288080751895904,0.9920150053590568,0.979,0.999680801537986,0.998356,0.9920154332565243,0.9789368104312939
39
+ 38,0.02184967446467357,0.07284429347515106,0.9915862808145767,0.974,0.9996958534226165,0.998204,0.9915894359029302,0.9738955823293173
40
+ 39,0.02177039008928984,0.06998497086763382,0.9914255091103966,0.977,0.9996905877041525,0.998366,0.9914273467638234,0.9767911200807265
41
+ 40,0.019916062515791973,0.06419473651051522,0.992658092175777,0.977,0.9997414197652124,0.998528,0.9926576986976794,0.9767911200807265
42
+ 41,0.02011498119382613,0.07022419507056475,0.9926045016077171,0.978,0.9997390346115803,0.998282,0.9926072748700916,0.9779116465863453
43
+ 42,0.021137025901430864,0.06808076894283295,0.9916398713826367,0.978,0.9997082760379509,0.99831,0.9916291049581455,0.9779116465863453
44
+ 43,0.018708350665424128,0.06744997225701808,0.9931939978563773,0.98,0.9997793984185901,0.9983439999999999,0.9931918087273507,0.9799196787148594
45
+ 44,0.018468665781691432,0.06981276272237301,0.9930600214362272,0.98,0.9997648376372361,0.998352,0.9930628096959957,0.9798387096774194
46
+ 45,0.01840671534108004,0.06784664452075959,0.9929796355841372,0.98,0.9997877586620852,0.9983979999999999,0.992978883052846,0.9798792756539235
47
+ 46,0.01743951269284705,0.07026282000541687,0.9936227224008575,0.981,0.9997826911081014,0.998418,0.9936220388037303,0.9808274470232089
48
+ 47,0.019610805441760172,0.06931583965010941,0.9930868167202572,0.98,0.999745398850531,0.998354,0.9930894091176943,0.9798792756539235
49
+ 48,0.017864001673194085,0.06890859639644623,0.9933011789924974,0.979,0.9998022016935308,0.998378,0.993299383543286,0.978894472361809
50
+ 49,0.018858803294215746,0.06938542400300503,0.9928992497320471,0.981,0.9997645734179295,0.9983660000000001,0.9928971561821545,0.9808660624370594
51
+ 50,0.016824277843142055,0.06828806474804879,0.9933815648445874,0.981,0.9998234024209381,0.9983879999999999,0.9933789036322209,0.9808660624370594
52
+ 51,0.017368377836187554,0.07072594451904297,0.9936763129689175,0.982,0.999803878911738,0.998382,0.9936773294754326,0.9818548387096774
53
+ 52,0.01653399652549907,0.07108432674780488,0.9936763129689175,0.981,0.9998266003361329,0.998364,0.993675635116304,0.9808660624370594
54
+ 53,0.018573750247669756,0.0684389100074768,0.9928992497320471,0.981,0.9997809679387104,0.99839,0.9929002009377094,0.9808660624370594
55
+ 54,0.01766859353534949,0.06880642592906952,0.9936495176848874,0.981,0.9997677584093538,0.99842,0.9936503683858005,0.9808660624370594
56
+ 55,0.016026736973182947,0.07103883915767073,0.9938638799571275,0.982,0.9998384629214154,0.998364,0.9938630577515745,0.9818913480885312
training_metrics.csv ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,train_loss,val_loss,train_accuracy,val_accuracy,train_auc,val_auc,train_f1,val_f1
2
+ 1,0.2992559493115094,0.20876842200756074,0.8865487674169347,0.919,0.9521798566667701,0.980348,0.8863599763809115,0.9218900675024108
3
+ 2,0.12608231064590994,0.1582107105255127,0.9475884244372991,0.937,0.9902657514914032,0.9894020000000001,0.947512477861858,0.938894277400582
4
+ 3,0.10436653757401984,0.12591355288028716,0.9578510182207931,0.954,0.9932662259373755,0.9909999999999999,0.9578927644082769,0.9536290322580645
5
+ 4,0.09696241902289283,0.11985185146331787,0.9608788853161844,0.952,0.9942321571553459,0.991154,0.9608285039708092,0.9524752475247524
6
+ 5,0.0849455009031909,0.1156887508481741,0.9664790996784566,0.957,0.9955087571468451,0.9942759999999999,0.9664764048556957,0.9579667644183774
7
+ 6,0.07620173852255874,0.09895226681232452,0.970016077170418,0.966,0.996354714131942,0.99409,0.9699975869372872,0.9657258064516129
8
+ 7,0.07403722440290873,0.12703941571712493,0.9697481243301179,0.949,0.9965995118835506,0.995868,0.9697278455557046,0.9511025886864813
9
+ 8,0.06730379444993194,0.10513309073448181,0.9730171489817792,0.956,0.9971600100633093,0.9936699999999999,0.9730294346091009,0.9566929133858267
10
+ 9,0.06610564867879033,0.07069086188450456,0.9747320471596999,0.975,0.9972713986046924,0.997084,0.9747476099938409,0.9747729566094854
11
+ 10,0.0625152113479098,0.12434254777431487,0.9758574490889603,0.957,0.9974816941971697,0.9933960000000001,0.9758386742109356,0.9577187807276303
12
+ 11,0.06051965755473379,0.09611465513706208,0.9763933547695606,0.971,0.9976965044934514,0.9939479999999999,0.9763851287962044,0.971201588877855
13
+ 12,0.05805372083829147,0.07930001637339593,0.9771436227224009,0.969,0.9978529768900009,0.9960419999999999,0.9771466845277964,0.9690927218344965
14
+ 13,0.05441189735672673,0.07749747556447983,0.980064308681672,0.971,0.9981060429827373,0.9961779999999999,0.9800696490758103,0.971028971028971
15
+ 14,0.05495417037098354,0.0988565045595169,0.9785369774919614,0.967,0.9980853563341985,0.9964459999999999,0.9785248934287782,0.967551622418879
16
+ 15,0.05151973974642455,0.06946688985824585,0.9796891747052519,0.974,0.99831483798187,0.9978020000000001,0.9796859087741866,0.9743589743589743
17
+ 16,0.05302145266456236,0.09198650085926056,0.980117899249732,0.972,0.9982026036515108,0.996364,0.9801104379992495,0.9715447154471545
18
+ 17,0.05111499030286764,0.1082226443439722,0.9811897106109325,0.973,0.9983336621714921,0.99519,0.9811866859623734,0.972972972972973
19
+ 18,0.05241213196535202,0.1071859257221222,0.9806002143622723,0.965,0.9982049974209898,0.995126,0.9806002143622723,0.9656526005888125
20
+ 19,0.04990142801734604,0.07836620014905929,0.9808949624866024,0.97,0.9983857636340034,0.9972000000000001,0.9808964981378775,0.9703557312252964
21
+ 20,0.046595271926528965,0.0676819824874401,0.9821811361200429,0.972,0.9986023229472171,0.99761,0.9821816135687682,0.9722222222222222
22
+ 21,0.04662280432951795,0.07094095158576966,0.9825562700964631,0.973,0.9985773829422313,0.997676,0.982560475769509,0.972972972972973
23
+ 22,0.04306263393120559,0.0786460086107254,0.9832797427652733,0.972,0.9988035704644169,0.997056,0.9832878033103005,0.9719438877755511
24
+ 23,0.044755603147856295,0.11225838279724121,0.9831189710610932,0.959,0.9987064081223314,0.995554,0.9831253013339047,0.9601554907677357
25
+ 24,0.043202851166487505,0.08695128572359681,0.9837084673097535,0.966,0.998731348127317,0.996696,0.9837145765254192,0.9661354581673307
26
+ 25,0.044267977352502645,0.07391600479185581,0.9836012861736334,0.976,0.9986831194880118,0.9976659999999999,0.9836091917081793,0.9762376237623762
27
+ 26,0.039225034822688895,0.09814529120922089,0.984914255091104,0.968,0.9989991846536832,0.9954779999999999,0.9849162759544541,0.9678714859437751
28
+ 27,0.029010193963861543,0.07500301361083984,0.9891747052518757,0.975,0.9994347056654364,0.997734,0.9891747052518757,0.9749247743229689
29
+ 28,0.02800916622713256,0.07024864779412747,0.989871382636656,0.977,0.9994683189563568,0.99807,0.9898757231626313,0.9769307923771314
30
+ 29,0.025017252657861956,0.06947895266115665,0.9904876741693462,0.975,0.9995947076872884,0.998088,0.9904925145290447,0.9749247743229689
31
+ 30,0.02682399880739941,0.06291172051429748,0.9898177920685959,0.977,0.999542028963031,0.998342,0.9898188832922517,0.977022977022977
32
+ 31,0.0245672710880733,0.07222659438848496,0.9910503751339764,0.976,0.999612209344403,0.998126,0.9910522931847406,0.9760479041916168
33
+ 32,0.022775541941584404,0.06732542610354722,0.9907824222936763,0.978,0.9996702040462314,0.9983400000000001,0.9907878528198811,0.9777327935222672
34
+ 33,0.02264225509532778,0.0671888653114438,0.9913719185423365,0.977,0.9996683013800289,0.9982840000000001,0.991372380901345,0.9767911200807265
35
+ 34,0.022598851041514965,0.06741214002668858,0.9916666666666667,0.981,0.9996392731556631,0.997224,0.99166956847829,0.981018981018981
36
+ 35,0.02185062739003893,0.062260420559905466,0.9919078242229368,0.982,0.9996926023763655,0.998424,0.99190652302085,0.9818913480885312
37
+ 36,0.021898518254758845,0.08399707531929015,0.9919346195069668,0.978,0.9996654480987122,0.9980199999999999,0.9919322415502961,0.9776422764227642
38
+ 37,0.022006946835079953,0.06288080751895904,0.9920150053590568,0.979,0.999680801537986,0.998356,0.9920154332565243,0.9789368104312939
39
+ 38,0.02184967446467357,0.07284429347515106,0.9915862808145767,0.974,0.9996958534226165,0.998204,0.9915894359029302,0.9738955823293173
40
+ 39,0.02177039008928984,0.06998497086763382,0.9914255091103966,0.977,0.9996905877041525,0.998366,0.9914273467638234,0.9767911200807265
41
+ 40,0.019916062515791973,0.06419473651051522,0.992658092175777,0.977,0.9997414197652124,0.998528,0.9926576986976794,0.9767911200807265
42
+ 41,0.02011498119382613,0.07022419507056475,0.9926045016077171,0.978,0.9997390346115803,0.998282,0.9926072748700916,0.9779116465863453
43
+ 42,0.021137025901430864,0.06808076894283295,0.9916398713826367,0.978,0.9997082760379509,0.99831,0.9916291049581455,0.9779116465863453
44
+ 43,0.018708350665424128,0.06744997225701808,0.9931939978563773,0.98,0.9997793984185901,0.9983439999999999,0.9931918087273507,0.9799196787148594
45
+ 44,0.018468665781691432,0.06981276272237301,0.9930600214362272,0.98,0.9997648376372361,0.998352,0.9930628096959957,0.9798387096774194
46
+ 45,0.01840671534108004,0.06784664452075959,0.9929796355841372,0.98,0.9997877586620852,0.9983979999999999,0.992978883052846,0.9798792756539235
47
+ 46,0.01743951269284705,0.07026282000541687,0.9936227224008575,0.981,0.9997826911081014,0.998418,0.9936220388037303,0.9808274470232089
48
+ 47,0.019610805441760172,0.06931583965010941,0.9930868167202572,0.98,0.999745398850531,0.998354,0.9930894091176943,0.9798792756539235
49
+ 48,0.017864001673194085,0.06890859639644623,0.9933011789924974,0.979,0.9998022016935308,0.998378,0.993299383543286,0.978894472361809
50
+ 49,0.018858803294215746,0.06938542400300503,0.9928992497320471,0.981,0.9997645734179295,0.9983660000000001,0.9928971561821545,0.9808660624370594
51
+ 50,0.016824277843142055,0.06828806474804879,0.9933815648445874,0.981,0.9998234024209381,0.9983879999999999,0.9933789036322209,0.9808660624370594
52
+ 51,0.017368377836187554,0.07072594451904297,0.9936763129689175,0.982,0.999803878911738,0.998382,0.9936773294754326,0.9818548387096774
53
+ 52,0.01653399652549907,0.07108432674780488,0.9936763129689175,0.981,0.9998266003361329,0.998364,0.993675635116304,0.9808660624370594
54
+ 53,0.018573750247669756,0.0684389100074768,0.9928992497320471,0.981,0.9997809679387104,0.99839,0.9929002009377094,0.9808660624370594
55
+ 54,0.01766859353534949,0.06880642592906952,0.9936495176848874,0.981,0.9997677584093538,0.99842,0.9936503683858005,0.9808660624370594
56
+ 55,0.016026736973182947,0.07103883915767073,0.9938638799571275,0.982,0.9998384629214154,0.998364,0.9938630577515745,0.9818913480885312
training_notebook_s2.ipynb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e73a8ba21cc96bb4ae07864a3b6ce6808bc99f75aa44ab247d52de389940a08
3
+ size 21054875