stulcrad commited on
Commit
15ace24
·
verified ·
1 Parent(s): 905f979

End of training

Browse files
README.md CHANGED
@@ -20,14 +20,14 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [xlm-roberta-large](https://huggingface.co/xlm-roberta-large) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 0.4935
24
- - Accuracy: 0.9044
25
- - Micro Precision: 0.9044
26
- - Micro Recall: 0.9044
27
- - Micro F1: 0.9044
28
- - Macro Precision: 0.8822
29
- - Macro Recall: 0.8527
30
- - Macro F1: 0.8599
31
 
32
  ## Model description
33
 
@@ -46,7 +46,7 @@ More information needed
46
  ### Training hyperparameters
47
 
48
  The following hyperparameters were used during training:
49
- - learning_rate: 1e-05
50
  - train_batch_size: 12
51
  - eval_batch_size: 12
52
  - seed: 42
@@ -55,17 +55,16 @@ The following hyperparameters were used during training:
55
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
56
  - lr_scheduler_type: linear
57
  - lr_scheduler_warmup_steps: 500
58
- - num_epochs: 5
59
 
60
  ### Training results
61
 
62
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | Micro Precision | Micro Recall | Micro F1 | Macro Precision | Macro Recall | Macro F1 |
63
  |:-------------:|:------:|:------:|:---------------:|:--------:|:---------------:|:------------:|:--------:|:---------------:|:------------:|:--------:|
64
- | 0.3304 | 1.0000 | 40166 | 0.3911 | 0.8883 | 0.8883 | 0.8883 | 0.8883 | 0.8651 | 0.8285 | 0.8359 |
65
- | 0.2376 | 2.0 | 80333 | 0.3998 | 0.9014 | 0.9014 | 0.9014 | 0.9014 | 0.8714 | 0.8480 | 0.8523 |
66
- | 0.1733 | 3.0000 | 120499 | 0.4297 | 0.9059 | 0.9059 | 0.9059 | 0.9059 | 0.8718 | 0.8607 | 0.8610 |
67
- | 0.1268 | 4.0 | 160666 | 0.4822 | 0.9088 | 0.9088 | 0.9088 | 0.9088 | 0.8807 | 0.8608 | 0.8649 |
68
- | 0.1038 | 4.9999 | 200830 | 0.5242 | 0.9087 | 0.9087 | 0.9087 | 0.9087 | 0.8753 | 0.8646 | 0.8660 |
69
 
70
 
71
  ### Framework versions
 
20
 
21
  This model is a fine-tuned version of [xlm-roberta-large](https://huggingface.co/xlm-roberta-large) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 3.4922
24
+ - Accuracy: 0.0913
25
+ - Micro Precision: 0.0913
26
+ - Micro Recall: 0.0913
27
+ - Micro F1: 0.0913
28
+ - Macro Precision: 0.0017
29
+ - Macro Recall: 0.0162
30
+ - Macro F1: 0.0030
31
 
32
  ## Model description
33
 
 
46
  ### Training hyperparameters
47
 
48
  The following hyperparameters were used during training:
49
+ - learning_rate: 0.0001
50
  - train_batch_size: 12
51
  - eval_batch_size: 12
52
  - seed: 42
 
55
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
56
  - lr_scheduler_type: linear
57
  - lr_scheduler_warmup_steps: 500
58
+ - num_epochs: 4
59
 
60
  ### Training results
61
 
62
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | Micro Precision | Micro Recall | Micro F1 | Macro Precision | Macro Recall | Macro F1 |
63
  |:-------------:|:------:|:------:|:---------------:|:--------:|:---------------:|:------------:|:--------:|:---------------:|:------------:|:--------:|
64
+ | 3.3304 | 1.0000 | 40166 | 3.4135 | 0.0955 | 0.0955 | 0.0955 | 0.0955 | 0.0035 | 0.0183 | 0.0048 |
65
+ | 3.3 | 2.0 | 80333 | 5.1442 | 0.0124 | 0.0124 | 0.0124 | 0.0124 | 0.0002 | 0.0156 | 0.0004 |
66
+ | 3.2756 | 3.0000 | 120499 | 4.6440 | 0.0124 | 0.0124 | 0.0124 | 0.0124 | 0.0002 | 0.0156 | 0.0004 |
67
+ | 3.2737 | 4.0000 | 160664 | 5.3371 | 0.0124 | 0.0124 | 0.0124 | 0.0124 | 0.0002 | 0.0156 | 0.0004 |
 
68
 
69
 
70
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:beb92342efa5ce3788ec5a2100c32f556cd32c6fea6e76ae777474e01fed902b
3
  size 2240416848
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7b67b03482a444a6daa39318212c221e5690325a79dbb60156b2a74d2c3a608
3
  size 2240416848
runs/Mar11_18-54-36_n28/events.out.tfevents.1741715709.n28.1412156.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c609a82739d14be24f4a16f92e9e341aff95b6c70ede7828f8b0d30023f63fd
3
+ size 40765
runs/Mar13_16-35-17_n31/events.out.tfevents.1741880144.n31.3016205.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49561005f3669b6f82104c9e02ffaa6b446ca4816f0b2b1147096f890aea9dfc
3
+ size 9849
runs/Mar13_16-52-00_n31/events.out.tfevents.1741881127.n31.3016205.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19b5e3ee2c21ee763ac6461b3038c7f9d773efc11f9f7e7347b30e9f98a13b68
3
+ size 9851
runs/Mar13_16-53-33_n31/events.out.tfevents.1741881219.n31.3016205.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a72d64c58d9ec32ba5426780075fdfe0388b3dd6d7102ad10cf5274f028e8274
3
+ size 8280
runs/Mar13_16-53-52_n31/events.out.tfevents.1741881234.n31.3016205.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:460780e66daac589de062f7ad10d1e27e7888e0a481580a48b3e9544630ebe6a
3
+ size 8280
runs/Mar13_17-09-18_n31/events.out.tfevents.1741882161.n31.3018265.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82620d42d4e1f0e9564d52916d6c5b81774059fbc7311d0607aec7641f25ad9e
3
+ size 81773
runs/Mar13_17-09-18_n31/events.out.tfevents.1741961103.n31.3018265.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:139386164b25cc8c2490da0d5ccb8bd8ee3d97710a52f23a166be14ff3e25d7c
3
+ size 757
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b2476dc9ac594194dc0989969d790e0761fea836a1e19e5e237763dbf6ff5f5b
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30931b0a55768477150d657db26b7166ee6107d4cacc0abcb309a6f979644aca
3
  size 5304