End of training
Browse files- README.md +14 -15
- model.safetensors +1 -1
- runs/Mar11_18-54-36_n28/events.out.tfevents.1741715709.n28.1412156.0 +3 -0
- runs/Mar13_16-35-17_n31/events.out.tfevents.1741880144.n31.3016205.0 +3 -0
- runs/Mar13_16-52-00_n31/events.out.tfevents.1741881127.n31.3016205.1 +3 -0
- runs/Mar13_16-53-33_n31/events.out.tfevents.1741881219.n31.3016205.2 +3 -0
- runs/Mar13_16-53-52_n31/events.out.tfevents.1741881234.n31.3016205.3 +3 -0
- runs/Mar13_17-09-18_n31/events.out.tfevents.1741882161.n31.3018265.0 +3 -0
- runs/Mar13_17-09-18_n31/events.out.tfevents.1741961103.n31.3018265.1 +3 -0
- training_args.bin +1 -1
README.md
CHANGED
@@ -20,14 +20,14 @@ should probably proofread and complete it, then remove this comment. -->
|
|
20 |
|
21 |
This model is a fine-tuned version of [xlm-roberta-large](https://huggingface.co/xlm-roberta-large) on the generator dataset.
|
22 |
It achieves the following results on the evaluation set:
|
23 |
-
- Loss:
|
24 |
-
- Accuracy: 0.
|
25 |
-
- Micro Precision: 0.
|
26 |
-
- Micro Recall: 0.
|
27 |
-
- Micro F1: 0.
|
28 |
-
- Macro Precision: 0.
|
29 |
-
- Macro Recall: 0.
|
30 |
-
- Macro F1: 0.
|
31 |
|
32 |
## Model description
|
33 |
|
@@ -46,7 +46,7 @@ More information needed
|
|
46 |
### Training hyperparameters
|
47 |
|
48 |
The following hyperparameters were used during training:
|
49 |
-
- learning_rate:
|
50 |
- train_batch_size: 12
|
51 |
- eval_batch_size: 12
|
52 |
- seed: 42
|
@@ -55,17 +55,16 @@ The following hyperparameters were used during training:
|
|
55 |
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
56 |
- lr_scheduler_type: linear
|
57 |
- lr_scheduler_warmup_steps: 500
|
58 |
-
- num_epochs:
|
59 |
|
60 |
### Training results
|
61 |
|
62 |
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Micro Precision | Micro Recall | Micro F1 | Macro Precision | Macro Recall | Macro F1 |
|
63 |
|:-------------:|:------:|:------:|:---------------:|:--------:|:---------------:|:------------:|:--------:|:---------------:|:------------:|:--------:|
|
64 |
-
|
|
65 |
-
|
|
66 |
-
|
|
67 |
-
|
|
68 |
-
| 0.1038 | 4.9999 | 200830 | 0.5242 | 0.9087 | 0.9087 | 0.9087 | 0.9087 | 0.8753 | 0.8646 | 0.8660 |
|
69 |
|
70 |
|
71 |
### Framework versions
|
|
|
20 |
|
21 |
This model is a fine-tuned version of [xlm-roberta-large](https://huggingface.co/xlm-roberta-large) on the generator dataset.
|
22 |
It achieves the following results on the evaluation set:
|
23 |
+
- Loss: 3.4922
|
24 |
+
- Accuracy: 0.0913
|
25 |
+
- Micro Precision: 0.0913
|
26 |
+
- Micro Recall: 0.0913
|
27 |
+
- Micro F1: 0.0913
|
28 |
+
- Macro Precision: 0.0017
|
29 |
+
- Macro Recall: 0.0162
|
30 |
+
- Macro F1: 0.0030
|
31 |
|
32 |
## Model description
|
33 |
|
|
|
46 |
### Training hyperparameters
|
47 |
|
48 |
The following hyperparameters were used during training:
|
49 |
+
- learning_rate: 0.0001
|
50 |
- train_batch_size: 12
|
51 |
- eval_batch_size: 12
|
52 |
- seed: 42
|
|
|
55 |
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
56 |
- lr_scheduler_type: linear
|
57 |
- lr_scheduler_warmup_steps: 500
|
58 |
+
- num_epochs: 4
|
59 |
|
60 |
### Training results
|
61 |
|
62 |
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Micro Precision | Micro Recall | Micro F1 | Macro Precision | Macro Recall | Macro F1 |
|
63 |
|:-------------:|:------:|:------:|:---------------:|:--------:|:---------------:|:------------:|:--------:|:---------------:|:------------:|:--------:|
|
64 |
+
| 3.3304 | 1.0000 | 40166 | 3.4135 | 0.0955 | 0.0955 | 0.0955 | 0.0955 | 0.0035 | 0.0183 | 0.0048 |
|
65 |
+
| 3.3 | 2.0 | 80333 | 5.1442 | 0.0124 | 0.0124 | 0.0124 | 0.0124 | 0.0002 | 0.0156 | 0.0004 |
|
66 |
+
| 3.2756 | 3.0000 | 120499 | 4.6440 | 0.0124 | 0.0124 | 0.0124 | 0.0124 | 0.0002 | 0.0156 | 0.0004 |
|
67 |
+
| 3.2737 | 4.0000 | 160664 | 5.3371 | 0.0124 | 0.0124 | 0.0124 | 0.0124 | 0.0002 | 0.0156 | 0.0004 |
|
|
|
68 |
|
69 |
|
70 |
### Framework versions
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 2240416848
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d7b67b03482a444a6daa39318212c221e5690325a79dbb60156b2a74d2c3a608
|
3 |
size 2240416848
|
runs/Mar11_18-54-36_n28/events.out.tfevents.1741715709.n28.1412156.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6c609a82739d14be24f4a16f92e9e341aff95b6c70ede7828f8b0d30023f63fd
|
3 |
+
size 40765
|
runs/Mar13_16-35-17_n31/events.out.tfevents.1741880144.n31.3016205.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:49561005f3669b6f82104c9e02ffaa6b446ca4816f0b2b1147096f890aea9dfc
|
3 |
+
size 9849
|
runs/Mar13_16-52-00_n31/events.out.tfevents.1741881127.n31.3016205.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:19b5e3ee2c21ee763ac6461b3038c7f9d773efc11f9f7e7347b30e9f98a13b68
|
3 |
+
size 9851
|
runs/Mar13_16-53-33_n31/events.out.tfevents.1741881219.n31.3016205.2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a72d64c58d9ec32ba5426780075fdfe0388b3dd6d7102ad10cf5274f028e8274
|
3 |
+
size 8280
|
runs/Mar13_16-53-52_n31/events.out.tfevents.1741881234.n31.3016205.3
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:460780e66daac589de062f7ad10d1e27e7888e0a481580a48b3e9544630ebe6a
|
3 |
+
size 8280
|
runs/Mar13_17-09-18_n31/events.out.tfevents.1741882161.n31.3018265.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:82620d42d4e1f0e9564d52916d6c5b81774059fbc7311d0607aec7641f25ad9e
|
3 |
+
size 81773
|
runs/Mar13_17-09-18_n31/events.out.tfevents.1741961103.n31.3018265.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:139386164b25cc8c2490da0d5ccb8bd8ee3d97710a52f23a166be14ff3e25d7c
|
3 |
+
size 757
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5304
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:30931b0a55768477150d657db26b7166ee6107d4cacc0abcb309a6f979644aca
|
3 |
size 5304
|