Update README.md
Browse files
README.md
CHANGED
@@ -33,14 +33,14 @@ The model is trained using the following datasets:
|
|
33 |
|
34 |
2. **Open Human Facial Emotion Recognition Dataset**:
|
35 |
- Contains nearly 10,000 images with emotion labels gathered from wild scenes to enhance the model's capability in facial emotion recognition.
|
36 |
-
|
37 |
-
3.**SFEW
|
38 |
- The Static Facial Expressions in the Wild (SFEW) dataset is a dataset for facial expression recognition. It is created by selecting static frames from the AFEW database by computing keyframes based on facial point clustering.
|
39 |
-
|
40 |
-
4.**Neutral add
|
41 |
- Contains 50K images without obvious emotional fluctuations as a supplementary category.
|
42 |
|
43 |
-
|
44 |
## training method
|
45 |
|
46 |
Combining the fine-tuning methods of layer_norm tuning, prefix tuning, and prompt tuning, the practical results show that the mixture of the three training methods can be comparable to or even exceed the performance of full fine-tuning in generalized visual emotion recognition by introducing only a small number of parameters. In addition, thanks to the adjustment of layer norm, it converges faster than prefix tuning and prompt tuning, achieving higher performance than EmotionCLIP-V1.
|
|
|
33 |
|
34 |
2. **Open Human Facial Emotion Recognition Dataset**:
|
35 |
- Contains nearly 10,000 images with emotion labels gathered from wild scenes to enhance the model's capability in facial emotion recognition.
|
36 |
+
|
37 |
+
3.**SFEW**:
|
38 |
- The Static Facial Expressions in the Wild (SFEW) dataset is a dataset for facial expression recognition. It is created by selecting static frames from the AFEW database by computing keyframes based on facial point clustering.
|
39 |
+
|
40 |
+
4.**Neutral add**:
|
41 |
- Contains 50K images without obvious emotional fluctuations as a supplementary category.
|
42 |
|
43 |
+
|
44 |
## training method
|
45 |
|
46 |
Combining the fine-tuning methods of layer_norm tuning, prefix tuning, and prompt tuning, the practical results show that the mixture of the three training methods can be comparable to or even exceed the performance of full fine-tuning in generalized visual emotion recognition by introducing only a small number of parameters. In addition, thanks to the adjustment of layer norm, it converges faster than prefix tuning and prompt tuning, achieving higher performance than EmotionCLIP-V1.
|