Update README.md
Browse files
README.md
CHANGED
@@ -12,7 +12,9 @@ tags:
|
|
12 |
- EmoSet
|
13 |
---
|
14 |
|
15 |
-
|
|
|
|
|
16 |
|
17 |
## Training Details
|
18 |
|
@@ -23,7 +25,8 @@ PreceptCLIP-Emotions is a model designed to predict the emotions that an image e
|
|
23 |
- *Learning Rate*: 0.0001
|
24 |
- *Batch Size*: 32
|
25 |
|
26 |
-
## Requirements
|
|
|
27 |
- python=3.9.15
|
28 |
- cudatoolkit=11.7
|
29 |
- torchvision=0.14.0
|
|
|
12 |
- EmoSet
|
13 |
---
|
14 |
|
15 |
+
**PerceptCLIP-Emotions** is a model designed to predict the **emotions** that an image evokes in users. This is the official model from the paper:
|
16 |
+
📄 **["Don't Judge Before You CLIP: A Unified Approach for Perceptual Tasks"](https://arxiv.org/abs/2503.13260)**
|
17 |
+
We apply **LoRA adaptation** on the **CLIP visual encoder** and add an **MLP head** for emotion classification. Our model achieves **state-of-the-art results**.
|
18 |
|
19 |
## Training Details
|
20 |
|
|
|
25 |
- *Learning Rate*: 0.0001
|
26 |
- *Batch Size*: 32
|
27 |
|
28 |
+
## Installation & Requirements
|
29 |
+
You can set up the environment using environment.yml or manually install dependencies:
|
30 |
- python=3.9.15
|
31 |
- cudatoolkit=11.7
|
32 |
- torchvision=0.14.0
|