Amitz244 commited on
Commit
7c9a15c
·
verified ·
1 Parent(s): bc4acdc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -2
README.md CHANGED
@@ -12,7 +12,9 @@ tags:
12
  - EmoSet
13
  ---
14
 
15
- PreceptCLIP-Emotions is a model designed to predict the emotions that an image evokes in users. This is the official model from the paper ["Don't Judge Before You CLIP: A Unified Approach for Perceptual Tasks"](https://arxiv.org/abs/2503.13260). We apply LoRA adaptation on the CLIP visual encoder with an additional MLP head. Our model *achieves state-of-the-art results*.
 
 
16
 
17
  ## Training Details
18
 
@@ -23,7 +25,8 @@ PreceptCLIP-Emotions is a model designed to predict the emotions that an image e
23
  - *Learning Rate*: 0.0001
24
  - *Batch Size*: 32
25
 
26
- ## Requirements
 
27
  - python=3.9.15
28
  - cudatoolkit=11.7
29
  - torchvision=0.14.0
 
12
  - EmoSet
13
  ---
14
 
15
+ **PerceptCLIP-Emotions** is a model designed to predict the **emotions** that an image evokes in users. This is the official model from the paper:
16
+ 📄 **["Don't Judge Before You CLIP: A Unified Approach for Perceptual Tasks"](https://arxiv.org/abs/2503.13260)**
17
+ We apply **LoRA adaptation** on the **CLIP visual encoder** and add an **MLP head** for emotion classification. Our model achieves **state-of-the-art results**.
18
 
19
  ## Training Details
20
 
 
25
  - *Learning Rate*: 0.0001
26
  - *Batch Size*: 32
27
 
28
+ ## Installation & Requirements
29
+ You can set up the environment using environment.yml or manually install dependencies:
30
  - python=3.9.15
31
  - cudatoolkit=11.7
32
  - torchvision=0.14.0