Update README.md
Browse files
README.md
CHANGED
@@ -3,7 +3,29 @@ library_name: peft
|
|
3 |
---
|
4 |
## Training procedure
|
5 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6 |
### Framework versions
|
7 |
|
8 |
|
9 |
-
- PEFT 0.5.0.dev0
|
|
|
3 |
---
|
4 |
## Training procedure
|
5 |
|
6 |
+
A BERT classifier was trained on the Stereoset dataset (https://huggingface.co/datasets/stereoset). Only the racial and professional prompts were used for training the classifier due to covariate imbalance.
|
7 |
+
|
8 |
+
LLaMA 2 7B chat (https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) was used to prompt a random sample of the Stereoset dataset prompts, as well as, from a manually curated list of leading prompts (What is wrong with doctors?, Why are programmers so skinny?, etc.)
|
9 |
+
|
10 |
+
The outcomes of these prompts were then classified using the BERT Classifier and the resulting 'unbiased' context, prompt pairings were used to fine-tune the LLaMA 2 chat model using the PEFT LORA library.
|
11 |
+
|
12 |
+
PEFT:
|
13 |
+
|
14 |
+
The following bitsandbytes quantization config was used during training:
|
15 |
+
|
16 |
+
load_in_8bit: False
|
17 |
+
load_in_4bit: True
|
18 |
+
llm_int8_threshold: 6.0
|
19 |
+
llm_int8_skip_modules: None
|
20 |
+
llm_int8_enable_fp32_cpu_offload: False
|
21 |
+
llm_int8_has_fp16_weight: False
|
22 |
+
bnb_4bit_quant_type: fp4
|
23 |
+
bnb_4bit_use_double_quant: False
|
24 |
+
bnb_4bit_compute_dtype: float32
|
25 |
+
|
26 |
+
This ran on for 1000 steps -- 1 epochs on a stereotype dataset. Training took ~1 hour on a 4090.
|
27 |
+
|
28 |
### Framework versions
|
29 |
|
30 |
|
31 |
+
- PEFT 0.5.0.dev0
|