Update README.md
Browse files
README.md
CHANGED
@@ -12,3 +12,98 @@ license: apache-2.0
|
|
12 |
|
13 |
|
14 |
<img src="https://i.imgur.com/pXcjpoV.png" alt="LLAMA-3_8B_Unaligned_Alpha_RP_Soup" style="width: 50%; min-width: 400px; display: block; margin: auto;">
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
12 |
|
13 |
|
14 |
<img src="https://i.imgur.com/pXcjpoV.png" alt="LLAMA-3_8B_Unaligned_Alpha_RP_Soup" style="width: 50%; min-width: 400px; display: block; margin: auto;">
|
15 |
+
|
16 |
+
|
17 |
+
# Model Details
|
18 |
+
This model is the outcome of multiple merges, starting with the base model SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha. The merging process was conducted in several stages:
|
19 |
+
|
20 |
+
Merge 1: LLAMA-3_8B_Unaligned_Alpha was SLERP merged with EtherealRainbow-v0.3-8B.
|
21 |
+
Merge 2: LLAMA-3_8B_Unaligned_Alpha was SLERP merged with TheDrummer/Llama-3SOME-8B-v2.
|
22 |
+
Soup 1: Merge 1 was combined with Merge 2.
|
23 |
+
Final Merge: Soup 1 was SLERP merged with Nitral-Archive/Hathor_Enigmatica-L3-8B-v0.4.
|
24 |
+
|
25 |
+
The final model is surprisingly coherent (although slightly more censored), which is a bit unexpected, since all the intermediate merge steps were pretty incoherent.
|
26 |
+
|
27 |
+
# Merge 1
|
28 |
+
```yaml
|
29 |
+
slices:
|
30 |
+
- sources:
|
31 |
+
- model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
|
32 |
+
layer_range: [0, 32]
|
33 |
+
- model: BeaverAI/Llama-3SOME-8B-v2d
|
34 |
+
layer_range: [0, 32]
|
35 |
+
merge_method: slerp
|
36 |
+
base_model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
|
37 |
+
parameters:
|
38 |
+
t:
|
39 |
+
- filter: self_attn
|
40 |
+
value: [0, 0.5, 0.3, 0.7, 1]
|
41 |
+
- filter: mlp
|
42 |
+
value: [1, 0.5, 0.7, 0.3, 0]
|
43 |
+
- value: 0.5 # fallback for rest of tensors
|
44 |
+
dtype: float16
|
45 |
+
|
46 |
+
```
|
47 |
+
|
48 |
+
# Merge 2
|
49 |
+
```yaml
|
50 |
+
slices:
|
51 |
+
- sources:
|
52 |
+
- model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
|
53 |
+
layer_range: [0, 32]
|
54 |
+
- model: invisietch/EtherealRainbow-v0.3-8B
|
55 |
+
layer_range: [0, 32]
|
56 |
+
merge_method: slerp
|
57 |
+
base_model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
|
58 |
+
parameters:
|
59 |
+
t:
|
60 |
+
- filter: self_attn
|
61 |
+
value: [0, 0.5, 0.3, 0.7, 1]
|
62 |
+
- filter: mlp
|
63 |
+
value: [1, 0.5, 0.7, 0.3, 0]
|
64 |
+
- value: 0.5 # fallback for rest of tensors
|
65 |
+
dtype: float16
|
66 |
+
|
67 |
+
```
|
68 |
+
|
69 |
+
# Soup 1
|
70 |
+
```yaml
|
71 |
+
slices:
|
72 |
+
- sources:
|
73 |
+
- model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
|
74 |
+
layer_range: [0, 32]
|
75 |
+
- model: Nitral-Archive/Hathor_Enigmatica-L3-8B-v0.4
|
76 |
+
layer_range: [0, 32]
|
77 |
+
merge_method: slerp
|
78 |
+
base_model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
|
79 |
+
parameters:
|
80 |
+
t:
|
81 |
+
- filter: self_attn
|
82 |
+
value: [0, 0.5, 0.3, 0.7, 1]
|
83 |
+
- filter: mlp
|
84 |
+
value: [1, 0.5, 0.7, 0.3, 0]
|
85 |
+
- value: 0.5 # fallback for rest of tensors
|
86 |
+
dtype: float16
|
87 |
+
|
88 |
+
```
|
89 |
+
# Final Merge
|
90 |
+
```yaml
|
91 |
+
slices:
|
92 |
+
- sources:
|
93 |
+
- model: Soup 1
|
94 |
+
layer_range: [0, 32]
|
95 |
+
- model: Nitral-Archive/Hathor_Enigmatica-L3-8B-v0.4
|
96 |
+
layer_range: [0, 32]
|
97 |
+
merge_method: slerp
|
98 |
+
base_model: Soup 1
|
99 |
+
parameters:
|
100 |
+
t:
|
101 |
+
- filter: self_attn
|
102 |
+
value: [0, 0.5, 0.3, 0.7, 1]
|
103 |
+
- filter: mlp
|
104 |
+
value: [1, 0.5, 0.7, 0.3, 0]
|
105 |
+
- value: 0.5 # fallback for rest of tensors
|
106 |
+
dtype: float16
|
107 |
+
|
108 |
+
```
|
109 |
+
|