SicariusSicariiStuff commited on
Commit
7143865
·
verified ·
1 Parent(s): e8150c5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +95 -0
README.md CHANGED
@@ -12,3 +12,98 @@ license: apache-2.0
12
 
13
 
14
  <img src="https://i.imgur.com/pXcjpoV.png" alt="LLAMA-3_8B_Unaligned_Alpha_RP_Soup" style="width: 50%; min-width: 400px; display: block; margin: auto;">
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
 
13
 
14
  <img src="https://i.imgur.com/pXcjpoV.png" alt="LLAMA-3_8B_Unaligned_Alpha_RP_Soup" style="width: 50%; min-width: 400px; display: block; margin: auto;">
15
+
16
+
17
+ # Model Details
18
+ This model is the outcome of multiple merges, starting with the base model SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha. The merging process was conducted in several stages:
19
+
20
+ Merge 1: LLAMA-3_8B_Unaligned_Alpha was SLERP merged with EtherealRainbow-v0.3-8B.
21
+ Merge 2: LLAMA-3_8B_Unaligned_Alpha was SLERP merged with TheDrummer/Llama-3SOME-8B-v2.
22
+ Soup 1: Merge 1 was combined with Merge 2.
23
+ Final Merge: Soup 1 was SLERP merged with Nitral-Archive/Hathor_Enigmatica-L3-8B-v0.4.
24
+
25
+ The final model is surprisingly coherent (although slightly more censored), which is a bit unexpected, since all the intermediate merge steps were pretty incoherent.
26
+
27
+ # Merge 1
28
+ ```yaml
29
+ slices:
30
+ - sources:
31
+ - model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
32
+ layer_range: [0, 32]
33
+ - model: BeaverAI/Llama-3SOME-8B-v2d
34
+ layer_range: [0, 32]
35
+ merge_method: slerp
36
+ base_model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
37
+ parameters:
38
+ t:
39
+ - filter: self_attn
40
+ value: [0, 0.5, 0.3, 0.7, 1]
41
+ - filter: mlp
42
+ value: [1, 0.5, 0.7, 0.3, 0]
43
+ - value: 0.5 # fallback for rest of tensors
44
+ dtype: float16
45
+
46
+ ```
47
+
48
+ # Merge 2
49
+ ```yaml
50
+ slices:
51
+ - sources:
52
+ - model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
53
+ layer_range: [0, 32]
54
+ - model: invisietch/EtherealRainbow-v0.3-8B
55
+ layer_range: [0, 32]
56
+ merge_method: slerp
57
+ base_model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
58
+ parameters:
59
+ t:
60
+ - filter: self_attn
61
+ value: [0, 0.5, 0.3, 0.7, 1]
62
+ - filter: mlp
63
+ value: [1, 0.5, 0.7, 0.3, 0]
64
+ - value: 0.5 # fallback for rest of tensors
65
+ dtype: float16
66
+
67
+ ```
68
+
69
+ # Soup 1
70
+ ```yaml
71
+ slices:
72
+ - sources:
73
+ - model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
74
+ layer_range: [0, 32]
75
+ - model: Nitral-Archive/Hathor_Enigmatica-L3-8B-v0.4
76
+ layer_range: [0, 32]
77
+ merge_method: slerp
78
+ base_model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
79
+ parameters:
80
+ t:
81
+ - filter: self_attn
82
+ value: [0, 0.5, 0.3, 0.7, 1]
83
+ - filter: mlp
84
+ value: [1, 0.5, 0.7, 0.3, 0]
85
+ - value: 0.5 # fallback for rest of tensors
86
+ dtype: float16
87
+
88
+ ```
89
+ # Final Merge
90
+ ```yaml
91
+ slices:
92
+ - sources:
93
+ - model: Soup 1
94
+ layer_range: [0, 32]
95
+ - model: Nitral-Archive/Hathor_Enigmatica-L3-8B-v0.4
96
+ layer_range: [0, 32]
97
+ merge_method: slerp
98
+ base_model: Soup 1
99
+ parameters:
100
+ t:
101
+ - filter: self_attn
102
+ value: [0, 0.5, 0.3, 0.7, 1]
103
+ - filter: mlp
104
+ value: [1, 0.5, 0.7, 0.3, 0]
105
+ - value: 0.5 # fallback for rest of tensors
106
+ dtype: float16
107
+
108
+ ```
109
+