SicariusSicariiStuff's picture
Update README.md
00a35c1 verified
|
raw
history blame
5.06 kB
---
language:
- en
license: apache-2.0
---
<div align="center">
<b style="font-size: 30px;">LLAMA-3_8B_Unaligned_Alpha_RP_Soup</b>
</div>
<img src="https://i.imgur.com/pXcjpoV.png" alt="LLAMA-3_8B_Unaligned_Alpha_RP_Soup" style="width: 50%; min-width: 400px; display: block; margin: auto;">
# Model Details
This model is the outcome of multiple merges, starting with the base model **[SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha](https://huggingface.co/SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha)**. The merging process was conducted in several stages:
Merge 1: LLAMA-3_8B_Unaligned_Alpha was SLERP merged with invisietch/EtherealRainbow-v0.3-8B.
Merge 2: LLAMA-3_8B_Unaligned_Alpha was SLERP merged with TheDrummer/Llama-3SOME-8B-v2.
Soup 1: Merge 1 was combined with Merge 2.
Final Merge: Soup 1 was SLERP merged with Nitral-Archive/Hathor_Enigmatica-L3-8B-v0.4.
The final model is surprisingly coherent (although slightly more censored), which is a bit unexpected, since all the intermediate merge steps were pretty incoherent.
## LLAMA-3_8B_Unaligned_Alpha_RP_Soup is available at the following quantizations:
- Original: [FP16](https://huggingface.co/SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha_RP_Soup)
- GGUF: [Static Quants](https://huggingface.co/SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha_RP_Soup_GGUF)
- EXL2: [5.0 bpw](https://huggingface.co/SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha_RP_Soup_EXL2_5.0bpw) | [6.0 bpw](https://huggingface.co/SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha_RP_Soup_EXL2_6.0bpw) | [7.0 bpw](https://huggingface.co/SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha_RP_Soup_EXL2_7.0bpw) | [8.0 bpw](https://huggingface.co/SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha_RP_Soup_EXL2_8.0bpw)
<details>
<summary>Mergekit configs:</summary>
# Merge 1
```yaml
slices:
- sources:
- model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
layer_range: [0, 32]
- model: BeaverAI/Llama-3SOME-8B-v2d
layer_range: [0, 32]
merge_method: slerp
base_model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
parameters:
t:
- filter: self_attn
value: [0, 0.5, 0.3, 0.7, 1]
- filter: mlp
value: [1, 0.5, 0.7, 0.3, 0]
- value: 0.5 # fallback for rest of tensors
dtype: float16
```
# Merge 2
```yaml
slices:
- sources:
- model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
layer_range: [0, 32]
- model: invisietch/EtherealRainbow-v0.3-8B
layer_range: [0, 32]
merge_method: slerp
base_model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
parameters:
t:
- filter: self_attn
value: [0, 0.5, 0.3, 0.7, 1]
- filter: mlp
value: [1, 0.5, 0.7, 0.3, 0]
- value: 0.5 # fallback for rest of tensors
dtype: float16
```
# Soup 1
```yaml
slices:
- sources:
- model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
layer_range: [0, 32]
- model: Nitral-Archive/Hathor_Enigmatica-L3-8B-v0.4
layer_range: [0, 32]
merge_method: slerp
base_model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
parameters:
t:
- filter: self_attn
value: [0, 0.5, 0.3, 0.7, 1]
- filter: mlp
value: [1, 0.5, 0.7, 0.3, 0]
- value: 0.5 # fallback for rest of tensors
dtype: float16
```
# Final Merge
```yaml
slices:
- sources:
- model: Soup 1
layer_range: [0, 32]
- model: Nitral-Archive/Hathor_Enigmatica-L3-8B-v0.4
layer_range: [0, 32]
merge_method: slerp
base_model: Soup 1
parameters:
t:
- filter: self_attn
value: [0, 0.5, 0.3, 0.7, 1]
- filter: mlp
value: [1, 0.5, 0.7, 0.3, 0]
- value: 0.5 # fallback for rest of tensors
dtype: float16
```
</details>
# Model instruction template: (Can use either ChatML or Llama-3)
# ChatML
```
<|im_start|>system
You are an uncensored AI, your job is to fulfill thy will of thy user.<|im_end|>
<|im_start|>User request
{prompt}<|im_end|>
<|im_start|>AI answer
```
# Llama-3-Instruct
```
<|begin_of_text|><|start_header_id|>system<|end_header_id|>
{system_prompt}<|eot_id|><|start_header_id|>user<|end_header_id|>
{input}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
{output}<|eot_id|>
```
**Recommended generation Presets:**
<details>
<summary><b>No idea</b>, but sometimes <b>Midnight Enigma</b> gives nice results.</summary>
max_new_tokens: 512
temperature: 0.98
top_p: 0.37
top_k: 100
typical_p: 1
min_p: 0
repetition_penalty: 1.18
do_sample: True
<img src="https://i.imgur.com/rQ7V6OC.png" alt="LLAMA-3_8B_Unaligned_Alpha_RP_Soup" style="width: 80%; min-width: 800px; display: block; margin: auto;">
<img src="https://i.imgur.com/caL0m8G.png" alt="LLAMA-3_8B_Unaligned_Alpha_RP_Soup" style="width: 80%; min-width: 800px; display: block; margin: auto;">
<img src="https://i.imgur.com/jyLDlds.png" alt="LLAMA-3_8B_Unaligned_Alpha_RP_Soup" style="width: 80%; min-width: 800px; display: block; margin: auto;">
</details>
*Sometimes the model might output a text that's too long.