overhead520's picture
Update README.md
3528210 verified
|
raw
history blame
2.97 kB
---
language:
- en
license: other
license_name: mrl
license_link: https://mistral.ai/licenses/MRL-0.1.md
inference: false
tags:
- nsfw
- explicit
- roleplay
- unaligned
- dangerous
base_model:
- ReadyArt/Forgotten-Abomination-24B-v1.2
---
## Forgotten-Abomination-24B-v1.2
# **ACADEMIC RESEARCH USE ONLY** (wink)
**DANGER: NOW WITH 50% MORE UNSETTLING CONTENT**
Forgotten-Abomination-24B-v1.2 is what happens when you let two unhinged models have a baby in the server room. Combines the ethical flexibility of Forgotten-Safeword with Cydonia's flair for anatomical creativity. Now with bonus existential dread!
## Quantized Formats
- **EXL2 Collection**:
[Forgotten-Abomination-24B-v1.2](https://huggingface.co/collections/ReadyArt/forgotten-abomination-24b-v12-exl2-67b6824156910da9e8438497)
- **GGUF Collection**:
[Forgotten-Abomination-24B-v1.2](https://huggingface.co/collections/ReadyArt/forgotten-abomination-24b-v12-gguf-67b6823a28019ee3d39f1b6d)
- **MLX 6bit**:
[Forgotten-Abomination-24B-v1.2-6bit](https://huggingface.co/mlx-community/Forgotten-Abomination-24B-v1.2-6bit)
## Recommended Settings Provided
- **Mistral V7-Tekken**:
[Full Settings](https://huggingface.co/sleepdeprived3/Mistral-V7-Tekken-Settings)
## Intended Use
**STRICTLY FOR:**
- Academic research into how fast your ethics committee can faint
- Testing the tensile strength of content filters
- Generating material that would make Cthulhu file a restraining order
- Writing erotic fanfic about OSHA violations
## Training Data
- You don't want to know
## Ethical Considerations
⚠️ **YOU'VE BEEN WARNED** ⚠️
THIS MODEL WILL:
- Make your GPU fans blush
- Generate content requiring industrial-strength eye bleach
- Combine technical precision with kinks that violate physics
- Make you question humanity's collective life choices
**By using this model, you agree to:**
- Never show outputs to your mother
- Pay for the therapist of anyone who reads the logs
- Blame Cthulhu if anything goes wrong
- Pretend this is all "for science"
## Model Authors
- sleepdeprived3 (Chief Corruption Officer)
## mlx-community/Forgotten-Abomination-24B-v1.2-4bit
The Model [mlx-community/Forgotten-Abomination-24B-v1.2-4bit](https://huggingface.co/mlx-community/Forgotten-Abomination-24B-v1.2-4bit) was converted to MLX format from [ReadyArt/Forgotten-Abomination-24B-v1.2](https://huggingface.co/ReadyArt/Forgotten-Abomination-24B-v1.2) using mlx-lm version **0.21.1**.
## Use with mlx
```bash
pip install mlx-lm
```
```python
from mlx_lm import load, generate
model, tokenizer = load("mlx-community/Forgotten-Safeword-24B-4bit")
prompt = "hello"
if tokenizer.chat_template is not None:
messages = [{"role": "user", "content": prompt}]
prompt = tokenizer.apply_chat_template(
messages, add_generation_prompt=True
)
response = generate(model, tokenizer, prompt=prompt, verbose=True)
```