QwQ-32B-Preview LoRA for separating thinking/answer parts

This LoRA file was fine-tuned to make QwQ constantly separate its private thoughts from the final answer using <THINKING>...</THINKING><ANSWER>...</ANSWER> tags.

A Q4_K_M GGUF version (which can be used as an adapter for Ollama) is available on shakedzy/QwQ-32B-Preview-with-Tags-LoRA-GGUF.

Downloads last month
0
Safetensors
Model size
18.9B params
Tensor type
F32
·
FP16
·
U8
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no pipeline_tag.

Model tree for shakedzy/QwQ-32b-Preview-bnb-4bit-wTags

Base model

Qwen/Qwen2.5-32B
Adapter
(22)
this model