1. Introduction

This model is quantized version of DeepSeek-R1-Distill-Qwen-14B with dataset for imatrix TFMC/imatrix-dataset-for-japanese-llm. Usgin English/Japanese mixed and quantization is tuned for Japanese.

2. License

This code repository and the model weights are licensed under the MIT License. DeepSeek-R1 series support commercial use, allow for any modifications and derivative works, including, but not limited to, distillation for training other LLMs. Please note that:

  • DeepSeek-R1-Distill-Qwen-1.5B, DeepSeek-R1-Distill-Qwen-7B, DeepSeek-R1-Distill-Qwen-14B and DeepSeek-R1-Distill-Qwen-32B are derived from Qwen-2.5 series, which are originally licensed under Apache 2.0 License, and now finetuned with 800k samples curated with DeepSeek-R1.
  • DeepSeek-R1-Distill-Llama-8B is derived from Llama3.1-8B-Base and is originally licensed under llama3.1 license.
  • DeepSeek-R1-Distill-Llama-70B is derived from Llama3.3-70B-Instruct and is originally licensed under llama3.3 license.
Downloads last month
10
GGUF
Model size
14.8B params
Architecture
qwen2

4-bit

Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for okamototk/DeepSeek-R1-Distill-Qwen-14B-imatrix-gguf

Quantized
(109)
this model

Dataset used to train okamototk/DeepSeek-R1-Distill-Qwen-14B-imatrix-gguf