This quantized models are converted from deepseek-ai/DeepSeek-R1-Distill-Llama-8B with https://huggingface.co/spaces/ggml-org/gguf-my-repo

The bf16, f16 models are converted by llama.cpp with version 25ff6f7659f6a5c47d6a73eada5813f0495331f0

Downloads last month
9
GGUF
Model size
8B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including skymizer/DeepSeek-R1-Distill-Llama-8B-GGUF