--- license: mit base_model: - deepseek-ai/DeepSeek-V3.2-Speciale --- **Note that the MTP layers of this model are also PTPC-quantized.** # Model Overview - **Model Architecture:** DeepSeek-V3.2-Speciale - **Input:** Text - **Output:** Text - **Supported Hardware Microarchitecture:** AMD MI350/MI355 - **ROCm**: 7.0 - **Operating System(s):** Linux - **Inference Engine:** [SGLang](https://docs.sglang.ai/)/[vLLM](https://docs.vllm.ai/en/latest/) - **Model Optimizer:** [AMD-Quark](https://quark.docs.amd.com/latest/index.html) (V0.10) - **Weight quantization:** Perchannel, FP8E4M3, Static - **Activation quantization:** Pertoken, FP8E4M3, Dynamic - **Calibration Dataset:** [Pile](https://huggingface.co/datasets/mit-han-lab/pile-val-backup) This model was built with deepseek-ai/DeepSeek-V3.2-Speciale model by applying [AMD-Quark](https://quark.docs.amd.com/latest/index.html) for FP8E4M3 PTPC quantization. # Model Quantization The model was quantized from [deepseek-ai/DeepSeek-V3.2-Speciale](https://huggingface.co/deepseek-ai/DeepSeek-V3.2-Speciale) using [AMD-Quark](https://quark.docs.amd.com/latest/index.html). The weights are quantized to FP8 and activations are quantized to FP8. ### Accuracy
| Benchmark | DeepSeek-V3.2-Speciale | DeepSeek-V3.2-Speciale-ptpc(this model) |
| gsm8k | 96.00 | 95.75 |