mlx-community/DeepSeek-R1-Distill-Qwen-32B-abliterated-4bit Text Generation • 5B • Updated Feb 20, 2025 • 359 • 5
mlx-community/DeepSeek-R1-Distill-Qwen-32B-abliterated Text Generation • 33B • Updated Feb 20, 2025 • 1.52k • • 3
mlx-community/DeepSeek-R1-Distill-Qwen-32B-4bit Text Generation • 5B • Updated Feb 26, 2025 • 1.56k • 45
Running 3.67k The Ultra-Scale Playbook 🌌 3.67k The ultimate guide to training LLM on large GPU Clusters