GGUF models of the following model : https://huggingface.co/mridul3301/BioMistral-7B-finetuned
3 format of quantization:
- fp8
- fp16
- fp32
Converted the safetensors to GGUF for inference in CPU using llama_cpp
- Downloads last month
- 43
Hardware compatibility
Log In to add your hardware
We're not able to determine the quantization variants.
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support