Codellama 34b base model fine-tuned on the text chunk from the OpenAssistant-Guanaco dataset instead of Q&A pair, so it struggles to determine the end of the answer. recommend using a stop string like "### Human:" to prevent the model from talking to itself.

Prompt template:

### Human: {prompt}
### Assistant:
Downloads last month
970
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support