Mistral-7b-v0.1-DPO is a finetuned adapter from the original Mistral-7b model. In this adaptor, I am finetuning the LM head in addition to the regular modules that are normally finetuned. Below is the list of the finetuned modules: 'k_proj', 'gate_proj', 'v_proj', 'up_proj', 'q_proj', 'o_proj', 'down_proj', 'lm_head'

Downloads last month
1,971
Safetensors
Model size
7.24B params
Tensor type
FP16
Β·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for walebadr/Mistral-7B-v0.1-DPO

Quantizations
2 models

Spaces using walebadr/Mistral-7B-v0.1-DPO 6