File size: 2,474 Bytes
4d2c3d9 e0c2f8a 2651e8e e0c2f8a 2651e8e 59f6109 2651e8e e0c2f8a 2651e8e ab380b2 2651e8e |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 |
---
license: cc-by-nc-nd-4.0
---
# Mistral-7B-Instruct-v0.2-code-ft
I'm thrilled to introduce the latest iteration of our model, Mistral-7B-Instruct-v0.2-code-ft. This updated version is designed to further enhance coding assistance and co-pilot functionalities. We're eager for developers and enthusiasts to try it out and provide feedback!
## Additional Information
This version builds upon the previous Mistral-7B models, incorporating new datasets and features for a more refined experience.
## Prompt template: ChatML
```
<|im_start|>system
{system_message}<|im_end|>
<|im_start|>user
{prompt}<|im_end|>
<|im_start|>assistant
```
## Quantised Model Links:
1. https://huggingface.co/LoneStriker/Mistral-7B-Instruct-v0.2-code-ft-8.0bpw-h8-exl2
2. https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-code-ft-GGUF
3. https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-code-ft-AWQ
4. https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-code-ft-GPTQ
## Eval Plus Performance
For detailed performance metrics, visit Eval Plus page: [Mistral-7B-Instruct-v0.2-code-ft Eval Plus](https://github.com/evalplus/evalplus)
Score: 0.421

## Dataset:
The model has been trained on a new dataset to improve its performance and versatility:
- path: ajibawa-2023/Code-74k-ShareGPT
type: sharegpt
conversation: chatml
Find more about the dataset here: [Code-74k-ShareGPT Dataset](https://huggingface.co/datasets/ajibawa-2023/Code-74k-ShareGPT)
## Model Architecture
- Base Model: mistralai/Mistral-7B-Instruct-v0.2
- Tokenizer Type: LlamaTokenizer
- Model Type: MistralForCausalLM
- Is Mistral Derived Model: true
- Sequence Length: 16384 with sample packing
## Enhanced Features
- Adapter: qlora
- Learning Rate: 0.0002 with cosine lr scheduler
- Optimizer: adamw_bnb_8bit
- Training Enhancements: bf16 training, gradient checkpointing, and flash attention
## Download Information
You can download and explore this model through these links on Hugging Face.
## Contributions and Feedback
We welcome contributions and feedback from the community. Please feel free to open issues or pull requests on repository.
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
|