|
--- |
|
language: |
|
- en |
|
library_name: pytorch |
|
pipeline_tag: text-generation |
|
tags: |
|
- code |
|
- deepseek |
|
- gguf |
|
- f32 |
|
- f16 |
|
- q2 |
|
- q8 |
|
- q6 |
|
- q4_k_m |
|
- humaneval |
|
license: mit |
|
--- |
|
|
|
# Model Card for wavecoder-ds-6.7b-GGUF |
|
|
|
WaveCoder 🌊 is a series of large language models (LLMs) for the coding domain. |
|
|
|
## Model Details |
|
|
|
- WaveCoder-6.7b-ds = Trained using CodeOcean dataset |
|
- WaveCoder-6.7b-pro = Trained using GPT-4 synthetic data |
|
- WaveCoder-6.7b-ultra = Trained using enhanced GPT-4 synthetic data |
|
|
|
### Model Description |
|
|
|
WaveCoder 🌊 is a series of large language models (LLMs) for the coding domain, designed to solve relevant problems in the field of code through instruction-following learning. Its training dataset was generated from a subset of code-search-net data using a generator-discriminator framework based on LLMs that we proposed, covering four general code-related tasks: code generation, code summary, code translation, and code repair. |
|
|
|
- **Developed by:** Yu, Zhaojian and Zhang, Xin and Shang, Ning and Huang, Yangyu and Xu, Can and Zhao, Yishujie and Hu, Wenxiang and Yin, Qiufeng |
|
- **Model type:** Large Language Model |
|
- **Language(s) (NLP):** English |
|
- **License:** DeepSeek License (Model) |
|
|
|
### Model Sources |
|
|
|
- **Repository:** [https://huggingface.co/microsoft/wavecoder-ds-6.7b](https://huggingface.co/microsoft/wavecoder-ds-6.7b) |
|
- **Paper :** [More Information Needed] |
|
- **Demo :** [More Information Needed] |
|
|
|
## Uses |
|
|
|
Coding/Refactoring/Cleanup/Fixing Code |
|
|
|
## Original: [https://huggingface.co/microsoft/wavecoder-ds-6.7b](https://huggingface.co/microsoft/wavecoder-ds-6.7b) |