|
--- |
|
base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B |
|
library_name: peft |
|
pipeline_tag: text-generation |
|
language: en |
|
tags: |
|
- deepseek |
|
- text-generation |
|
- conversational |
|
--- |
|
|
|
# DeepSeek Chatbot |
|
|
|
This is a fine-tuned version of DeepSeek-R1-Distill-Qwen-1.5B, optimized for conversational AI applications. The model maintains the base model's capabilities while being tuned for improved dialogue interactions. |
|
|
|
## Model Details |
|
|
|
### Model Description |
|
|
|
- **Developed by:** Trinoid |
|
- **Model type:** Conversational Language Model |
|
- **Language(s):** English |
|
- **License:** Same as base model (DeepSeek-R1-Distill-Qwen-1.5B) |
|
- **Finetuned from model:** deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B |
|
|
|
## Uses |
|
|
|
### Direct Use |
|
|
|
This model can be used for: |
|
- General conversation |
|
- Text generation |
|
- Question answering |
|
- Chat-based applications |
|
|
|
Example usage: |
|
```python |
|
from huggingface_hub import InferenceClient |
|
|
|
client = InferenceClient("Trinoid/Deepseek_Chatbot") |
|
|
|
messages = [ |
|
{"role": "system", "content": "You are a helpful assistant."}, |
|
{"role": "user", "content": "Hello, how are you?"} |
|
] |
|
|
|
response = client.chat_completion( |
|
messages, |
|
max_tokens=512, |
|
temperature=0.7, |
|
top_p=0.95 |
|
) |
|
``` |
|
|
|
### Out-of-Scope Use |
|
|
|
This model should not be used for: |
|
- Generation of harmful or malicious content |
|
- Spreading misinformation |
|
- Production of illegal content |
|
- Making critical decisions without human oversight |
|
|
|
## Training Details |
|
|
|
### Training Procedure |
|
|
|
#### Training Hyperparameters |
|
|
|
- **Training regime:** fp16 mixed precision |
|
- **Framework:** PEFT (Parameter-Efficient Fine-Tuning) |
|
- **PEFT Method:** LoRA |
|
- **Version:** PEFT 0.14.0 |
|
|
|
## Technical Specifications |
|
|
|
### Model Architecture and Objective |
|
|
|
- Base architecture: DeepSeek-R1-Distill-Qwen-1.5B |
|
- Fine-tuning method: PEFT/LoRA |
|
- Primary objective: Conversational AI |
|
|
|
### Compute Infrastructure |
|
|
|
#### Software |
|
- PEFT 0.14.0 |
|
- Transformers |
|
- Python 3.x |
|
|
|
## Model Card Contact |
|
|
|
For questions or issues about this model, please open an issue in the model repository. |