Update README.md
Browse files
README.md
CHANGED
@@ -7,8 +7,10 @@ base_model:
|
|
7 |
pipeline_tag: question-answering
|
8 |
---
|
9 |
|
|
|
|
|
10 |
## Model Overview
|
11 |
-
This model is a fine-tuned version of the LLaMA 3.1-8B model, trained on a curated selection of 1,
|
12 |
|
13 |
- **Base Model**: LLaMA 3.1-8B
|
14 |
- **Fine-tuning Dataset**: 1,122 samples from ChatDoctor dataset
|
@@ -26,7 +28,7 @@ This model is designed to assist in:
|
|
26 |
|------------------------------|----------------------------|
|
27 |
| **Model Type** | Causal Language Model |
|
28 |
| **Architecture** | LLaMA 3.1-8B |
|
29 |
-
| **Training Data** | ChatDoctor (1,
|
30 |
| **Quantization** | Q4_0 |
|
31 |
| **Deployment Format** | GGUF |
|
32 |
|
@@ -48,30 +50,8 @@ The model was fine-tuned with the following hyperparameters:
|
|
48 |
|
49 |
Validation was performed using a separate subset of the dataset. The final training and validation loss are as follows:
|
50 |
|
51 |
-
![Training and Validation Loss](
|
52 |
|
53 |
## Usage
|
54 |
### Loading the Model
|
55 |
-
This model is hosted in **GGUF format** for optimal deployment. You can load and run the model using **LLaMA.cpp**.
|
56 |
-
|
57 |
-
#### Steps to Use
|
58 |
-
1. Clone the LLaMA.cpp repository:
|
59 |
-
```bash
|
60 |
-
git clone https://github.com/ggerganov/llama.cpp
|
61 |
-
cd llama.cpp
|
62 |
-
make
|
63 |
-
```
|
64 |
-
|
65 |
-
2. Download the model from Hugging Face:
|
66 |
-
```bash
|
67 |
-
huggingface-cli login
|
68 |
-
wget https://huggingface.co/your-username/llama-3.1-8B-gguf/resolve/main/output_model.gguf
|
69 |
-
```
|
70 |
-
|
71 |
-
3. Run inference:
|
72 |
-
```bash
|
73 |
-
./main -m output_model.gguf -p "What are the symptoms of a common cold?" -t 4 -n 100
|
74 |
-
```
|
75 |
-
|
76 |
-
### Quantization Details
|
77 |
-
The model is quantized to **Q4_0** for faster inference while maintaining reasonable accuracy. You can run it efficiently on CPUs with low memory requirements.
|
|
|
7 |
pipeline_tag: question-answering
|
8 |
---
|
9 |
|
10 |
+
# LLaMA 3.1-8B Fine-Tuned on ChatDoctor Dataset
|
11 |
+
|
12 |
## Model Overview
|
13 |
+
This model is a fine-tuned version of the LLaMA 3.1-8B model, trained on a curated selection of 1,122 samples from the **ChatDoctor (HealthCareMagic-100k)** dataset. It has been optimized for tasks related to medical consultations.
|
14 |
|
15 |
- **Base Model**: LLaMA 3.1-8B
|
16 |
- **Fine-tuning Dataset**: 1,122 samples from ChatDoctor dataset
|
|
|
28 |
|------------------------------|----------------------------|
|
29 |
| **Model Type** | Causal Language Model |
|
30 |
| **Architecture** | LLaMA 3.1-8B |
|
31 |
+
| **Training Data** | ChatDoctor (1,122 samples) |
|
32 |
| **Quantization** | Q4_0 |
|
33 |
| **Deployment Format** | GGUF |
|
34 |
|
|
|
50 |
|
51 |
Validation was performed using a separate subset of the dataset. The final training and validation loss are as follows:
|
52 |
|
53 |
+
![Training and Validation Loss](train-val-curve.png)
|
54 |
|
55 |
## Usage
|
56 |
### Loading the Model
|
57 |
+
This model is hosted in **GGUF format** for optimal deployment. You can load and run the model using **LLaMA.cpp**.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|