Yassinj commited on
Commit
ba2b748
verified
1 Parent(s): e35da25

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -26
README.md CHANGED
@@ -7,8 +7,10 @@ base_model:
7
  pipeline_tag: question-answering
8
  ---
9
 
 
 
10
  ## Model Overview
11
- This model is a fine-tuned version of the LLaMA 3.1-8B model, trained on a curated selection of 1,000 samples from the **ChatDoctor (HealthCareMagic-100k)** dataset. It has been optimized for tasks related to medical consultations.
12
 
13
  - **Base Model**: LLaMA 3.1-8B
14
  - **Fine-tuning Dataset**: 1,122 samples from ChatDoctor dataset
@@ -26,7 +28,7 @@ This model is designed to assist in:
26
  |------------------------------|----------------------------|
27
  | **Model Type** | Causal Language Model |
28
  | **Architecture** | LLaMA 3.1-8B |
29
- | **Training Data** | ChatDoctor (1,000 samples) |
30
  | **Quantization** | Q4_0 |
31
  | **Deployment Format** | GGUF |
32
 
@@ -48,30 +50,8 @@ The model was fine-tuned with the following hyperparameters:
48
 
49
  Validation was performed using a separate subset of the dataset. The final training and validation loss are as follows:
50
 
51
- ![Training and Validation Loss](https://your-image-link-here.com/training-validation-loss.png)
52
 
53
  ## Usage
54
  ### Loading the Model
55
- This model is hosted in **GGUF format** for optimal deployment. You can load and run the model using **LLaMA.cpp**.
56
-
57
- #### Steps to Use
58
- 1. Clone the LLaMA.cpp repository:
59
- ```bash
60
- git clone https://github.com/ggerganov/llama.cpp
61
- cd llama.cpp
62
- make
63
- ```
64
-
65
- 2. Download the model from Hugging Face:
66
- ```bash
67
- huggingface-cli login
68
- wget https://huggingface.co/your-username/llama-3.1-8B-gguf/resolve/main/output_model.gguf
69
- ```
70
-
71
- 3. Run inference:
72
- ```bash
73
- ./main -m output_model.gguf -p "What are the symptoms of a common cold?" -t 4 -n 100
74
- ```
75
-
76
- ### Quantization Details
77
- The model is quantized to **Q4_0** for faster inference while maintaining reasonable accuracy. You can run it efficiently on CPUs with low memory requirements.
 
7
  pipeline_tag: question-answering
8
  ---
9
 
10
+ # LLaMA 3.1-8B Fine-Tuned on ChatDoctor Dataset
11
+
12
  ## Model Overview
13
+ This model is a fine-tuned version of the LLaMA 3.1-8B model, trained on a curated selection of 1,122 samples from the **ChatDoctor (HealthCareMagic-100k)** dataset. It has been optimized for tasks related to medical consultations.
14
 
15
  - **Base Model**: LLaMA 3.1-8B
16
  - **Fine-tuning Dataset**: 1,122 samples from ChatDoctor dataset
 
28
  |------------------------------|----------------------------|
29
  | **Model Type** | Causal Language Model |
30
  | **Architecture** | LLaMA 3.1-8B |
31
+ | **Training Data** | ChatDoctor (1,122 samples) |
32
  | **Quantization** | Q4_0 |
33
  | **Deployment Format** | GGUF |
34
 
 
50
 
51
  Validation was performed using a separate subset of the dataset. The final training and validation loss are as follows:
52
 
53
+ ![Training and Validation Loss](train-val-curve.png)
54
 
55
  ## Usage
56
  ### Loading the Model
57
+ This model is hosted in **GGUF format** for optimal deployment. You can load and run the model using **LLaMA.cpp**.