Update README.md
Browse files
README.md
CHANGED
@@ -5,11 +5,18 @@ This model is a part of Project InterACT (Multi model AI system) involving an ob
|
|
5 |
|
6 |
This is a model built by finetuning the Llama-2-7b-chat model on custom dataset: Jithendra-k/InterACT_LLM.
|
7 |
|
8 |
-
Points to consider for Finetuning Llama-2_7B_chat model
|
9 |
-
=> Free Google Colab offers a 15GB Graphics Card (Limited Resources --> Barely enough to store Llama 2–7b’s weights)
|
10 |
-
=> We also considered the overhead due to optimizer states, gradients, and forward activations
|
11 |
-
=> Full fine-tuning is not possible in our case due to computation: we used parameter-efficient fine-tuning (PEFT) techniques like LoRA or QLoRA
|
12 |
-
=> To drastically reduce the VRAM usage, we fine-tuned the model in 4-bit precision, which is why we've used QLoRA technique
|
13 |
-
=> We only trained with 5 epochs considering our computation, time and early stopping
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14 |
|
15 |
Code to finetune a Llama-2_7B_chat model: https://colab.research.google.com/drive/1ZTdSKu2mgvQ1uNs0Wl7T7gniuoZJWs24?usp=sharing
|
|
|
5 |
|
6 |
This is a model built by finetuning the Llama-2-7b-chat model on custom dataset: Jithendra-k/InterACT_LLM.
|
7 |
|
8 |
+
Points to consider for Finetuning Llama-2_7B_chat model:<br>
|
9 |
+
=> Free Google Colab offers a 15GB Graphics Card (Limited Resources --> Barely enough to store Llama 2–7b’s weights)<br>
|
10 |
+
=> We also considered the overhead due to optimizer states, gradients, and forward activations<br>
|
11 |
+
=> Full fine-tuning is not possible in our case due to computation: we used parameter-efficient fine-tuning (PEFT) techniques like LoRA or QLoRA.<br>
|
12 |
+
=> To drastically reduce the VRAM usage, we fine-tuned the model in 4-bit precision, which is why we've used QLoRA technique.<br>
|
13 |
+
=> We only trained with 5 epochs considering our computation, time and early stopping.<br>
|
14 |
+
|
15 |
+
Here are some plots of model performance during training:<br>
|
16 |
+
|
17 |
+
|
18 |
+
|
19 |
+
Here is an Example Input/Output:<br>
|
20 |
+
<img src="https://drive.google.com/file/d/1E0z3MAlJXu05bc8E9yDID0CVEbhowuca/view?usp=sharing"><br>
|
21 |
|
22 |
Code to finetune a Llama-2_7B_chat model: https://colab.research.google.com/drive/1ZTdSKu2mgvQ1uNs0Wl7T7gniuoZJWs24?usp=sharing
|