Jithendra-k commited on
Commit
a2eaa61
·
verified ·
1 Parent(s): 1a83a32

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -6
README.md CHANGED
@@ -5,11 +5,18 @@ This model is a part of Project InterACT (Multi model AI system) involving an ob
5
 
6
  This is a model built by finetuning the Llama-2-7b-chat model on custom dataset: Jithendra-k/InterACT_LLM.
7
 
8
- Points to consider for Finetuning Llama-2_7B_chat model:
9
- => Free Google Colab offers a 15GB Graphics Card (Limited Resources --> Barely enough to store Llama 2–7b’s weights)
10
- => We also considered the overhead due to optimizer states, gradients, and forward activations
11
- => Full fine-tuning is not possible in our case due to computation: we used parameter-efficient fine-tuning (PEFT) techniques like LoRA or QLoRA.
12
- => To drastically reduce the VRAM usage, we fine-tuned the model in 4-bit precision, which is why we've used QLoRA technique.
13
- => We only trained with 5 epochs considering our computation, time and early stopping.
 
 
 
 
 
 
 
14
 
15
  Code to finetune a Llama-2_7B_chat model: https://colab.research.google.com/drive/1ZTdSKu2mgvQ1uNs0Wl7T7gniuoZJWs24?usp=sharing
 
5
 
6
  This is a model built by finetuning the Llama-2-7b-chat model on custom dataset: Jithendra-k/InterACT_LLM.
7
 
8
+ Points to consider for Finetuning Llama-2_7B_chat model:<br>
9
+ => Free Google Colab offers a 15GB Graphics Card (Limited Resources --> Barely enough to store Llama 2–7b’s weights)<br>
10
+ => We also considered the overhead due to optimizer states, gradients, and forward activations<br>
11
+ => Full fine-tuning is not possible in our case due to computation: we used parameter-efficient fine-tuning (PEFT) techniques like LoRA or QLoRA.<br>
12
+ => To drastically reduce the VRAM usage, we fine-tuned the model in 4-bit precision, which is why we've used QLoRA technique.<br>
13
+ => We only trained with 5 epochs considering our computation, time and early stopping.<br>
14
+
15
+ Here are some plots of model performance during training:<br>
16
+
17
+
18
+
19
+ Here is an Example Input/Output:<br>
20
+ <img src="https://drive.google.com/file/d/1E0z3MAlJXu05bc8E9yDID0CVEbhowuca/view?usp=sharing"><br>
21
 
22
  Code to finetune a Llama-2_7B_chat model: https://colab.research.google.com/drive/1ZTdSKu2mgvQ1uNs0Wl7T7gniuoZJWs24?usp=sharing