Update README.md
Browse files
README.md
CHANGED
@@ -8,10 +8,14 @@ base_model:
|
|
8 |
- google/gemma-2-2b-it
|
9 |
---
|
10 |
|
|
|
11 |
The goal of this project is to adapt large language models for the Arabic language. Due to the scarcity of Arabic instruction fine-tuning data, the focus is on creating a high-quality instruction fine-tuning (IFT) dataset. The project aims to finetune models on this dataset and evaluate their performance across various benchmarks.
|
12 |
|
|
|
13 |
This model is the 2B version. It was trained for 2 days on 1 A100 GPU using LoRA with a rank of 128, a learning rate of 1e-4, and a cosine learning rate schedule.
|
14 |
|
|
|
|
|
15 |
| Metric | Slim205/Barka-2b-it |
|
16 |
|----------------------|---------------------|
|
17 |
| Average | 46.98 |
|
|
|
8 |
- google/gemma-2-2b-it
|
9 |
---
|
10 |
|
11 |
+
# Motivation :
|
12 |
The goal of this project is to adapt large language models for the Arabic language. Due to the scarcity of Arabic instruction fine-tuning data, the focus is on creating a high-quality instruction fine-tuning (IFT) dataset. The project aims to finetune models on this dataset and evaluate their performance across various benchmarks.
|
13 |
|
14 |
+
# Training :
|
15 |
This model is the 2B version. It was trained for 2 days on 1 A100 GPU using LoRA with a rank of 128, a learning rate of 1e-4, and a cosine learning rate schedule.
|
16 |
|
17 |
+
# Evaluation :
|
18 |
+
|
19 |
| Metric | Slim205/Barka-2b-it |
|
20 |
|----------------------|---------------------|
|
21 |
| Average | 46.98 |
|