End of training
Browse files
README.md
CHANGED
|
@@ -5,6 +5,8 @@ base_model: google/gemma-3-1b-it
|
|
| 5 |
tags:
|
| 6 |
- axolotl
|
| 7 |
- generated_from_trainer
|
|
|
|
|
|
|
| 8 |
model-index:
|
| 9 |
- name: gemma3_1b_lora_sft_sitcom
|
| 10 |
results: []
|
|
@@ -74,7 +76,7 @@ weight_decay: 0.0
|
|
| 74 |
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/deepakkarkala-personal/finetuning_llama31_8b_sitcom/runs/sft_gemma3_1b_2)
|
| 75 |
# gemma3_1b_lora_sft_sitcom
|
| 76 |
|
| 77 |
-
This model is a fine-tuned version of [google/gemma-3-1b-it](https://huggingface.co/google/gemma-3-1b-it) on
|
| 78 |
|
| 79 |
## Model description
|
| 80 |
|
|
@@ -104,6 +106,10 @@ The following hyperparameters were used during training:
|
|
| 104 |
- lr_scheduler_warmup_steps: 26
|
| 105 |
- training_steps: 264
|
| 106 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 107 |
### Framework versions
|
| 108 |
|
| 109 |
- PEFT 0.15.2
|
|
|
|
| 5 |
tags:
|
| 6 |
- axolotl
|
| 7 |
- generated_from_trainer
|
| 8 |
+
datasets:
|
| 9 |
+
- deepakkarkala/sft_sitcom_chandlerbing_jsonl
|
| 10 |
model-index:
|
| 11 |
- name: gemma3_1b_lora_sft_sitcom
|
| 12 |
results: []
|
|
|
|
| 76 |
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/deepakkarkala-personal/finetuning_llama31_8b_sitcom/runs/sft_gemma3_1b_2)
|
| 77 |
# gemma3_1b_lora_sft_sitcom
|
| 78 |
|
| 79 |
+
This model is a fine-tuned version of [google/gemma-3-1b-it](https://huggingface.co/google/gemma-3-1b-it) on the deepakkarkala/sft_sitcom_chandlerbing_jsonl dataset.
|
| 80 |
|
| 81 |
## Model description
|
| 82 |
|
|
|
|
| 106 |
- lr_scheduler_warmup_steps: 26
|
| 107 |
- training_steps: 264
|
| 108 |
|
| 109 |
+
### Training results
|
| 110 |
+
|
| 111 |
+
|
| 112 |
+
|
| 113 |
### Framework versions
|
| 114 |
|
| 115 |
- PEFT 0.15.2
|