A-Mahla HF Staff commited on
Commit
50bb8e0
·
verified ·
1 Parent(s): d5eb063

Model save

Browse files
README.md CHANGED
@@ -1,13 +1,9 @@
1
  ---
2
  base_model: Qwen/Qwen2.5-VL-3B-Instruct
3
- datasets: smolagents/aguvis-stage-2
4
  library_name: transformers
5
  model_name: Qwen2.5-VL-3B-Instruct-Agentic
6
  tags:
7
  - generated_from_trainer
8
- - open-r1
9
- - vision-language
10
- - vlm
11
  - trl
12
  - sft
13
  licence: license
@@ -15,7 +11,7 @@ licence: license
15
 
16
  # Model Card for Qwen2.5-VL-3B-Instruct-Agentic
17
 
18
- This model is a fine-tuned version of [Qwen/Qwen2.5-VL-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-VL-3B-Instruct) on the [smolagents/aguvis-stage-2](https://huggingface.co/datasets/smolagents/aguvis-stage-2) dataset.
19
  It has been trained using [TRL](https://github.com/huggingface/trl).
20
 
21
  ## Quick start
@@ -31,7 +27,7 @@ print(output["generated_text"])
31
 
32
  ## Training procedure
33
 
34
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/huggingface/huggingface/runs/mkymtipi)
35
 
36
 
37
  This model was trained with SFT.
@@ -41,8 +37,8 @@ This model was trained with SFT.
41
  - TRL: 0.18.0
42
  - Transformers: 4.52.3
43
  - Pytorch: 2.6.0
44
- - Datasets: 3.6.0
45
- - Tokenizers: 0.21.1
46
 
47
  ## Citations
48
 
 
1
  ---
2
  base_model: Qwen/Qwen2.5-VL-3B-Instruct
 
3
  library_name: transformers
4
  model_name: Qwen2.5-VL-3B-Instruct-Agentic
5
  tags:
6
  - generated_from_trainer
 
 
 
7
  - trl
8
  - sft
9
  licence: license
 
11
 
12
  # Model Card for Qwen2.5-VL-3B-Instruct-Agentic
13
 
14
+ This model is a fine-tuned version of [Qwen/Qwen2.5-VL-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-VL-3B-Instruct).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/huggingface/huggingface/runs/a5t7btov)
31
 
32
 
33
  This model was trained with SFT.
 
37
  - TRL: 0.18.0
38
  - Transformers: 4.52.3
39
  - Pytorch: 2.6.0
40
+ - Datasets: 4.0.0
41
+ - Tokenizers: 0.21.2
42
 
43
  ## Citations
44
 
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "total_flos": 309043048284160.0,
3
- "train_loss": 0.11322491052526659,
4
- "train_runtime": 3926.7142,
5
- "train_samples": 20495,
6
- "train_samples_per_second": 31.316,
7
- "train_steps_per_second": 0.49
8
  }
 
1
  {
2
+ "total_flos": 1.1033250327691264e+16,
3
+ "train_loss": 0.6782059136778116,
4
+ "train_runtime": 78870.76,
5
+ "train_samples": 327501,
6
+ "train_samples_per_second": 4.152,
7
+ "train_steps_per_second": 0.008
8
  }
config.json CHANGED
@@ -70,7 +70,7 @@
70
  },
71
  "torch_dtype": "bfloat16",
72
  "transformers_version": "4.52.3",
73
- "use_cache": true,
74
  "use_sliding_window": false,
75
  "video_token_id": 151656,
76
  "vision_config": {
 
70
  },
71
  "torch_dtype": "bfloat16",
72
  "transformers_version": "4.52.3",
73
+ "use_cache": false,
74
  "use_sliding_window": false,
75
  "video_token_id": 151656,
76
  "vision_config": {
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:736013d0a86054665b4246806a0307536ccbb0b71d1ff03895e2fab630a64c18
3
  size 4997756664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50c6779f3e08adbe8eafe5bf9c1f48b70f50c4d8a78e1fecbcfe19c4a3c1e140
3
  size 4997756664
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c17f0453a22b53ee5311492941a8abc7cd6ec9e6a0cd2358ff3b1e237b9afa60
3
  size 2511590128
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b59679cf1cb5e5267415e940d8158c1d1e1a5a9192116ce281aed6294eba115
3
  size 2511590128
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba0c439f7be467bf47d12a7e6f9adc6116201056fc60c67f431c679b7c16afc8
3
- size 11422064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
tokenizer_config.json CHANGED
@@ -199,9 +199,10 @@
199
  "eos_token": "<|im_end|>",
200
  "errors": "replace",
201
  "extra_special_tokens": {},
 
 
202
  "model_max_length": 131072,
203
  "pad_token": "<|endoftext|>",
204
- "processor_class": "Qwen2_5_VLProcessor",
205
  "split_special_tokens": false,
206
  "tokenizer_class": "Qwen2Tokenizer",
207
  "unk_token": null
 
199
  "eos_token": "<|im_end|>",
200
  "errors": "replace",
201
  "extra_special_tokens": {},
202
+ "max_pixels": 1003520,
203
+ "min_pixels": 200704,
204
  "model_max_length": 131072,
205
  "pad_token": "<|endoftext|>",
 
206
  "split_special_tokens": false,
207
  "tokenizer_class": "Qwen2Tokenizer",
208
  "unk_token": null
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "total_flos": 309043048284160.0,
3
- "train_loss": 0.11322491052526659,
4
- "train_runtime": 3926.7142,
5
- "train_samples": 20495,
6
- "train_samples_per_second": 31.316,
7
- "train_steps_per_second": 0.49
8
  }
 
1
  {
2
+ "total_flos": 1.1033250327691264e+16,
3
+ "train_loss": 0.6782059136778116,
4
+ "train_runtime": 78870.76,
5
+ "train_samples": 327501,
6
+ "train_samples_per_second": 4.152,
7
+ "train_steps_per_second": 0.008
8
  }
trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0a173f08d1eae91c16a9c6221f280e019a42ffd0ba8c5ded0e47d5670fa39f3b
3
- size 7352
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c0ce0e66c77eae20e729b7f22ffbfec3eec08941b70655111f37221a08e545b
3
+ size 7416