Update README.md
Browse files
README.md
CHANGED
@@ -129,9 +129,10 @@ base_model:
|
|
129 |
|
130 |
## Model Summary
|
131 |
|
132 |
-
In contrast to lmms-lab/LLaVA-NeXT-Video-7B-Qwen2, this is a 7B model trained on [LLaVA-
|
133 |
|
134 |
-
|
|
|
135 |
|
136 |
- **Repository:** [LLaVA-VL/LLaVA-NeXT](https://github.com/LLaVA-VL/LLaVA-NeXT?tab=readme-ov-file)
|
137 |
- **Point of Contact:** [Yuanhan Zhang](https://zhangyuanhan-ai.github.io/)
|
@@ -229,5 +230,3 @@ print(text_outputs)
|
|
229 |
- **GPUs:** 256 * Nvidia Tesla A100 (for whole model series training)
|
230 |
- **Orchestration:** [Huggingface Trainer](https://huggingface.co/docs/transformers/main_classes/trainer)
|
231 |
- **Neural networks:** [PyTorch](https://github.com/pytorch/pytorch)
|
232 |
-
|
233 |
-
# Citation
|
|
|
129 |
|
130 |
## Model Summary
|
131 |
|
132 |
+
In contrast to lmms-lab/LLaVA-NeXT-Video-7B-Qwen2, this is a 7B model trained on [LLaVA-Video-178K](https://huggingface.co/datasets/lmms-lab/LLaVA-NeXT-Video-SFT-Data) only, based on Qwen2 language model with a context window of 32K tokens.
|
133 |
|
134 |
+
|
135 |
+
This model supports up to 110 frames and achieves comparable results to those of lmms-lab/LLaVA-NeXT-Video-7B-Qwen2 in terms of video benchmarks."
|
136 |
|
137 |
- **Repository:** [LLaVA-VL/LLaVA-NeXT](https://github.com/LLaVA-VL/LLaVA-NeXT?tab=readme-ov-file)
|
138 |
- **Point of Contact:** [Yuanhan Zhang](https://zhangyuanhan-ai.github.io/)
|
|
|
230 |
- **GPUs:** 256 * Nvidia Tesla A100 (for whole model series training)
|
231 |
- **Orchestration:** [Huggingface Trainer](https://huggingface.co/docs/transformers/main_classes/trainer)
|
232 |
- **Neural networks:** [PyTorch](https://github.com/pytorch/pytorch)
|
|
|
|