ZhangYuanhan commited on
Commit
4ae7f23
·
verified ·
1 Parent(s): e54db93

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -4
README.md CHANGED
@@ -129,9 +129,10 @@ base_model:
129
 
130
  ## Model Summary
131
 
132
- In contrast to lmms-lab/LLaVA-NeXT-Video-7B-Qwen2, this is a 7B model trained on [LLaVA-NeXT-Video-178K](https://huggingface.co/datasets/lmms-lab/LLaVA-NeXT-Video-SFT-Data) only, based on Qwen2 language model with a context window of 32K tokens.
133
 
134
- This model support at most 110 frames.
 
135
 
136
  - **Repository:** [LLaVA-VL/LLaVA-NeXT](https://github.com/LLaVA-VL/LLaVA-NeXT?tab=readme-ov-file)
137
  - **Point of Contact:** [Yuanhan Zhang](https://zhangyuanhan-ai.github.io/)
@@ -229,5 +230,3 @@ print(text_outputs)
229
  - **GPUs:** 256 * Nvidia Tesla A100 (for whole model series training)
230
  - **Orchestration:** [Huggingface Trainer](https://huggingface.co/docs/transformers/main_classes/trainer)
231
  - **Neural networks:** [PyTorch](https://github.com/pytorch/pytorch)
232
-
233
- # Citation
 
129
 
130
  ## Model Summary
131
 
132
+ In contrast to lmms-lab/LLaVA-NeXT-Video-7B-Qwen2, this is a 7B model trained on [LLaVA-Video-178K](https://huggingface.co/datasets/lmms-lab/LLaVA-NeXT-Video-SFT-Data) only, based on Qwen2 language model with a context window of 32K tokens.
133
 
134
+
135
+ This model supports up to 110 frames and achieves comparable results to those of lmms-lab/LLaVA-NeXT-Video-7B-Qwen2 in terms of video benchmarks."
136
 
137
  - **Repository:** [LLaVA-VL/LLaVA-NeXT](https://github.com/LLaVA-VL/LLaVA-NeXT?tab=readme-ov-file)
138
  - **Point of Contact:** [Yuanhan Zhang](https://zhangyuanhan-ai.github.io/)
 
230
  - **GPUs:** 256 * Nvidia Tesla A100 (for whole model series training)
231
  - **Orchestration:** [Huggingface Trainer](https://huggingface.co/docs/transformers/main_classes/trainer)
232
  - **Neural networks:** [PyTorch](https://github.com/pytorch/pytorch)