ZhangYuanhan
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -10,7 +10,7 @@ metrics:
|
|
10 |
tags:
|
11 |
- multimodal
|
12 |
model-index:
|
13 |
-
- name: LLaVA-
|
14 |
results:
|
15 |
- task:
|
16 |
type: multimodal
|
@@ -116,7 +116,7 @@ base_model:
|
|
116 |
- lmms-lab/llava-onevision-qwen2-7b-si
|
117 |
---
|
118 |
|
119 |
-
# LLaVA-
|
120 |
|
121 |
## Table of Contents
|
122 |
|
@@ -132,7 +132,7 @@ base_model:
|
|
132 |
In contrast to lmms-lab/LLaVA-NeXT-Video-7B-Qwen2, this is a 7B model trained on [LLaVA-Video-178K](https://huggingface.co/datasets/lmms-lab/LLaVA-NeXT-Video-SFT-Data) only, based on Qwen2 language model with a context window of 32K tokens.
|
133 |
|
134 |
|
135 |
-
This model supports up to 110 frames and achieves comparable results to those of lmms-lab/LLaVA-
|
136 |
|
137 |
- **Repository:** [LLaVA-VL/LLaVA-NeXT](https://github.com/LLaVA-VL/LLaVA-NeXT?tab=readme-ov-file)
|
138 |
- **Point of Contact:** [Yuanhan Zhang](https://zhangyuanhan-ai.github.io/)
|
@@ -184,7 +184,7 @@ def load_video(self, video_path, max_frames_num,fps=1,force_sample=False):
|
|
184 |
spare_frames = vr.get_batch(frame_idx).asnumpy()
|
185 |
# import pdb;pdb.set_trace()
|
186 |
return spare_frames,frame_time,video_time
|
187 |
-
pretrained = "lmms-lab/LLaVA-
|
188 |
model_name = "llava_qwen"
|
189 |
device = "cuda"
|
190 |
device_map = "auto"
|
|
|
10 |
tags:
|
11 |
- multimodal
|
12 |
model-index:
|
13 |
+
- name: LLaVA-Video-7B-Qwen2
|
14 |
results:
|
15 |
- task:
|
16 |
type: multimodal
|
|
|
116 |
- lmms-lab/llava-onevision-qwen2-7b-si
|
117 |
---
|
118 |
|
119 |
+
# LLaVA-Video-7B-Qwen2-Video-Only
|
120 |
|
121 |
## Table of Contents
|
122 |
|
|
|
132 |
In contrast to lmms-lab/LLaVA-NeXT-Video-7B-Qwen2, this is a 7B model trained on [LLaVA-Video-178K](https://huggingface.co/datasets/lmms-lab/LLaVA-NeXT-Video-SFT-Data) only, based on Qwen2 language model with a context window of 32K tokens.
|
133 |
|
134 |
|
135 |
+
This model supports up to 110 frames and achieves comparable results to those of lmms-lab/LLaVA-Video-7B-Qwen2 in terms of video benchmarks.
|
136 |
|
137 |
- **Repository:** [LLaVA-VL/LLaVA-NeXT](https://github.com/LLaVA-VL/LLaVA-NeXT?tab=readme-ov-file)
|
138 |
- **Point of Contact:** [Yuanhan Zhang](https://zhangyuanhan-ai.github.io/)
|
|
|
184 |
spare_frames = vr.get_batch(frame_idx).asnumpy()
|
185 |
# import pdb;pdb.set_trace()
|
186 |
return spare_frames,frame_time,video_time
|
187 |
+
pretrained = "lmms-lab/LLaVA-Video-7B-Qwen2-Video-Only "
|
188 |
model_name = "llava_qwen"
|
189 |
device = "cuda"
|
190 |
device_map = "auto"
|