Update modeling_videochat_flash.py
Browse files
modeling_videochat_flash.py
CHANGED
@@ -641,7 +641,7 @@ class VideoChatFlashQwenForCausalLM(LlavaMetaForCausalLM, Qwen2ForCausalLM_Flash
|
|
641 |
conv = conv_templates["qwen_2"].copy()
|
642 |
|
643 |
if chat_history is None or len(chat_history) == 0:
|
644 |
-
user_prompt = f'{DEFAULT_IMAGE_TOKEN}\n{time_msg.
|
645 |
else:
|
646 |
assert DEFAULT_IMAGE_TOKEN in chat_history[0]['content'], chat_history
|
647 |
for msg in chat_history:
|
|
|
641 |
conv = conv_templates["qwen_2"].copy()
|
642 |
|
643 |
if chat_history is None or len(chat_history) == 0:
|
644 |
+
user_prompt = f'{DEFAULT_IMAGE_TOKEN}\n{time_msg.strip()} {user_prompt}'
|
645 |
else:
|
646 |
assert DEFAULT_IMAGE_TOKEN in chat_history[0]['content'], chat_history
|
647 |
for msg in chat_history:
|