diff --git "a/mix_sft/charades.json" "b/mix_sft/charades.json" new file mode 100644--- /dev/null +++ "b/mix_sft/charades.json" @@ -0,0 +1,186122 @@ +[ + { + "video_id": "AO8RW", + "question_id": "AO8RW", + "video_file": "Charades/videos/AO8RW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is putting a book on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.9>." + } + ] + }, + { + "video_id": "Y6R7T", + "question_id": "Y6R7T", + "video_file": "Charades/videos/Y6R7T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins to play on a phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.8> to <30.0>." + } + ] + }, + { + "video_id": "Y6R7T", + "question_id": "Y6R7T", + "video_file": "Charades/videos/Y6R7T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person start playing on their phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.8> to <30.0>." + } + ] + }, + { + "video_id": "Y6R7T", + "question_id": "Y6R7T", + "video_file": "Charades/videos/Y6R7T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pouring it into a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <17.3>." + } + ] + }, + { + "video_id": "Y6R7T", + "question_id": "Y6R7T", + "video_file": "Charades/videos/Y6R7T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person pours some water into the glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.3> to <17.3>." + } + ] + }, + { + "video_id": "Y6R7T", + "question_id": "Y6R7T", + "video_file": "Charades/videos/Y6R7T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pours water into a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.3> to <17.3>." + } + ] + }, + { + "video_id": "IGDIE", + "question_id": "IGDIE", + "video_file": "Charades/videos/IGDIE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts watching television' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.7> to <34.0>." + } + ] + }, + { + "video_id": "7UPGT", + "question_id": "7UPGT", + "video_file": "Charades/videos/7UPGT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a cup out the fridge' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.9> to <19.7>." + } + ] + }, + { + "video_id": "KBPDM", + "question_id": "KBPDM", + "video_file": "Charades/videos/KBPDM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sneezing on a mirror' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.1> to <13.6>." + } + ] + }, + { + "video_id": "KBPDM", + "question_id": "KBPDM", + "video_file": "Charades/videos/KBPDM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is awakening' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.1>." + } + ] + }, + { + "video_id": "KBPDM", + "question_id": "KBPDM", + "video_file": "Charades/videos/KBPDM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'holding a small personal mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.1> to <18.7>." + } + ] + }, + { + "video_id": "IBIWF", + "question_id": "IBIWF", + "video_file": "Charades/videos/IBIWF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is playing with their camera' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "372CC", + "question_id": "372CC", + "video_file": "Charades/videos/372CC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting in front of a computer sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.5> to <29.9>." + } + ] + }, + { + "video_id": "372CC", + "question_id": "372CC", + "video_file": "Charades/videos/372CC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.5> to <29.9>." + } + ] + }, + { + "video_id": "N14BK", + "question_id": "N14BK", + "video_file": "Charades/videos/N14BK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put a notebook in a bag'." + }, + { + "from": "gpt", + "value": "From <39.3> to <45.0>." + } + ] + }, + { + "video_id": "NQT1S", + "question_id": "NQT1S", + "video_file": "Charades/videos/NQT1S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person all of a sudden they start sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.3> to <31.0>." + } + ] + }, + { + "video_id": "F1VEE", + "question_id": "F1VEE", + "video_file": "Charades/videos/F1VEE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.0> to <32.0>." + } + ] + }, + { + "video_id": "F1VEE", + "question_id": "F1VEE", + "video_file": "Charades/videos/F1VEE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put on their shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.1> to <23.8>." + } + ] + }, + { + "video_id": "F1VEE", + "question_id": "F1VEE", + "video_file": "Charades/videos/F1VEE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts on shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.1> to <23.8>." + } + ] + }, + { + "video_id": "F1VEE", + "question_id": "F1VEE", + "video_file": "Charades/videos/F1VEE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take a mobile phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.2> to <28.2>." + } + ] + }, + { + "video_id": "F1VEE", + "question_id": "F1VEE", + "video_file": "Charades/videos/F1VEE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put some shoes on'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.1> to <23.8>." + } + ] + }, + { + "video_id": "YDWN5", + "question_id": "YDWN5", + "video_file": "Charades/videos/YDWN5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person tidying a wardrobe in an entryway is smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.2> to <12.6>." + } + ] + }, + { + "video_id": "QRWQ3", + "question_id": "QRWQ3", + "video_file": "Charades/videos/QRWQ3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person cooking on the stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.1> to <29.0>." + } + ] + }, + { + "video_id": "QRWQ3", + "question_id": "QRWQ3", + "video_file": "Charades/videos/QRWQ3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is cooking something on a stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.1> to <29.0>." + } + ] + }, + { + "video_id": "SFHHR", + "question_id": "SFHHR", + "video_file": "Charades/videos/SFHHR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the coffee on the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.5> to <28.0>." + } + ] + }, + { + "video_id": "HWYTN", + "question_id": "HWYTN", + "video_file": "Charades/videos/HWYTN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person runs out the room'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "HWYTN", + "question_id": "HWYTN", + "video_file": "Charades/videos/HWYTN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another man running past' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "HWYTN", + "question_id": "HWYTN", + "video_file": "Charades/videos/HWYTN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person is running in shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "GG4UR", + "question_id": "GG4UR", + "video_file": "Charades/videos/GG4UR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens door to closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <15.4>." + } + ] + }, + { + "video_id": "GG4UR", + "question_id": "GG4UR", + "video_file": "Charades/videos/GG4UR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is seen watching television' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "GG4UR", + "question_id": "GG4UR", + "video_file": "Charades/videos/GG4UR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is watching television on the sofa' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "GG4UR", + "question_id": "GG4UR", + "video_file": "Charades/videos/GG4UR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person is sitting on couch watching television' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "GG4UR", + "question_id": "GG4UR", + "video_file": "Charades/videos/GG4UR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person seen opening the door to a hallway closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <15.4>." + } + ] + }, + { + "video_id": "XDVAY", + "question_id": "XDVAY", + "video_file": "Charades/videos/XDVAY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is awakening in bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "MLBTH", + "question_id": "MLBTH", + "video_file": "Charades/videos/MLBTH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "MLBTH", + "question_id": "MLBTH", + "video_file": "Charades/videos/MLBTH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "MLBTH", + "question_id": "MLBTH", + "video_file": "Charades/videos/MLBTH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens up a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "IR31O", + "question_id": "IR31O", + "video_file": "Charades/videos/IR31O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.6> to <12.9>." + } + ] + }, + { + "video_id": "IR31O", + "question_id": "IR31O", + "video_file": "Charades/videos/IR31O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is opening a book'." + }, + { + "from": "gpt", + "value": "From <1.3> to <7.7>." + } + ] + }, + { + "video_id": "KX8VW", + "question_id": "KX8VW", + "video_file": "Charades/videos/KX8VW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put back on the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.3> to <22.1>." + } + ] + }, + { + "video_id": "KX8VW", + "question_id": "KX8VW", + "video_file": "Charades/videos/KX8VW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours some water into a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.9> to <15.4>." + } + ] + }, + { + "video_id": "KX8VW", + "question_id": "KX8VW", + "video_file": "Charades/videos/KX8VW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sitting on the sofa eating out of a dish' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "NVBBD", + "question_id": "NVBBD", + "video_file": "Charades/videos/NVBBD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens up a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "NVBBD", + "question_id": "NVBBD", + "video_file": "Charades/videos/NVBBD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts his shoes on' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.1> to <32.0>." + } + ] + }, + { + "video_id": "NVBBD", + "question_id": "NVBBD", + "video_file": "Charades/videos/NVBBD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts on a shoe' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.1> to <32.0>." + } + ] + }, + { + "video_id": "NVBBD", + "question_id": "NVBBD", + "video_file": "Charades/videos/NVBBD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they open the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "NVBBD", + "question_id": "NVBBD", + "video_file": "Charades/videos/NVBBD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person start putting shoes on'." + }, + { + "from": "gpt", + "value": "From <19.1> to <32.0>." + } + ] + }, + { + "video_id": "NVBBD", + "question_id": "NVBBD", + "video_file": "Charades/videos/NVBBD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a door by grasping a doorknob'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "4C4LC", + "question_id": "4C4LC", + "video_file": "Charades/videos/4C4LC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person also turning off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <27.3> to <32.0>." + } + ] + }, + { + "video_id": "J4GX8", + "question_id": "J4GX8", + "video_file": "Charades/videos/J4GX8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking a glass of water'." + }, + { + "from": "gpt", + "value": "From <18.8> to <27.3>." + } + ] + }, + { + "video_id": "J4GX8", + "question_id": "J4GX8", + "video_file": "Charades/videos/J4GX8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.4>." + } + ] + }, + { + "video_id": "J4GX8", + "question_id": "J4GX8", + "video_file": "Charades/videos/J4GX8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person appear to take some medicine by swallowing it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.6> to <22.2>." + } + ] + }, + { + "video_id": "J4GX8", + "question_id": "J4GX8", + "video_file": "Charades/videos/J4GX8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person can take their medicine' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.6> to <22.2>." + } + ] + }, + { + "video_id": "HPEE5", + "question_id": "HPEE5", + "video_file": "Charades/videos/HPEE5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person close the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.3> to <25.0>." + } + ] + }, + { + "video_id": "HPEE5", + "question_id": "HPEE5", + "video_file": "Charades/videos/HPEE5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person open a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.6> to <21.4>." + } + ] + }, + { + "video_id": "MBGPJ", + "question_id": "MBGPJ", + "video_file": "Charades/videos/MBGPJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person a guy is sitting in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.8>." + } + ] + }, + { + "video_id": "MBGPJ", + "question_id": "MBGPJ", + "video_file": "Charades/videos/MBGPJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes out a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.7> to <20.9>." + } + ] + }, + { + "video_id": "MBGPJ", + "question_id": "MBGPJ", + "video_file": "Charades/videos/MBGPJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.5> to <18.1>." + } + ] + }, + { + "video_id": "NEM29", + "question_id": "NEM29", + "video_file": "Charades/videos/NEM29.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person watches television' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.0> to <17.8>." + } + ] + }, + { + "video_id": "NEM29", + "question_id": "NEM29", + "video_file": "Charades/videos/NEM29.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is undressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "SA7GL", + "question_id": "SA7GL", + "video_file": "Charades/videos/SA7GL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they close the laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.6> to <17.0>." + } + ] + }, + { + "video_id": "SA7GL", + "question_id": "SA7GL", + "video_file": "Charades/videos/SA7GL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is playing on a laptop'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "SA7GL", + "question_id": "SA7GL", + "video_file": "Charades/videos/SA7GL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sits in a chair working on a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "0QA8P", + "question_id": "0QA8P", + "video_file": "Charades/videos/0QA8P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person undressed from all clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.3> to <28.1>." + } + ] + }, + { + "video_id": "0QA8P", + "question_id": "0QA8P", + "video_file": "Charades/videos/0QA8P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put the clothes'." + }, + { + "from": "gpt", + "value": "From <38.1> to <43.6>." + } + ] + }, + { + "video_id": "0QA8P", + "question_id": "0QA8P", + "video_file": "Charades/videos/0QA8P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.8>." + } + ] + }, + { + "video_id": "BWHUX", + "question_id": "BWHUX", + "video_file": "Charades/videos/BWHUX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person he puts something on a table next to dishes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.7> to <10.1>." + } + ] + }, + { + "video_id": "KJ8G2", + "question_id": "KJ8G2", + "video_file": "Charades/videos/KJ8G2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting them on a table'." + }, + { + "from": "gpt", + "value": "From <5.2> to <10.4>." + } + ] + }, + { + "video_id": "D87UN", + "question_id": "D87UN", + "video_file": "Charades/videos/D87UN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they are also drinking from a cup'." + }, + { + "from": "gpt", + "value": "From <1.5> to <11.4>." + } + ] + }, + { + "video_id": "D87UN", + "question_id": "D87UN", + "video_file": "Charades/videos/D87UN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they're drinking a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.5> to <11.4>." + } + ] + }, + { + "video_id": "XJOJL", + "question_id": "XJOJL", + "video_file": "Charades/videos/XJOJL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the sandwich on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.4> to <35.1>." + } + ] + }, + { + "video_id": "XJOJL", + "question_id": "XJOJL", + "video_file": "Charades/videos/XJOJL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person snuggles up in a freshly cleaned blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <45.6> to <49.0>." + } + ] + }, + { + "video_id": "67MSU", + "question_id": "67MSU", + "video_file": "Charades/videos/67MSU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.7> to <31.0>." + } + ] + }, + { + "video_id": "67MSU", + "question_id": "67MSU", + "video_file": "Charades/videos/67MSU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they open the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "67MSU", + "question_id": "67MSU", + "video_file": "Charades/videos/67MSU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "9NRLA", + "question_id": "9NRLA", + "video_file": "Charades/videos/9NRLA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats it'." + }, + { + "from": "gpt", + "value": "From <13.6> to <19.7>." + } + ] + }, + { + "video_id": "9NRLA", + "question_id": "9NRLA", + "video_file": "Charades/videos/9NRLA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they stop to eat some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.6> to <19.7>." + } + ] + }, + { + "video_id": "9NRLA", + "question_id": "9NRLA", + "video_file": "Charades/videos/9NRLA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person puts clothes into a washing machine' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.1>." + } + ] + }, + { + "video_id": "I6NJ6", + "question_id": "I6NJ6", + "video_file": "Charades/videos/I6NJ6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person awakens in bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.7> to <13.4>." + } + ] + }, + { + "video_id": "KG20W", + "question_id": "KG20W", + "video_file": "Charades/videos/KG20W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a woman walks around holding a blanket/pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <31.8>." + } + ] + }, + { + "video_id": "68JOO", + "question_id": "68JOO", + "video_file": "Charades/videos/68JOO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they stand up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "68JOO", + "question_id": "68JOO", + "video_file": "Charades/videos/68JOO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <16.1>." + } + ] + }, + { + "video_id": "68JOO", + "question_id": "68JOO", + "video_file": "Charades/videos/68JOO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sit back down in the chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.1> to <16.1>." + } + ] + }, + { + "video_id": "68JOO", + "question_id": "68JOO", + "video_file": "Charades/videos/68JOO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sitting in a chair snuggles a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.1> to <16.1>." + } + ] + }, + { + "video_id": "68JOO", + "question_id": "68JOO", + "video_file": "Charades/videos/68JOO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person snuggles with a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.8> to <19.6>." + } + ] + }, + { + "video_id": "QHARF", + "question_id": "QHARF", + "video_file": "Charades/videos/QHARF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person holds a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.3>." + } + ] + }, + { + "video_id": "QHARF", + "question_id": "QHARF", + "video_file": "Charades/videos/QHARF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person one holding a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.3>." + } + ] + }, + { + "video_id": "QHARF", + "question_id": "QHARF", + "video_file": "Charades/videos/QHARF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is holding a bag of food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.3>." + } + ] + }, + { + "video_id": "P1EK9", + "question_id": "P1EK9", + "video_file": "Charades/videos/P1EK9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throwing his bag down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.8> to <27.0>." + } + ] + }, + { + "video_id": "P1EK9", + "question_id": "P1EK9", + "video_file": "Charades/videos/P1EK9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they throw it on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.8> to <27.0>." + } + ] + }, + { + "video_id": "P1EK9", + "question_id": "P1EK9", + "video_file": "Charades/videos/P1EK9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they throw the bag on the floor too'." + }, + { + "from": "gpt", + "value": "From <20.8> to <27.0>." + } + ] + }, + { + "video_id": "P1EK9", + "question_id": "P1EK9", + "video_file": "Charades/videos/P1EK9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person walks into a room opens a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.8> to <11.2>." + } + ] + }, + { + "video_id": "KZSA9", + "question_id": "KZSA9", + "video_file": "Charades/videos/KZSA9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <17.7>." + } + ] + }, + { + "video_id": "KZSA9", + "question_id": "KZSA9", + "video_file": "Charades/videos/KZSA9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a blanket out' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.6> to <30.9>." + } + ] + }, + { + "video_id": "KZSA9", + "question_id": "KZSA9", + "video_file": "Charades/videos/KZSA9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <29.6>." + } + ] + }, + { + "video_id": "LQ0LM", + "question_id": "LQ0LM", + "video_file": "Charades/videos/LQ0LM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person close the cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.1> to <17.9>." + } + ] + }, + { + "video_id": "LQ0LM", + "question_id": "LQ0LM", + "video_file": "Charades/videos/LQ0LM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <21.0>." + } + ] + }, + { + "video_id": "LQ0LM", + "question_id": "LQ0LM", + "video_file": "Charades/videos/LQ0LM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open the cabinet below' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.0> to <19.0>." + } + ] + }, + { + "video_id": "ORW6Y", + "question_id": "ORW6Y", + "video_file": "Charades/videos/ORW6Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person puts a box on a shelf'." + }, + { + "from": "gpt", + "value": "From <18.7> to <25.8>." + } + ] + }, + { + "video_id": "ORW6Y", + "question_id": "ORW6Y", + "video_file": "Charades/videos/ORW6Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person holding a pillow starts sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <31.0>." + } + ] + }, + { + "video_id": "D8HN1", + "question_id": "D8HN1", + "video_file": "Charades/videos/D8HN1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eat a sandwich they walk out of the bathroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.2> to <13.1>." + } + ] + }, + { + "video_id": "D8HN1", + "question_id": "D8HN1", + "video_file": "Charades/videos/D8HN1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating a sandwich in front of a mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.2> to <13.1>." + } + ] + }, + { + "video_id": "D8HN1", + "question_id": "D8HN1", + "video_file": "Charades/videos/D8HN1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person use their foot to close the door'." + }, + { + "from": "gpt", + "value": "From <23.9> to <28.9>." + } + ] + }, + { + "video_id": "D8HN1", + "question_id": "D8HN1", + "video_file": "Charades/videos/D8HN1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person close a door with their feet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.9> to <28.9>." + } + ] + }, + { + "video_id": "BY6ZZ", + "question_id": "BY6ZZ", + "video_file": "Charades/videos/BY6ZZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person lays on some clothes on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.5> to <27.0>." + } + ] + }, + { + "video_id": "T8B7B", + "question_id": "T8B7B", + "video_file": "Charades/videos/T8B7B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the camera in a box on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.6> to <25.4>." + } + ] + }, + { + "video_id": "T8B7B", + "question_id": "T8B7B", + "video_file": "Charades/videos/T8B7B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person closes the box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.3> to <28.8>." + } + ] + }, + { + "video_id": "T8B7B", + "question_id": "T8B7B", + "video_file": "Charades/videos/T8B7B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the coffee cup in a box'." + }, + { + "from": "gpt", + "value": "From <20.0> to <27.7>." + } + ] + }, + { + "video_id": "D8OSC", + "question_id": "D8OSC", + "video_file": "Charades/videos/D8OSC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.1> to <17.9>." + } + ] + }, + { + "video_id": "TGDWN", + "question_id": "TGDWN", + "video_file": "Charades/videos/TGDWN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.4> to <29.2>." + } + ] + }, + { + "video_id": "X2Q00", + "question_id": "X2Q00", + "video_file": "Charades/videos/X2Q00.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person after eating' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <9.6>." + } + ] + }, + { + "video_id": "X2Q00", + "question_id": "X2Q00", + "video_file": "Charades/videos/X2Q00.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they are on their laptop eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <1.4> to <9.6>." + } + ] + }, + { + "video_id": "X2Q00", + "question_id": "X2Q00", + "video_file": "Charades/videos/X2Q00.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person debating on tidying up or reading a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.7> to <33.8>." + } + ] + }, + { + "video_id": "S8TT8", + "question_id": "S8TT8", + "video_file": "Charades/videos/S8TT8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sit down in a chair'." + }, + { + "from": "gpt", + "value": "From <23.5> to <31.6>." + } + ] + }, + { + "video_id": "S8TT8", + "question_id": "S8TT8", + "video_file": "Charades/videos/S8TT8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts putting away clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.3> to <23.6>." + } + ] + }, + { + "video_id": "S8TT8", + "question_id": "S8TT8", + "video_file": "Charades/videos/S8TT8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person sits down on a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.5> to <31.6>." + } + ] + }, + { + "video_id": "S8TT8", + "question_id": "S8TT8", + "video_file": "Charades/videos/S8TT8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts some clothes on a bench' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.3> to <23.6>." + } + ] + }, + { + "video_id": "QG4N1", + "question_id": "QG4N1", + "video_file": "Charades/videos/QG4N1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put it on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.2> to <31.0>." + } + ] + }, + { + "video_id": "QG4N1", + "question_id": "QG4N1", + "video_file": "Charades/videos/QG4N1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the camera on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <31.0>." + } + ] + }, + { + "video_id": "AOMNM", + "question_id": "AOMNM", + "video_file": "Charades/videos/AOMNM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs into the closet'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "X8MNV", + "question_id": "X8MNV", + "video_file": "Charades/videos/X8MNV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks through a doorway into a living room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <8.1>." + } + ] + }, + { + "video_id": "NHUV1", + "question_id": "NHUV1", + "video_file": "Charades/videos/NHUV1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opening a bag'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.1>." + } + ] + }, + { + "video_id": "NNG97", + "question_id": "NNG97", + "video_file": "Charades/videos/NNG97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts a book on a shelf'." + }, + { + "from": "gpt", + "value": "From <1.4> to <6.6>." + } + ] + }, + { + "video_id": "NNG97", + "question_id": "NNG97", + "video_file": "Charades/videos/NNG97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts the book back on the shelf'." + }, + { + "from": "gpt", + "value": "From <1.4> to <6.6>." + } + ] + }, + { + "video_id": "NNG97", + "question_id": "NNG97", + "video_file": "Charades/videos/NNG97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person throws a book onto the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.6>." + } + ] + }, + { + "video_id": "A1NKP", + "question_id": "A1NKP", + "video_file": "Charades/videos/A1NKP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a bowl of soup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <17.6>." + } + ] + }, + { + "video_id": "EDXBD", + "question_id": "EDXBD", + "video_file": "Charades/videos/EDXBD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walks over to open a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.2> to <14.0>." + } + ] + }, + { + "video_id": "BW5MF", + "question_id": "BW5MF", + "video_file": "Charades/videos/BW5MF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "UZMMO", + "question_id": "UZMMO", + "video_file": "Charades/videos/UZMMO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits on the bed to read it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.9> to <37.0>." + } + ] + }, + { + "video_id": "UZMMO", + "question_id": "UZMMO", + "video_file": "Charades/videos/UZMMO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes a book from a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.4> to <25.1>." + } + ] + }, + { + "video_id": "UZMMO", + "question_id": "UZMMO", + "video_file": "Charades/videos/UZMMO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person sat down to read a book'." + }, + { + "from": "gpt", + "value": "From <22.5> to <37.0>." + } + ] + }, + { + "video_id": "CHBAS", + "question_id": "CHBAS", + "video_file": "Charades/videos/CHBAS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person throws their clothes in a pantry' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <13.2>." + } + ] + }, + { + "video_id": "CHBAS", + "question_id": "CHBAS", + "video_file": "Charades/videos/CHBAS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is in a bathroom doorway undressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "M2F66", + "question_id": "M2F66", + "video_file": "Charades/videos/M2F66.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they close the door to the laundry room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.1> to <23.0>." + } + ] + }, + { + "video_id": "M2F66", + "question_id": "M2F66", + "video_file": "Charades/videos/M2F66.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <17.1> to <23.0>." + } + ] + }, + { + "video_id": "M2F66", + "question_id": "M2F66", + "video_file": "Charades/videos/M2F66.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts detergent on the clothes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <23.5>." + } + ] + }, + { + "video_id": "M2F66", + "question_id": "M2F66", + "video_file": "Charades/videos/M2F66.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is putting clothes in the washing machine' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <23.5>." + } + ] + }, + { + "video_id": "M2F66", + "question_id": "M2F66", + "video_file": "Charades/videos/M2F66.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts washing a load of clothes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.6> to <24.5>." + } + ] + }, + { + "video_id": "M2F66", + "question_id": "M2F66", + "video_file": "Charades/videos/M2F66.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they close the washer door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.1> to <23.0>." + } + ] + }, + { + "video_id": "KMZHU", + "question_id": "KMZHU", + "video_file": "Charades/videos/KMZHU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pouring a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <11.0>." + } + ] + }, + { + "video_id": "0CG15", + "question_id": "0CG15", + "video_file": "Charades/videos/0CG15.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <14.3>." + } + ] + }, + { + "video_id": "0CG15", + "question_id": "0CG15", + "video_file": "Charades/videos/0CG15.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person he opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.8> to <29.7>." + } + ] + }, + { + "video_id": "0CG15", + "question_id": "0CG15", + "video_file": "Charades/videos/0CG15.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closing a cabinet'." + }, + { + "from": "gpt", + "value": "From <25.8> to <33.3>." + } + ] + }, + { + "video_id": "9XXJ7", + "question_id": "9XXJ7", + "video_file": "Charades/videos/9XXJ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <8.2>." + } + ] + }, + { + "video_id": "9XXJ7", + "question_id": "9XXJ7", + "video_file": "Charades/videos/9XXJ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.6> to <6.8>." + } + ] + }, + { + "video_id": "HKBKA", + "question_id": "HKBKA", + "video_file": "Charades/videos/HKBKA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person lies on a bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.9> to <24.9>." + } + ] + }, + { + "video_id": "N39RU", + "question_id": "N39RU", + "video_file": "Charades/videos/N39RU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person get up from their seats' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "E0MK4", + "question_id": "E0MK4", + "video_file": "Charades/videos/E0MK4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take a book out their bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <13.8>." + } + ] + }, + { + "video_id": "ECMQU", + "question_id": "ECMQU", + "video_file": "Charades/videos/ECMQU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinks out of a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.8> to <25.6>." + } + ] + }, + { + "video_id": "ECMQU", + "question_id": "ECMQU", + "video_file": "Charades/videos/ECMQU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a smiling person is dressing in their basement' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "ECMQU", + "question_id": "ECMQU", + "video_file": "Charades/videos/ECMQU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting on clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.0>." + } + ] + }, + { + "video_id": "M6LFI", + "question_id": "M6LFI", + "video_file": "Charades/videos/M6LFI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is undressing in front of a mirror' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.0> to <8.0>." + } + ] + }, + { + "video_id": "M6LFI", + "question_id": "M6LFI", + "video_file": "Charades/videos/M6LFI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening a sandwich in a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <9.5>." + } + ] + }, + { + "video_id": "618HP", + "question_id": "618HP", + "video_file": "Charades/videos/618HP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sitting on chair jumps up turns off light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "618HP", + "question_id": "618HP", + "video_file": "Charades/videos/618HP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person the stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <21.2>." + } + ] + }, + { + "video_id": "618HP", + "question_id": "618HP", + "video_file": "Charades/videos/618HP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turn off the light'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "UVEPN", + "question_id": "UVEPN", + "video_file": "Charades/videos/UVEPN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking a cup of coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.8> to <17.0>." + } + ] + }, + { + "video_id": "Q8HQM", + "question_id": "Q8HQM", + "video_file": "Charades/videos/Q8HQM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person the other one keeps sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.3> to <10.6>." + } + ] + }, + { + "video_id": "GJJA1", + "question_id": "GJJA1", + "video_file": "Charades/videos/GJJA1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person plays with a phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "GJJA1", + "question_id": "GJJA1", + "video_file": "Charades/videos/GJJA1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person start playing with their phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "GJJA1", + "question_id": "GJJA1", + "video_file": "Charades/videos/GJJA1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person holds a pillow'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.2>." + } + ] + }, + { + "video_id": "HMSM8", + "question_id": "HMSM8", + "video_file": "Charades/videos/HMSM8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <24.8>." + } + ] + }, + { + "video_id": "HMSM8", + "question_id": "HMSM8", + "video_file": "Charades/videos/HMSM8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person opens the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <24.8>." + } + ] + }, + { + "video_id": "HMSM8", + "question_id": "HMSM8", + "video_file": "Charades/videos/HMSM8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they open the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <24.8>." + } + ] + }, + { + "video_id": "3P8NL", + "question_id": "3P8NL", + "video_file": "Charades/videos/3P8NL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sitting at a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.4>." + } + ] + }, + { + "video_id": "3P8NL", + "question_id": "3P8NL", + "video_file": "Charades/videos/3P8NL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person sits in a chair'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.1>." + } + ] + }, + { + "video_id": "3P8NL", + "question_id": "3P8NL", + "video_file": "Charades/videos/3P8NL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes out the cards from the box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.3>." + } + ] + }, + { + "video_id": "VSS68", + "question_id": "VSS68", + "video_file": "Charades/videos/VSS68.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person put groceries in the cabinet in the kitchen' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <8.3>." + } + ] + }, + { + "video_id": "VSS68", + "question_id": "VSS68", + "video_file": "Charades/videos/VSS68.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person was opening the cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.7> to <6.1>." + } + ] + }, + { + "video_id": "P4DL9", + "question_id": "P4DL9", + "video_file": "Charades/videos/P4DL9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the refrigerator door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "P4DL9", + "question_id": "P4DL9", + "video_file": "Charades/videos/P4DL9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts cooking at the stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.2> to <18.1>." + } + ] + }, + { + "video_id": "P4DL9", + "question_id": "P4DL9", + "video_file": "Charades/videos/P4DL9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts cooking with a pan' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.2> to <18.1>." + } + ] + }, + { + "video_id": "P4DL9", + "question_id": "P4DL9", + "video_file": "Charades/videos/P4DL9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person closes the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "5TLQE", + "question_id": "5TLQE", + "video_file": "Charades/videos/5TLQE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person start to close the book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.9> to <22.9>." + } + ] + }, + { + "video_id": "5TLQE", + "question_id": "5TLQE", + "video_file": "Charades/videos/5TLQE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.9> to <22.9>." + } + ] + }, + { + "video_id": "LTJEO", + "question_id": "LTJEO", + "video_file": "Charades/videos/LTJEO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person places the pair of shoes on the dresser' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <32.0>." + } + ] + }, + { + "video_id": "LTJEO", + "question_id": "LTJEO", + "video_file": "Charades/videos/LTJEO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts dressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.0> to <32.0>." + } + ] + }, + { + "video_id": "LTJEO", + "question_id": "LTJEO", + "video_file": "Charades/videos/LTJEO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.2>." + } + ] + }, + { + "video_id": "LTJEO", + "question_id": "LTJEO", + "video_file": "Charades/videos/LTJEO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a smiling person is opening the wardrobe door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.2>." + } + ] + }, + { + "video_id": "LTJEO", + "question_id": "LTJEO", + "video_file": "Charades/videos/LTJEO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes out some clothes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.2> to <19.0>." + } + ] + }, + { + "video_id": "GIIMN", + "question_id": "GIIMN", + "video_file": "Charades/videos/GIIMN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating in the bathroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.8> to <23.2>." + } + ] + }, + { + "video_id": "CBVUS", + "question_id": "CBVUS", + "video_file": "Charades/videos/CBVUS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sneezing in the pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <7.5>." + } + ] + }, + { + "video_id": "CBVUS", + "question_id": "CBVUS", + "video_file": "Charades/videos/CBVUS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.1> to <7.5>." + } + ] + }, + { + "video_id": "DPJH4", + "question_id": "DPJH4", + "video_file": "Charades/videos/DPJH4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <12.7>." + } + ] + }, + { + "video_id": "DPJH4", + "question_id": "DPJH4", + "video_file": "Charades/videos/DPJH4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes a can of food from a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.8> to <21.9>." + } + ] + }, + { + "video_id": "W0RSP", + "question_id": "W0RSP", + "video_file": "Charades/videos/W0RSP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakening in their bedroom'." + }, + { + "from": "gpt", + "value": "From <1.2> to <6.9>." + } + ] + }, + { + "video_id": "IH3GN", + "question_id": "IH3GN", + "video_file": "Charades/videos/IH3GN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is walking through a doorway holding a towel' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "VG94P", + "question_id": "VG94P", + "video_file": "Charades/videos/VG94P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'when they start laughing at something another person said' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <33.0>." + } + ] + }, + { + "video_id": "VG94P", + "question_id": "VG94P", + "video_file": "Charades/videos/VG94P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they both laugh'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <33.0>." + } + ] + }, + { + "video_id": "92TUQ", + "question_id": "92TUQ", + "video_file": "Charades/videos/92TUQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes off shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.0> to <11.0>." + } + ] + }, + { + "video_id": "92TUQ", + "question_id": "92TUQ", + "video_file": "Charades/videos/92TUQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person holding a bag takes off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.0> to <11.0>." + } + ] + }, + { + "video_id": "YEHFQ", + "question_id": "YEHFQ", + "video_file": "Charades/videos/YEHFQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person enters the front door holding a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.1> to <22.6>." + } + ] + }, + { + "video_id": "YEHFQ", + "question_id": "YEHFQ", + "video_file": "Charades/videos/YEHFQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.7> to <32.0>." + } + ] + }, + { + "video_id": "YEHFQ", + "question_id": "YEHFQ", + "video_file": "Charades/videos/YEHFQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person starts eating something'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.7> to <32.0>." + } + ] + }, + { + "video_id": "1E1VE", + "question_id": "1E1VE", + "video_file": "Charades/videos/1E1VE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person enters holding a towel'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <12.5>." + } + ] + }, + { + "video_id": "813XE", + "question_id": "813XE", + "video_file": "Charades/videos/813XE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is holding a camera'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "813XE", + "question_id": "813XE", + "video_file": "Charades/videos/813XE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person open a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.0> to <11.0>." + } + ] + }, + { + "video_id": "813XE", + "question_id": "813XE", + "video_file": "Charades/videos/813XE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a cabinet in a living room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "813XE", + "question_id": "813XE", + "video_file": "Charades/videos/813XE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put the camera in a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.0> to <6.0>." + } + ] + }, + { + "video_id": "U896B", + "question_id": "U896B", + "video_file": "Charades/videos/U896B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person undressing beside the chair next to the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "ZS1P1", + "question_id": "ZS1P1", + "video_file": "Charades/videos/ZS1P1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person works on a laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.6> to <35.2>." + } + ] + }, + { + "video_id": "ZS1P1", + "question_id": "ZS1P1", + "video_file": "Charades/videos/ZS1P1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person occasionally looking at an opened book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.8> to <37.8>." + } + ] + }, + { + "video_id": "ZS1P1", + "question_id": "ZS1P1", + "video_file": "Charades/videos/ZS1P1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person sits working on a laptop laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.6> to <35.2>." + } + ] + }, + { + "video_id": "FIJ3D", + "question_id": "FIJ3D", + "video_file": "Charades/videos/FIJ3D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes a book off of a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "FIJ3D", + "question_id": "FIJ3D", + "video_file": "Charades/videos/FIJ3D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the book down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <22.9>." + } + ] + }, + { + "video_id": "FIJ3D", + "question_id": "FIJ3D", + "video_file": "Charades/videos/FIJ3D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes a book off of a shelf,turns'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "FIJ3D", + "question_id": "FIJ3D", + "video_file": "Charades/videos/FIJ3D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <26.7>." + } + ] + }, + { + "video_id": "FIJ3D", + "question_id": "FIJ3D", + "video_file": "Charades/videos/FIJ3D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person plays with a laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.6> to <31.0>." + } + ] + }, + { + "video_id": "FIJ3D", + "question_id": "FIJ3D", + "video_file": "Charades/videos/FIJ3D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <11.1>." + } + ] + }, + { + "video_id": "0G2SC", + "question_id": "0G2SC", + "video_file": "Charades/videos/0G2SC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person #1 smiled as person #2 walked through the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <18.8>." + } + ] + }, + { + "video_id": "TANH9", + "question_id": "TANH9", + "video_file": "Charades/videos/TANH9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person is smiling at food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <22.6>." + } + ] + }, + { + "video_id": "WZN0I", + "question_id": "WZN0I", + "video_file": "Charades/videos/WZN0I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throwing them on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <7.7>." + } + ] + }, + { + "video_id": "WZN0I", + "question_id": "WZN0I", + "video_file": "Charades/videos/WZN0I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person throws a pillow through the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <8.2>." + } + ] + }, + { + "video_id": "WZN0I", + "question_id": "WZN0I", + "video_file": "Charades/videos/WZN0I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person holding a pillow throws it on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "1LARL", + "question_id": "1LARL", + "video_file": "Charades/videos/1LARL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes a sandwich'." + }, + { + "from": "gpt", + "value": "From <3.6> to <10.2>." + } + ] + }, + { + "video_id": "MMFXQ", + "question_id": "MMFXQ", + "video_file": "Charades/videos/MMFXQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts them on a shelf in a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.3> to <9.4>." + } + ] + }, + { + "video_id": "MMFXQ", + "question_id": "MMFXQ", + "video_file": "Charades/videos/MMFXQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.1> to <6.4>." + } + ] + }, + { + "video_id": "9INKU", + "question_id": "9INKU", + "video_file": "Charades/videos/9INKU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is running back' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.0> to <15.7>." + } + ] + }, + { + "video_id": "0S9KN", + "question_id": "0S9KN", + "video_file": "Charades/videos/0S9KN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws the pillow onto the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.7> to <16.8>." + } + ] + }, + { + "video_id": "4UYGY", + "question_id": "4UYGY", + "video_file": "Charades/videos/4UYGY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they awaken' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <19.3>." + } + ] + }, + { + "video_id": "4UYGY", + "question_id": "4UYGY", + "video_file": "Charades/videos/4UYGY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a phone off a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.5> to <22.7>." + } + ] + }, + { + "video_id": "4UYGY", + "question_id": "4UYGY", + "video_file": "Charades/videos/4UYGY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person awakens' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <19.3>." + } + ] + }, + { + "video_id": "4UYGY", + "question_id": "4UYGY", + "video_file": "Charades/videos/4UYGY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person awakens on a bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <19.3>." + } + ] + }, + { + "video_id": "QUGSS", + "question_id": "QUGSS", + "video_file": "Charades/videos/QUGSS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <11.2>." + } + ] + }, + { + "video_id": "WEYV3", + "question_id": "WEYV3", + "video_file": "Charades/videos/WEYV3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is seen eating a sandwich at their desk' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "WEYV3", + "question_id": "WEYV3", + "video_file": "Charades/videos/WEYV3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is eating' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.9>." + } + ] + }, + { + "video_id": "WEYV3", + "question_id": "WEYV3", + "video_file": "Charades/videos/WEYV3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating a snack' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.9>." + } + ] + }, + { + "video_id": "WEYV3", + "question_id": "WEYV3", + "video_file": "Charades/videos/WEYV3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person playing on their laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.5>." + } + ] + }, + { + "video_id": "WEYV3", + "question_id": "WEYV3", + "video_file": "Charades/videos/WEYV3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person talking on their phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.8> to <30.0>." + } + ] + }, + { + "video_id": "V6IIE", + "question_id": "V6IIE", + "video_file": "Charades/videos/V6IIE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they later sit on a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.7> to <31.0>." + } + ] + }, + { + "video_id": "V6IIE", + "question_id": "V6IIE", + "video_file": "Charades/videos/V6IIE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person undresses in the bedroom doorway'." + }, + { + "from": "gpt", + "value": "From <14.7> to <21.7>." + } + ] + }, + { + "video_id": "V6IIE", + "question_id": "V6IIE", + "video_file": "Charades/videos/V6IIE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits down on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.7> to <31.0>." + } + ] + }, + { + "video_id": "V6IIE", + "question_id": "V6IIE", + "video_file": "Charades/videos/V6IIE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person smiling at each other' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.0>." + } + ] + }, + { + "video_id": "V6IIE", + "question_id": "V6IIE", + "video_file": "Charades/videos/V6IIE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person sits in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.7> to <31.0>." + } + ] + }, + { + "video_id": "ZHOP3", + "question_id": "ZHOP3", + "video_file": "Charades/videos/ZHOP3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person takes a blanket from a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.5> to <11.1>." + } + ] + }, + { + "video_id": "ZHOP3", + "question_id": "ZHOP3", + "video_file": "Charades/videos/ZHOP3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes out a blanket from a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.5> to <11.1>." + } + ] + }, + { + "video_id": "1G9R7", + "question_id": "1G9R7", + "video_file": "Charades/videos/1G9R7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person pours a drink into a cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.7> to <12.6>." + } + ] + }, + { + "video_id": "HPAYB", + "question_id": "HPAYB", + "video_file": "Charades/videos/HPAYB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes some clothes out of a bin'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <12.7>." + } + ] + }, + { + "video_id": "SKLEN", + "question_id": "SKLEN", + "video_file": "Charades/videos/SKLEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the cleaner back on the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <15.0>." + } + ] + }, + { + "video_id": "SKLEN", + "question_id": "SKLEN", + "video_file": "Charades/videos/SKLEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person the tidying a shelf in the garage'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.1> to <21.4>." + } + ] + }, + { + "video_id": "SKLEN", + "question_id": "SKLEN", + "video_file": "Charades/videos/SKLEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the stuff back on the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <15.0>." + } + ] + }, + { + "video_id": "L347E", + "question_id": "L347E", + "video_file": "Charades/videos/L347E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person vegins to throw a broom on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.3> to <33.8>." + } + ] + }, + { + "video_id": "B5YYS", + "question_id": "B5YYS", + "video_file": "Charades/videos/B5YYS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person undresses quickly' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.5> to <15.5>." + } + ] + }, + { + "video_id": "2074D", + "question_id": "2074D", + "video_file": "Charades/videos/2074D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person turns the light on' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.3> to <11.2>." + } + ] + }, + { + "video_id": "2074D", + "question_id": "2074D", + "video_file": "Charades/videos/2074D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person flips on a light switch'." + }, + { + "from": "gpt", + "value": "From <3.3> to <11.2>." + } + ] + }, + { + "video_id": "9PAQ4", + "question_id": "9PAQ4", + "video_file": "Charades/videos/9PAQ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes off their shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.9> to <16.7>." + } + ] + }, + { + "video_id": "9PAQ4", + "question_id": "9PAQ4", + "video_file": "Charades/videos/9PAQ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take off their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.9> to <16.7>." + } + ] + }, + { + "video_id": "9PAQ4", + "question_id": "9PAQ4", + "video_file": "Charades/videos/9PAQ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take their shoes off' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.9> to <16.7>." + } + ] + }, + { + "video_id": "6AJX0", + "question_id": "6AJX0", + "video_file": "Charades/videos/6AJX0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a door'." + }, + { + "from": "gpt", + "value": "From <1.1> to <6.5>." + } + ] + }, + { + "video_id": "6AJX0", + "question_id": "6AJX0", + "video_file": "Charades/videos/6AJX0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the door back' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.8> to <13.2>." + } + ] + }, + { + "video_id": "6AJX0", + "question_id": "6AJX0", + "video_file": "Charades/videos/6AJX0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person close the door of the pantry' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.8> to <13.2>." + } + ] + }, + { + "video_id": "ACMHK", + "question_id": "ACMHK", + "video_file": "Charades/videos/ACMHK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person fix their hair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <23.0>." + } + ] + }, + { + "video_id": "TQGPM", + "question_id": "TQGPM", + "video_file": "Charades/videos/TQGPM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts it on the dining room table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.5> to <27.4>." + } + ] + }, + { + "video_id": "AK9IB", + "question_id": "AK9IB", + "video_file": "Charades/videos/AK9IB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes their shoes off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.1> to <29.0>." + } + ] + }, + { + "video_id": "UQW95", + "question_id": "UQW95", + "video_file": "Charades/videos/UQW95.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is cooking on the stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <11.7>." + } + ] + }, + { + "video_id": "UQW95", + "question_id": "UQW95", + "video_file": "Charades/videos/UQW95.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person the open a box of food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.0> to <15.8>." + } + ] + }, + { + "video_id": "UQW95", + "question_id": "UQW95", + "video_file": "Charades/videos/UQW95.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.0> to <15.8>." + } + ] + }, + { + "video_id": "UQW95", + "question_id": "UQW95", + "video_file": "Charades/videos/UQW95.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pour it in the pot they are cooking with' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <11.7>." + } + ] + }, + { + "video_id": "UQW95", + "question_id": "UQW95", + "video_file": "Charades/videos/UQW95.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is cooking at the stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.4> to <11.7>." + } + ] + }, + { + "video_id": "BTW1H", + "question_id": "BTW1H", + "video_file": "Charades/videos/BTW1H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person watches a car through the window'." + }, + { + "from": "gpt", + "value": "From <15.0> to <29.0>." + } + ] + }, + { + "video_id": "BTW1H", + "question_id": "BTW1H", + "video_file": "Charades/videos/BTW1H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes a vacuum to the entryway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <10.7>." + } + ] + }, + { + "video_id": "OE751", + "question_id": "OE751", + "video_file": "Charades/videos/OE751.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person is talking on the phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <23.4>." + } + ] + }, + { + "video_id": "OE751", + "question_id": "OE751", + "video_file": "Charades/videos/OE751.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person talking on the phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.0> to <23.4>." + } + ] + }, + { + "video_id": "OE751", + "question_id": "OE751", + "video_file": "Charades/videos/OE751.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the door behind them' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.1> to <34.0>." + } + ] + }, + { + "video_id": "Q3HZI", + "question_id": "Q3HZI", + "video_file": "Charades/videos/Q3HZI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'carries through a door to the outside/ the person opens'." + }, + { + "from": "gpt", + "value": "From <8.6> to <17.7>." + } + ] + }, + { + "video_id": "Q3HZI", + "question_id": "Q3HZI", + "video_file": "Charades/videos/Q3HZI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person takes a pillow off a sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "4P13T", + "question_id": "4P13T", + "video_file": "Charades/videos/4P13T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person runs through'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "4P13T", + "question_id": "4P13T", + "video_file": "Charades/videos/4P13T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "4P13T", + "question_id": "4P13T", + "video_file": "Charades/videos/4P13T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person runs out the front door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "4P13T", + "question_id": "4P13T", + "video_file": "Charades/videos/4P13T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs into a living room'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "VD6P1", + "question_id": "VD6P1", + "video_file": "Charades/videos/VD6P1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person fixes his hair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <13.1>." + } + ] + }, + { + "video_id": "NYDRK", + "question_id": "NYDRK", + "video_file": "Charades/videos/NYDRK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throw a book down the stairs too' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.6> to <15.9>." + } + ] + }, + { + "video_id": "P2DV7", + "question_id": "P2DV7", + "video_file": "Charades/videos/P2DV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a picture of something on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <17.4>." + } + ] + }, + { + "video_id": "TAZGF", + "question_id": "TAZGF", + "video_file": "Charades/videos/TAZGF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put the plate on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <10.5>." + } + ] + }, + { + "video_id": "TAZGF", + "question_id": "TAZGF", + "video_file": "Charades/videos/TAZGF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sit down on a sofa'." + }, + { + "from": "gpt", + "value": "From <7.3> to <13.0>." + } + ] + }, + { + "video_id": "TAZGF", + "question_id": "TAZGF", + "video_file": "Charades/videos/TAZGF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they sit on the couch' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.3> to <13.0>." + } + ] + }, + { + "video_id": "1L376", + "question_id": "1L376", + "video_file": "Charades/videos/1L376.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is dressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.8> to <14.8>." + } + ] + }, + { + "video_id": "MJX48", + "question_id": "MJX48", + "video_file": "Charades/videos/MJX48.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person puts a pair of shoes into a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.2> to <15.3>." + } + ] + }, + { + "video_id": "MJX48", + "question_id": "MJX48", + "video_file": "Charades/videos/MJX48.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws the bag up the stairs'." + }, + { + "from": "gpt", + "value": "From <18.5> to <23.5>." + } + ] + }, + { + "video_id": "MJX48", + "question_id": "MJX48", + "video_file": "Charades/videos/MJX48.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person he throws the bag as well' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.5> to <23.5>." + } + ] + }, + { + "video_id": "BD0S7", + "question_id": "BD0S7", + "video_file": "Charades/videos/BD0S7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'second person takes a picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.4>." + } + ] + }, + { + "video_id": "K2BY7", + "question_id": "K2BY7", + "video_file": "Charades/videos/K2BY7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is putting on clothes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "0AGCS", + "question_id": "0AGCS", + "video_file": "Charades/videos/0AGCS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person throws the blanket on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <22.5>." + } + ] + }, + { + "video_id": "BRLC0", + "question_id": "BRLC0", + "video_file": "Charades/videos/BRLC0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person eating sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.5> to <19.2>." + } + ] + }, + { + "video_id": "U7RVS", + "question_id": "U7RVS", + "video_file": "Charades/videos/U7RVS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a drink from a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "U7RVS", + "question_id": "U7RVS", + "video_file": "Charades/videos/U7RVS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking a cup of coffee'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "U7RVS", + "question_id": "U7RVS", + "video_file": "Charades/videos/U7RVS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person places the cup on a laptop computer'." + }, + { + "from": "gpt", + "value": "From <7.2> to <13.9>." + } + ] + }, + { + "video_id": "QEQJ3", + "question_id": "QEQJ3", + "video_file": "Charades/videos/QEQJ3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person close the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.1> to <24.3>." + } + ] + }, + { + "video_id": "QEQJ3", + "question_id": "QEQJ3", + "video_file": "Charades/videos/QEQJ3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is playing with their phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <9.3>." + } + ] + }, + { + "video_id": "QEQJ3", + "question_id": "QEQJ3", + "video_file": "Charades/videos/QEQJ3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the doorknob' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.1> to <24.3>." + } + ] + }, + { + "video_id": "GUCB4", + "question_id": "GUCB4", + "video_file": "Charades/videos/GUCB4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person drinks the glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.9> to <39.3>." + } + ] + }, + { + "video_id": "GUCB4", + "question_id": "GUCB4", + "video_file": "Charades/videos/GUCB4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person pours water into a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.8> to <26.5>." + } + ] + }, + { + "video_id": "GUCB4", + "question_id": "GUCB4", + "video_file": "Charades/videos/GUCB4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they pour something into a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <26.5>." + } + ] + }, + { + "video_id": "GUCB4", + "question_id": "GUCB4", + "video_file": "Charades/videos/GUCB4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person pours water into a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <26.5>." + } + ] + }, + { + "video_id": "YW5C4", + "question_id": "YW5C4", + "video_file": "Charades/videos/YW5C4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing their cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <9.6>." + } + ] + }, + { + "video_id": "YW5C4", + "question_id": "YW5C4", + "video_file": "Charades/videos/YW5C4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they start to take some medicine with a spoon' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.0> to <24.0>." + } + ] + }, + { + "video_id": "YW5C4", + "question_id": "YW5C4", + "video_file": "Charades/videos/YW5C4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <25.9>." + } + ] + }, + { + "video_id": "YW5C4", + "question_id": "YW5C4", + "video_file": "Charades/videos/YW5C4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.1> to <11.9>." + } + ] + }, + { + "video_id": "2MJ72", + "question_id": "2MJ72", + "video_file": "Charades/videos/2MJ72.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close the door to their bedroom'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <17.0>." + } + ] + }, + { + "video_id": "2MJ72", + "question_id": "2MJ72", + "video_file": "Charades/videos/2MJ72.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they close the door'." + }, + { + "from": "gpt", + "value": "From <11.0> to <17.0>." + } + ] + }, + { + "video_id": "EN80R", + "question_id": "EN80R", + "video_file": "Charades/videos/EN80R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is cooking on a stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "EN80R", + "question_id": "EN80R", + "video_file": "Charades/videos/EN80R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is cooking at a stove'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "EN80R", + "question_id": "EN80R", + "video_file": "Charades/videos/EN80R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take some medicine off a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.6> to <20.4>." + } + ] + }, + { + "video_id": "PLGZM", + "question_id": "PLGZM", + "video_file": "Charades/videos/PLGZM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is watching television' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.2>." + } + ] + }, + { + "video_id": "PLGZM", + "question_id": "PLGZM", + "video_file": "Charades/videos/PLGZM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes a camera from the shelf'." + }, + { + "from": "gpt", + "value": "From <2.5> to <8.1>." + } + ] + }, + { + "video_id": "3MWAY", + "question_id": "3MWAY", + "video_file": "Charades/videos/3MWAY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.5>." + } + ] + }, + { + "video_id": "3MWAY", + "question_id": "3MWAY", + "video_file": "Charades/videos/3MWAY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pours it into the cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.1> to <32.1>." + } + ] + }, + { + "video_id": "3MWAY", + "question_id": "3MWAY", + "video_file": "Charades/videos/3MWAY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.8> to <14.3>." + } + ] + }, + { + "video_id": "30TL5", + "question_id": "30TL5", + "video_file": "Charades/videos/30TL5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting on a toilet reading a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.7> to <15.4>." + } + ] + }, + { + "video_id": "WSZSL", + "question_id": "WSZSL", + "video_file": "Charades/videos/WSZSL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <22.9>." + } + ] + }, + { + "video_id": "WSZSL", + "question_id": "WSZSL", + "video_file": "Charades/videos/WSZSL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opens the window'." + }, + { + "from": "gpt", + "value": "From <12.8> to <22.9>." + } + ] + }, + { + "video_id": "54XD1", + "question_id": "54XD1", + "video_file": "Charades/videos/54XD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turn off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.8> to <31.0>." + } + ] + }, + { + "video_id": "54XD1", + "question_id": "54XD1", + "video_file": "Charades/videos/54XD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs out of the room through the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.4> to <27.2>." + } + ] + }, + { + "video_id": "54XD1", + "question_id": "54XD1", + "video_file": "Charades/videos/54XD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person but leaves it there to run across the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.4> to <27.2>." + } + ] + }, + { + "video_id": "54XD1", + "question_id": "54XD1", + "video_file": "Charades/videos/54XD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.8> to <31.0>." + } + ] + }, + { + "video_id": "54XD1", + "question_id": "54XD1", + "video_file": "Charades/videos/54XD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turning off all the lights' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.8> to <31.0>." + } + ] + }, + { + "video_id": "KZK6W", + "question_id": "KZK6W", + "video_file": "Charades/videos/KZK6W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person stands at a stove cooking some food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.7> to <23.8>." + } + ] + }, + { + "video_id": "KZK6W", + "question_id": "KZK6W", + "video_file": "Charades/videos/KZK6W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinks from a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "KZK6W", + "question_id": "KZK6W", + "video_file": "Charades/videos/KZK6W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is cooking food on the stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.7> to <23.8>." + } + ] + }, + { + "video_id": "KZK6W", + "question_id": "KZK6W", + "video_file": "Charades/videos/KZK6W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take a glass of water from the counter' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.7> to <30.4>." + } + ] + }, + { + "video_id": "6FUDS", + "question_id": "6FUDS", + "video_file": "Charades/videos/6FUDS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens the blinds for a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <17.0>." + } + ] + }, + { + "video_id": "6FUDS", + "question_id": "6FUDS", + "video_file": "Charades/videos/6FUDS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person continue eating some food out of a bowl' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <33.7>." + } + ] + }, + { + "video_id": "6FUDS", + "question_id": "6FUDS", + "video_file": "Charades/videos/6FUDS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person sits down to eat some food'." + }, + { + "from": "gpt", + "value": "From <17.7> to <33.7>." + } + ] + }, + { + "video_id": "6FUDS", + "question_id": "6FUDS", + "video_file": "Charades/videos/6FUDS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is opening the blinds to their window'." + }, + { + "from": "gpt", + "value": "From <9.7> to <17.0>." + } + ] + }, + { + "video_id": "QNUIU", + "question_id": "QNUIU", + "video_file": "Charades/videos/QNUIU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws it on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.6> to <31.0>." + } + ] + }, + { + "video_id": "QNUIU", + "question_id": "QNUIU", + "video_file": "Charades/videos/QNUIU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throw the homework on the floor'." + }, + { + "from": "gpt", + "value": "From <16.6> to <31.0>." + } + ] + }, + { + "video_id": "5GPOJ", + "question_id": "5GPOJ", + "video_file": "Charades/videos/5GPOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts the book down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.6> to <14.7>." + } + ] + }, + { + "video_id": "5GPOJ", + "question_id": "5GPOJ", + "video_file": "Charades/videos/5GPOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they put the book down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.6> to <14.7>." + } + ] + }, + { + "video_id": "CDIW7", + "question_id": "CDIW7", + "video_file": "Charades/videos/CDIW7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts a box in front of a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <13.5>." + } + ] + }, + { + "video_id": "GLGQJ", + "question_id": "GLGQJ", + "video_file": "Charades/videos/GLGQJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is seen washing a window in their room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <17.6>." + } + ] + }, + { + "video_id": "GLGQJ", + "question_id": "GLGQJ", + "video_file": "Charades/videos/GLGQJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the window'." + }, + { + "from": "gpt", + "value": "From <21.1> to <29.3>." + } + ] + }, + { + "video_id": "GLGQJ", + "question_id": "GLGQJ", + "video_file": "Charades/videos/GLGQJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they start closing the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.1> to <29.3>." + } + ] + }, + { + "video_id": "GLGQJ", + "question_id": "GLGQJ", + "video_file": "Charades/videos/GLGQJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person washing a window with a cloth' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <17.6>." + } + ] + }, + { + "video_id": "7XVS7", + "question_id": "7XVS7", + "video_file": "Charades/videos/7XVS7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is pouring coffee in to a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.1> to <18.9>." + } + ] + }, + { + "video_id": "7XVS7", + "question_id": "7XVS7", + "video_file": "Charades/videos/7XVS7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pour the coffee into a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.1> to <18.9>." + } + ] + }, + { + "video_id": "7XVS7", + "question_id": "7XVS7", + "video_file": "Charades/videos/7XVS7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they pour themselves a cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.1> to <18.9>." + } + ] + }, + { + "video_id": "ESAIY", + "question_id": "ESAIY", + "video_file": "Charades/videos/ESAIY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they fix their hair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.3> to <29.5>." + } + ] + }, + { + "video_id": "ESAIY", + "question_id": "ESAIY", + "video_file": "Charades/videos/ESAIY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they start fixing their hair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.3> to <29.5>." + } + ] + }, + { + "video_id": "ESAIY", + "question_id": "ESAIY", + "video_file": "Charades/videos/ESAIY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take the medicine with a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.7> to <11.6>." + } + ] + }, + { + "video_id": "XFLQH", + "question_id": "XFLQH", + "video_file": "Charades/videos/XFLQH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a box next to the bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.0> to <26.0>." + } + ] + }, + { + "video_id": "XFLQH", + "question_id": "XFLQH", + "video_file": "Charades/videos/XFLQH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts them on the shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.0> to <36.0>." + } + ] + }, + { + "video_id": "XFLQH", + "question_id": "XFLQH", + "video_file": "Charades/videos/XFLQH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take some items out of the box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.0> to <36.0>." + } + ] + }, + { + "video_id": "XFLQH", + "question_id": "XFLQH", + "video_file": "Charades/videos/XFLQH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the first person puts something on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.0> to <36.0>." + } + ] + }, + { + "video_id": "JXF3M", + "question_id": "JXF3M", + "video_file": "Charades/videos/JXF3M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they awaken'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <20.5>." + } + ] + }, + { + "video_id": "JUINA", + "question_id": "JUINA", + "video_file": "Charades/videos/JUINA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person is standing on chair putting dishes on shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "JUINA", + "question_id": "JUINA", + "video_file": "Charades/videos/JUINA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts dishes on a shelf'." + }, + { + "from": "gpt", + "value": "From <3.2> to <15.0>." + } + ] + }, + { + "video_id": "TKFY2", + "question_id": "TKFY2", + "video_file": "Charades/videos/TKFY2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person eats a bite of sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.7> to <17.2>." + } + ] + }, + { + "video_id": "E6TED", + "question_id": "E6TED", + "video_file": "Charades/videos/E6TED.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person drinks a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.0> to <31.0>." + } + ] + }, + { + "video_id": "7FNZZ", + "question_id": "7FNZZ", + "video_file": "Charades/videos/7FNZZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts it on the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.4> to <29.5>." + } + ] + }, + { + "video_id": "T3FJH", + "question_id": "T3FJH", + "video_file": "Charades/videos/T3FJH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person drinks from a wine glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "T3FJH", + "question_id": "T3FJH", + "video_file": "Charades/videos/T3FJH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person is drinking out of a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "ED6VQ", + "question_id": "ED6VQ", + "video_file": "Charades/videos/ED6VQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put the cup on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.8> to <13.0>." + } + ] + }, + { + "video_id": "87T5P", + "question_id": "87T5P", + "video_file": "Charades/videos/87T5P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sit on a couch' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.6> to <32.0>." + } + ] + }, + { + "video_id": "87T5P", + "question_id": "87T5P", + "video_file": "Charades/videos/87T5P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person watching television' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.7> to <32.0>." + } + ] + }, + { + "video_id": "2544C", + "question_id": "2544C", + "video_file": "Charades/videos/2544C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting items onto the shelf'." + }, + { + "from": "gpt", + "value": "From <11.0> to <45.0>." + } + ] + }, + { + "video_id": "2544C", + "question_id": "2544C", + "video_file": "Charades/videos/2544C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks through a doorway folding a towel' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <75.9> to <82.0>." + } + ] + }, + { + "video_id": "2544C", + "question_id": "2544C", + "video_file": "Charades/videos/2544C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts the towel on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.0> to <45.0>." + } + ] + }, + { + "video_id": "2544C", + "question_id": "2544C", + "video_file": "Charades/videos/2544C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person throws a towel in their home office'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <15.4>." + } + ] + }, + { + "video_id": "EE8OB", + "question_id": "EE8OB", + "video_file": "Charades/videos/EE8OB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats the sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.8> to <21.0>." + } + ] + }, + { + "video_id": "EE8OB", + "question_id": "EE8OB", + "video_file": "Charades/videos/EE8OB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a woman walking in a house with a sandwich eating'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.8> to <21.0>." + } + ] + }, + { + "video_id": "TLBWW", + "question_id": "TLBWW", + "video_file": "Charades/videos/TLBWW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is walking through a doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <15.3>." + } + ] + }, + { + "video_id": "TLBWW", + "question_id": "TLBWW", + "video_file": "Charades/videos/TLBWW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws a towel' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.2> to <27.0>." + } + ] + }, + { + "video_id": "9HIE0", + "question_id": "9HIE0", + "video_file": "Charades/videos/9HIE0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is cooking food on the stove'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.5>." + } + ] + }, + { + "video_id": "9HIE0", + "question_id": "9HIE0", + "video_file": "Charades/videos/9HIE0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person is smiling as they enter the room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <25.7>." + } + ] + }, + { + "video_id": "9HIE0", + "question_id": "9HIE0", + "video_file": "Charades/videos/9HIE0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is cooking something on a stove'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.5>." + } + ] + }, + { + "video_id": "04LAX", + "question_id": "04LAX", + "video_file": "Charades/videos/04LAX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pours it into a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <22.7>." + } + ] + }, + { + "video_id": "04LAX", + "question_id": "04LAX", + "video_file": "Charades/videos/04LAX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinks from the glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.2> to <27.3>." + } + ] + }, + { + "video_id": "UDDTR", + "question_id": "UDDTR", + "video_file": "Charades/videos/UDDTR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person beginning to dress themselves'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <37.5> to <51.1>." + } + ] + }, + { + "video_id": "UDDTR", + "question_id": "UDDTR", + "video_file": "Charades/videos/UDDTR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts putting on clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <13.2>." + } + ] + }, + { + "video_id": "XOFEX", + "question_id": "XOFEX", + "video_file": "Charades/videos/XOFEX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put it down on the table'." + }, + { + "from": "gpt", + "value": "From <12.0> to <17.2>." + } + ] + }, + { + "video_id": "KZODG", + "question_id": "KZODG", + "video_file": "Charades/videos/KZODG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they close the laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.3> to <22.9>." + } + ] + }, + { + "video_id": "KZODG", + "question_id": "KZODG", + "video_file": "Charades/videos/KZODG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin playing a game on the phone's touchscreen' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.8> to <32.0>." + } + ] + }, + { + "video_id": "KZODG", + "question_id": "KZODG", + "video_file": "Charades/videos/KZODG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person closes a laptop on the coffee table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.3> to <22.9>." + } + ] + }, + { + "video_id": "I20N2", + "question_id": "I20N2", + "video_file": "Charades/videos/I20N2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is running'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "I20N2", + "question_id": "I20N2", + "video_file": "Charades/videos/I20N2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws it on a clothesline' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.6> to <33.1>." + } + ] + }, + { + "video_id": "I20N2", + "question_id": "I20N2", + "video_file": "Charades/videos/I20N2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is running toward the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "F9K8L", + "question_id": "F9K8L", + "video_file": "Charades/videos/F9K8L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs into the kitchen'." + }, + { + "from": "gpt", + "value": "From <0.3> to <5.8>." + } + ] + }, + { + "video_id": "F9K8L", + "question_id": "F9K8L", + "video_file": "Charades/videos/F9K8L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts a pair of shoes on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <7.2>." + } + ] + }, + { + "video_id": "F9K8L", + "question_id": "F9K8L", + "video_file": "Charades/videos/F9K8L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person he opens a window screen' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.7> to <14.1>." + } + ] + }, + { + "video_id": "F9K8L", + "question_id": "F9K8L", + "video_file": "Charades/videos/F9K8L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs into his kitchen'." + }, + { + "from": "gpt", + "value": "From <0.3> to <5.8>." + } + ] + }, + { + "video_id": "KO80I", + "question_id": "KO80I", + "video_file": "Charades/videos/KO80I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting a pillow behind the back' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <23.1>." + } + ] + }, + { + "video_id": "1CITG", + "question_id": "1CITG", + "video_file": "Charades/videos/1CITG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns on the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.6> to <10.6>." + } + ] + }, + { + "video_id": "54VBR", + "question_id": "54VBR", + "video_file": "Charades/videos/54VBR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits in a chair at the table'." + }, + { + "from": "gpt", + "value": "From <16.3> to <23.8>." + } + ] + }, + { + "video_id": "54VBR", + "question_id": "54VBR", + "video_file": "Charades/videos/54VBR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting at a table eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <23.8>." + } + ] + }, + { + "video_id": "54VBR", + "question_id": "54VBR", + "video_file": "Charades/videos/54VBR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits down at the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <23.8>." + } + ] + }, + { + "video_id": "54VBR", + "question_id": "54VBR", + "video_file": "Charades/videos/54VBR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting at the table eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <16.3> to <23.8>." + } + ] + }, + { + "video_id": "54VBR", + "question_id": "54VBR", + "video_file": "Charades/videos/54VBR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a drink from a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <18.7>." + } + ] + }, + { + "video_id": "3OUXT", + "question_id": "3OUXT", + "video_file": "Charades/videos/3OUXT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.3> to <21.1>." + } + ] + }, + { + "video_id": "TETZ7", + "question_id": "TETZ7", + "video_file": "Charades/videos/TETZ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person laughs as they leave through the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <12.6>." + } + ] + }, + { + "video_id": "L5842", + "question_id": "L5842", + "video_file": "Charades/videos/L5842.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating food placed on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <26.3>." + } + ] + }, + { + "video_id": "L5842", + "question_id": "L5842", + "video_file": "Charades/videos/L5842.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person puts a pillow on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.1> to <17.8>." + } + ] + }, + { + "video_id": "L5842", + "question_id": "L5842", + "video_file": "Charades/videos/L5842.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats some food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.2> to <26.3>." + } + ] + }, + { + "video_id": "WNKL6", + "question_id": "WNKL6", + "video_file": "Charades/videos/WNKL6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating from a bag of food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.8> to <28.0>." + } + ] + }, + { + "video_id": "WNKL6", + "question_id": "WNKL6", + "video_file": "Charades/videos/WNKL6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is seen eating food in a bag'." + }, + { + "from": "gpt", + "value": "From <20.8> to <28.0>." + } + ] + }, + { + "video_id": "WNKL6", + "question_id": "WNKL6", + "video_file": "Charades/videos/WNKL6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person laughs' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.7>." + } + ] + }, + { + "video_id": "P8ZU6", + "question_id": "P8ZU6", + "video_file": "Charades/videos/P8ZU6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.3> to <18.4>." + } + ] + }, + { + "video_id": "DPBI3", + "question_id": "DPBI3", + "video_file": "Charades/videos/DPBI3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.0> to <18.7>." + } + ] + }, + { + "video_id": "Y1FKP", + "question_id": "Y1FKP", + "video_file": "Charades/videos/Y1FKP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person look out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.7> to <32.0>." + } + ] + }, + { + "video_id": "Y1FKP", + "question_id": "Y1FKP", + "video_file": "Charades/videos/Y1FKP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person looks out the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.7> to <32.0>." + } + ] + }, + { + "video_id": "Y1FKP", + "question_id": "Y1FKP", + "video_file": "Charades/videos/Y1FKP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person laugh softly'." + }, + { + "from": "gpt", + "value": "From <21.9> to <32.0>." + } + ] + }, + { + "video_id": "Y1FKP", + "question_id": "Y1FKP", + "video_file": "Charades/videos/Y1FKP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person laughs as they look out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.7> to <32.0>." + } + ] + }, + { + "video_id": "NDYOR", + "question_id": "NDYOR", + "video_file": "Charades/videos/NDYOR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes their phone from the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <23.9>." + } + ] + }, + { + "video_id": "CX5ZM", + "question_id": "CX5ZM", + "video_file": "Charades/videos/CX5ZM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person start cooking something on the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <24.6>." + } + ] + }, + { + "video_id": "0LNLR", + "question_id": "0LNLR", + "video_file": "Charades/videos/0LNLR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits on a sofa' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.9> to <13.9>." + } + ] + }, + { + "video_id": "OHQOO", + "question_id": "OHQOO", + "video_file": "Charades/videos/OHQOO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person looking out the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.7> to <10.2>." + } + ] + }, + { + "video_id": "OHQOO", + "question_id": "OHQOO", + "video_file": "Charades/videos/OHQOO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is holding a box of shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.8> to <12.2>." + } + ] + }, + { + "video_id": "OHQOO", + "question_id": "OHQOO", + "video_file": "Charades/videos/OHQOO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person start laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.1> to <22.1>." + } + ] + }, + { + "video_id": "Z60TB", + "question_id": "Z60TB", + "video_file": "Charades/videos/Z60TB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sneezing into the phone in the dining room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.3> to <10.8>." + } + ] + }, + { + "video_id": "8VWV2", + "question_id": "8VWV2", + "video_file": "Charades/videos/8VWV2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they begin putting their dishes away in the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <52.5> to <58.3>." + } + ] + }, + { + "video_id": "8VWV2", + "question_id": "8VWV2", + "video_file": "Charades/videos/8VWV2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is seen eating food at a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <23.6>." + } + ] + }, + { + "video_id": "8VWV2", + "question_id": "8VWV2", + "video_file": "Charades/videos/8VWV2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person another eats on the floor'." + }, + { + "from": "gpt", + "value": "From <0.0> to <23.6>." + } + ] + }, + { + "video_id": "UPJWV", + "question_id": "UPJWV", + "video_file": "Charades/videos/UPJWV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is cooking on a stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.7> to <19.2>." + } + ] + }, + { + "video_id": "UPJWV", + "question_id": "UPJWV", + "video_file": "Charades/videos/UPJWV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking food on the stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <19.2>." + } + ] + }, + { + "video_id": "UPJWV", + "question_id": "UPJWV", + "video_file": "Charades/videos/UPJWV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs out' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.0> to <29.4>." + } + ] + }, + { + "video_id": "UPJWV", + "question_id": "UPJWV", + "video_file": "Charades/videos/UPJWV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person run out of the kitchen' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.0> to <29.4>." + } + ] + }, + { + "video_id": "UPJWV", + "question_id": "UPJWV", + "video_file": "Charades/videos/UPJWV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they turn off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <24.9>." + } + ] + }, + { + "video_id": "ZF5CP", + "question_id": "ZF5CP", + "video_file": "Charades/videos/ZF5CP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person closes the closet door'." + }, + { + "from": "gpt", + "value": "From <14.2> to <21.1>." + } + ] + }, + { + "video_id": "ZF5CP", + "question_id": "ZF5CP", + "video_file": "Charades/videos/ZF5CP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "P7RQS", + "question_id": "P7RQS", + "video_file": "Charades/videos/P7RQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting at the table'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "P7RQS", + "question_id": "P7RQS", + "video_file": "Charades/videos/P7RQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person is at the refrigerator laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.8> to <7.1>." + } + ] + }, + { + "video_id": "VS7DA", + "question_id": "VS7DA", + "video_file": "Charades/videos/VS7DA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.0> to <10.8>." + } + ] + }, + { + "video_id": "VS7DA", + "question_id": "VS7DA", + "video_file": "Charades/videos/VS7DA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns the light on' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.7> to <18.1>." + } + ] + }, + { + "video_id": "VS7DA", + "question_id": "VS7DA", + "video_file": "Charades/videos/VS7DA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.7> to <18.1>." + } + ] + }, + { + "video_id": "VS7DA", + "question_id": "VS7DA", + "video_file": "Charades/videos/VS7DA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns light on grabs something from self' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.7> to <18.1>." + } + ] + }, + { + "video_id": "VS7DA", + "question_id": "VS7DA", + "video_file": "Charades/videos/VS7DA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person close the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.6> to <38.2>." + } + ] + }, + { + "video_id": "VS7DA", + "question_id": "VS7DA", + "video_file": "Charades/videos/VS7DA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the door to the pantry' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <10.8>." + } + ] + }, + { + "video_id": "72HTA", + "question_id": "72HTA", + "video_file": "Charades/videos/72HTA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take clothes off a table'." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.0>." + } + ] + }, + { + "video_id": "72HTA", + "question_id": "72HTA", + "video_file": "Charades/videos/72HTA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes off clothes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.0>." + } + ] + }, + { + "video_id": "72HTA", + "question_id": "72HTA", + "video_file": "Charades/videos/72HTA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person in a basement begins undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.0> to <12.5>." + } + ] + }, + { + "video_id": "72HTA", + "question_id": "72HTA", + "video_file": "Charades/videos/72HTA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts clothes in washing machine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <11.6>." + } + ] + }, + { + "video_id": "5QJNP", + "question_id": "5QJNP", + "video_file": "Charades/videos/5QJNP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person began to drink from a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.5> to <32.8>." + } + ] + }, + { + "video_id": "1K0SU", + "question_id": "1K0SU", + "video_file": "Charades/videos/1K0SU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking a glass of water'." + }, + { + "from": "gpt", + "value": "From <14.9> to <23.0>." + } + ] + }, + { + "video_id": "1K0SU", + "question_id": "1K0SU", + "video_file": "Charades/videos/1K0SU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <9.4>." + } + ] + }, + { + "video_id": "3QGLU", + "question_id": "3QGLU", + "video_file": "Charades/videos/3QGLU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.4> to <35.0>." + } + ] + }, + { + "video_id": "3QGLU", + "question_id": "3QGLU", + "video_file": "Charades/videos/3QGLU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the refrigerator door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.9> to <9.7>." + } + ] + }, + { + "video_id": "3QGLU", + "question_id": "3QGLU", + "video_file": "Charades/videos/3QGLU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts his phone back in his pocket'." + }, + { + "from": "gpt", + "value": "From <28.9> to <33.5>." + } + ] + }, + { + "video_id": "7WG7A", + "question_id": "7WG7A", + "video_file": "Charades/videos/7WG7A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.0> to <26.0>." + } + ] + }, + { + "video_id": "7WG7A", + "question_id": "7WG7A", + "video_file": "Charades/videos/7WG7A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turning off a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.0> to <26.0>." + } + ] + }, + { + "video_id": "YKVYR", + "question_id": "YKVYR", + "video_file": "Charades/videos/YKVYR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes the garbage page' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <15.1>." + } + ] + }, + { + "video_id": "YKVYR", + "question_id": "YKVYR", + "video_file": "Charades/videos/YKVYR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes the bag outside'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <15.1>." + } + ] + }, + { + "video_id": "WEU2E", + "question_id": "WEU2E", + "video_file": "Charades/videos/WEU2E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they start laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.4> to <31.0>." + } + ] + }, + { + "video_id": "WEU2E", + "question_id": "WEU2E", + "video_file": "Charades/videos/WEU2E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is smiling'." + }, + { + "from": "gpt", + "value": "From <14.3> to <23.4>." + } + ] + }, + { + "video_id": "WEU2E", + "question_id": "WEU2E", + "video_file": "Charades/videos/WEU2E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the box on top of a dresser' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.7> to <17.8>." + } + ] + }, + { + "video_id": "WEU2E", + "question_id": "WEU2E", + "video_file": "Charades/videos/WEU2E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins to laugh' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.4> to <31.0>." + } + ] + }, + { + "video_id": "7SXQS", + "question_id": "7SXQS", + "video_file": "Charades/videos/7SXQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes blinds at a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.0> to <7.0>." + } + ] + }, + { + "video_id": "7SXQS", + "question_id": "7SXQS", + "video_file": "Charades/videos/7SXQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is closing the window in the dining room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <7.0>." + } + ] + }, + { + "video_id": "7SXQS", + "question_id": "7SXQS", + "video_file": "Charades/videos/7SXQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the shutters of the window behind the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <7.0>." + } + ] + }, + { + "video_id": "7SXQS", + "question_id": "7SXQS", + "video_file": "Charades/videos/7SXQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person seen snuggling up with a blanket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <32.0>." + } + ] + }, + { + "video_id": "PZQ5A", + "question_id": "PZQ5A", + "video_file": "Charades/videos/PZQ5A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closed the door behind them' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.6> to <16.8>." + } + ] + }, + { + "video_id": "PZQ5A", + "question_id": "PZQ5A", + "video_file": "Charades/videos/PZQ5A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes their camera that was at a desk'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.6> to <10.3>." + } + ] + }, + { + "video_id": "77HDM", + "question_id": "77HDM", + "video_file": "Charades/videos/77HDM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turning off the light behind them' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <16.3>." + } + ] + }, + { + "video_id": "64F1C", + "question_id": "64F1C", + "video_file": "Charades/videos/64F1C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is throwing a blanket on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.6> to <27.7>." + } + ] + }, + { + "video_id": "64F1C", + "question_id": "64F1C", + "video_file": "Charades/videos/64F1C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throwing towels on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.3> to <27.4>." + } + ] + }, + { + "video_id": "V2JOF", + "question_id": "V2JOF", + "video_file": "Charades/videos/V2JOF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person putting the blanket on a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <26.5>." + } + ] + }, + { + "video_id": "V2JOF", + "question_id": "V2JOF", + "video_file": "Charades/videos/V2JOF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns on a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.6> to <8.8>." + } + ] + }, + { + "video_id": "MZGMN", + "question_id": "MZGMN", + "video_file": "Charades/videos/MZGMN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a box on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "MZGMN", + "question_id": "MZGMN", + "video_file": "Charades/videos/MZGMN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws a book into it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <13.7>." + } + ] + }, + { + "video_id": "MZGMN", + "question_id": "MZGMN", + "video_file": "Charades/videos/MZGMN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.6> to <14.0>." + } + ] + }, + { + "video_id": "MZGMN", + "question_id": "MZGMN", + "video_file": "Charades/videos/MZGMN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens up a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "MZGMN", + "question_id": "MZGMN", + "video_file": "Charades/videos/MZGMN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the door behind them' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.9> to <27.0>." + } + ] + }, + { + "video_id": "MZGMN", + "question_id": "MZGMN", + "video_file": "Charades/videos/MZGMN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.9> to <27.0>." + } + ] + }, + { + "video_id": "MZGMN", + "question_id": "MZGMN", + "video_file": "Charades/videos/MZGMN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws a book into the box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <13.7>." + } + ] + }, + { + "video_id": "O7OD2", + "question_id": "O7OD2", + "video_file": "Charades/videos/O7OD2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is holding a medicine bottle'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <10.9>." + } + ] + }, + { + "video_id": "O7OD2", + "question_id": "O7OD2", + "video_file": "Charades/videos/O7OD2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is holding a bottle of medicine'." + }, + { + "from": "gpt", + "value": "From <4.4> to <10.9>." + } + ] + }, + { + "video_id": "O7OD2", + "question_id": "O7OD2", + "video_file": "Charades/videos/O7OD2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <16.1>." + } + ] + }, + { + "video_id": "4878H", + "question_id": "4878H", + "video_file": "Charades/videos/4878H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person snuggling a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "4878H", + "question_id": "4878H", + "video_file": "Charades/videos/4878H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person snuggles a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "8NS1G", + "question_id": "8NS1G", + "video_file": "Charades/videos/8NS1G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eat a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.7> to <31.0>." + } + ] + }, + { + "video_id": "8NS1G", + "question_id": "8NS1G", + "video_file": "Charades/videos/8NS1G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they pour a glass of milk'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.3> to <25.5>." + } + ] + }, + { + "video_id": "WF3NY", + "question_id": "WF3NY", + "video_file": "Charades/videos/WF3NY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opening the cabinet'." + }, + { + "from": "gpt", + "value": "From <16.0> to <23.0>." + } + ] + }, + { + "video_id": "WF3NY", + "question_id": "WF3NY", + "video_file": "Charades/videos/WF3NY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.0> to <23.0>." + } + ] + }, + { + "video_id": "NC1OC", + "question_id": "NC1OC", + "video_file": "Charades/videos/NC1OC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits down on the couch'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <27.9>." + } + ] + }, + { + "video_id": "NC1OC", + "question_id": "NC1OC", + "video_file": "Charades/videos/NC1OC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sit down on the sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.6> to <27.9>." + } + ] + }, + { + "video_id": "NC1OC", + "question_id": "NC1OC", + "video_file": "Charades/videos/NC1OC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins to work on it at the coffee table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.2> to <49.0>." + } + ] + }, + { + "video_id": "NC1OC", + "question_id": "NC1OC", + "video_file": "Charades/videos/NC1OC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they put their homework on a side table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.7> to <25.5>." + } + ] + }, + { + "video_id": "S1N2U", + "question_id": "S1N2U", + "video_file": "Charades/videos/S1N2U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person lying in a bed awakens' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.3> to <13.3>." + } + ] + }, + { + "video_id": "S1N2U", + "question_id": "S1N2U", + "video_file": "Charades/videos/S1N2U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person grasps for a light on a nightstand'." + }, + { + "from": "gpt", + "value": "From <12.4> to <16.7>." + } + ] + }, + { + "video_id": "S1N2U", + "question_id": "S1N2U", + "video_file": "Charades/videos/S1N2U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts on their shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.8> to <33.3>." + } + ] + }, + { + "video_id": "S1N2U", + "question_id": "S1N2U", + "video_file": "Charades/videos/S1N2U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <16.7>." + } + ] + }, + { + "video_id": "G3CDA", + "question_id": "G3CDA", + "video_file": "Charades/videos/G3CDA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the folded towel down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.7> to <23.1>." + } + ] + }, + { + "video_id": "00ZCA", + "question_id": "00ZCA", + "video_file": "Charades/videos/00ZCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take out a vacuum'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.9> to <42.4>." + } + ] + }, + { + "video_id": "00ZCA", + "question_id": "00ZCA", + "video_file": "Charades/videos/00ZCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes out vacuum' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.9> to <42.4>." + } + ] + }, + { + "video_id": "00ZCA", + "question_id": "00ZCA", + "video_file": "Charades/videos/00ZCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person tidy up the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <35.6> to <54.0>." + } + ] + }, + { + "video_id": "00ZCA", + "question_id": "00ZCA", + "video_file": "Charades/videos/00ZCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.5> to <16.4>." + } + ] + }, + { + "video_id": "R6QCQ", + "question_id": "R6QCQ", + "video_file": "Charades/videos/R6QCQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens on a living room sofa'." + }, + { + "from": "gpt", + "value": "From <2.1> to <9.0>." + } + ] + }, + { + "video_id": "R6QCQ", + "question_id": "R6QCQ", + "video_file": "Charades/videos/R6QCQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person watches themselves in the mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <18.8>." + } + ] + }, + { + "video_id": "YV521", + "question_id": "YV521", + "video_file": "Charades/videos/YV521.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting on a shelf behind him' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.6> to <28.7>." + } + ] + }, + { + "video_id": "YV521", + "question_id": "YV521", + "video_file": "Charades/videos/YV521.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put the book on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.6> to <28.7>." + } + ] + }, + { + "video_id": "YV521", + "question_id": "YV521", + "video_file": "Charades/videos/YV521.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'this person is sitting on a bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <17.3>." + } + ] + }, + { + "video_id": "YV521", + "question_id": "YV521", + "video_file": "Charades/videos/YV521.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the book away' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.6> to <28.0>." + } + ] + }, + { + "video_id": "YV521", + "question_id": "YV521", + "video_file": "Charades/videos/YV521.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is opening a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.8>." + } + ] + }, + { + "video_id": "YV521", + "question_id": "YV521", + "video_file": "Charades/videos/YV521.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person start sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.6> to <30.2>." + } + ] + }, + { + "video_id": "67EEN", + "question_id": "67EEN", + "video_file": "Charades/videos/67EEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they stand up' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.8> to <18.7>." + } + ] + }, + { + "video_id": "ANJNU", + "question_id": "ANJNU", + "video_file": "Charades/videos/ANJNU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person runs in' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <9.5>." + } + ] + }, + { + "video_id": "ANJNU", + "question_id": "ANJNU", + "video_file": "Charades/videos/ANJNU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns the light off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <27.3>." + } + ] + }, + { + "video_id": "ANJNU", + "question_id": "ANJNU", + "video_file": "Charades/videos/ANJNU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person runs into the room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <9.5>." + } + ] + }, + { + "video_id": "ANJNU", + "question_id": "ANJNU", + "video_file": "Charades/videos/ANJNU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <27.3>." + } + ] + }, + { + "video_id": "J12SC", + "question_id": "J12SC", + "video_file": "Charades/videos/J12SC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a camera out of it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.4> to <21.4>." + } + ] + }, + { + "video_id": "J12SC", + "question_id": "J12SC", + "video_file": "Charades/videos/J12SC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes a box off the staircase' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.8> to <15.2>." + } + ] + }, + { + "video_id": "J12SC", + "question_id": "J12SC", + "video_file": "Charades/videos/J12SC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes their shoes off' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.1>." + } + ] + }, + { + "video_id": "J12SC", + "question_id": "J12SC", + "video_file": "Charades/videos/J12SC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take a camera out' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <21.4>." + } + ] + }, + { + "video_id": "J12SC", + "question_id": "J12SC", + "video_file": "Charades/videos/J12SC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.0> to <18.2>." + } + ] + }, + { + "video_id": "U1BQ0", + "question_id": "U1BQ0", + "video_file": "Charades/videos/U1BQ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts throwing them in boxes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.9>." + } + ] + }, + { + "video_id": "U1BQ0", + "question_id": "U1BQ0", + "video_file": "Charades/videos/U1BQ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is seen putting dishes away'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.3>." + } + ] + }, + { + "video_id": "AADCE", + "question_id": "AADCE", + "video_file": "Charades/videos/AADCE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person is drinking water from a water bottle' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "AADCE", + "question_id": "AADCE", + "video_file": "Charades/videos/AADCE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person playing on their phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.5> to <32.0>." + } + ] + }, + { + "video_id": "AADCE", + "question_id": "AADCE", + "video_file": "Charades/videos/AADCE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking from a glass of water'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "YV0ZX", + "question_id": "YV0ZX", + "video_file": "Charades/videos/YV0ZX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <19.1>." + } + ] + }, + { + "video_id": "LE5F4", + "question_id": "LE5F4", + "video_file": "Charades/videos/LE5F4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes a camera from the chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "BUL4V", + "question_id": "BUL4V", + "video_file": "Charades/videos/BUL4V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person closes the window blinds' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <26.1>." + } + ] + }, + { + "video_id": "BUL4V", + "question_id": "BUL4V", + "video_file": "Charades/videos/BUL4V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they close a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.2> to <26.1>." + } + ] + }, + { + "video_id": "BUL4V", + "question_id": "BUL4V", + "video_file": "Charades/videos/BUL4V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <26.1>." + } + ] + }, + { + "video_id": "PP2FF", + "question_id": "PP2FF", + "video_file": "Charades/videos/PP2FF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person the stand up' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.5> to <21.3>." + } + ] + }, + { + "video_id": "PP2FF", + "question_id": "PP2FF", + "video_file": "Charades/videos/PP2FF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looked out the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.4> to <32.0>." + } + ] + }, + { + "video_id": "PP2FF", + "question_id": "PP2FF", + "video_file": "Charades/videos/PP2FF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person is sitting on the bed eating'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "PP2FF", + "question_id": "PP2FF", + "video_file": "Charades/videos/PP2FF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person walk to look out the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.4> to <32.0>." + } + ] + }, + { + "video_id": "PP2FF", + "question_id": "PP2FF", + "video_file": "Charades/videos/PP2FF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person sat on their bed eating some food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "PP2FF", + "question_id": "PP2FF", + "video_file": "Charades/videos/PP2FF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person was eating sitting on the sofa' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "PP2FF", + "question_id": "PP2FF", + "video_file": "Charades/videos/PP2FF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person was watching out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.4> to <32.0>." + } + ] + }, + { + "video_id": "W5GEK", + "question_id": "W5GEK", + "video_file": "Charades/videos/W5GEK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person turned on a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.0> to <32.0>." + } + ] + }, + { + "video_id": "JMCRT", + "question_id": "JMCRT", + "video_file": "Charades/videos/JMCRT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is cooking food on the stove they' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "JMCRT", + "question_id": "JMCRT", + "video_file": "Charades/videos/JMCRT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open the cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "JMCRT", + "question_id": "JMCRT", + "video_file": "Charades/videos/JMCRT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they stop cooking'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "JMCRT", + "question_id": "JMCRT", + "video_file": "Charades/videos/JMCRT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they talk on their phone'." + }, + { + "from": "gpt", + "value": "From <19.3> to <29.3>." + } + ] + }, + { + "video_id": "0ET8W", + "question_id": "0ET8W", + "video_file": "Charades/videos/0ET8W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs into the living room through the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <13.8>." + } + ] + }, + { + "video_id": "FV51S", + "question_id": "FV51S", + "video_file": "Charades/videos/FV51S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes out a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <10.4>." + } + ] + }, + { + "video_id": "Q8UJ8", + "question_id": "Q8UJ8", + "video_file": "Charades/videos/Q8UJ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a bag which sits on the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <5.9>." + } + ] + }, + { + "video_id": "9GQRI", + "question_id": "9GQRI", + "video_file": "Charades/videos/9GQRI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person cooking on the stove begins sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <18.6>." + } + ] + }, + { + "video_id": "R390Z", + "question_id": "R390Z", + "video_file": "Charades/videos/R390Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person takes a picture with their phone camera' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.7> to <31.7>." + } + ] + }, + { + "video_id": "X6B0P", + "question_id": "X6B0P", + "video_file": "Charades/videos/X6B0P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is standing in the hallway undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.3> to <22.7>." + } + ] + }, + { + "video_id": "KZPBC", + "question_id": "KZPBC", + "video_file": "Charades/videos/KZPBC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a book to read' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.2> to <18.6>." + } + ] + }, + { + "video_id": "U3ANG", + "question_id": "U3ANG", + "video_file": "Charades/videos/U3ANG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the laptop in a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.3> to <36.0>." + } + ] + }, + { + "video_id": "U3ANG", + "question_id": "U3ANG", + "video_file": "Charades/videos/U3ANG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening the closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.4> to <11.4>." + } + ] + }, + { + "video_id": "U3ANG", + "question_id": "U3ANG", + "video_file": "Charades/videos/U3ANG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opened a closet door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <11.4>." + } + ] + }, + { + "video_id": "KA3ER", + "question_id": "KA3ER", + "video_file": "Charades/videos/KA3ER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a cabinet'." + }, + { + "from": "gpt", + "value": "From <14.9> to <22.1>." + } + ] + }, + { + "video_id": "KA3ER", + "question_id": "KA3ER", + "video_file": "Charades/videos/KA3ER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is putting groceries on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.6>." + } + ] + }, + { + "video_id": "R2KI4", + "question_id": "R2KI4", + "video_file": "Charades/videos/R2KI4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person watches something on their laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.4> to <33.6>." + } + ] + }, + { + "video_id": "R2KI4", + "question_id": "R2KI4", + "video_file": "Charades/videos/R2KI4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put down their camera'." + }, + { + "from": "gpt", + "value": "From <20.4> to <25.6>." + } + ] + }, + { + "video_id": "R2KI4", + "question_id": "R2KI4", + "video_file": "Charades/videos/R2KI4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking from a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.7> to <34.0>." + } + ] + }, + { + "video_id": "TVBLW", + "question_id": "TVBLW", + "video_file": "Charades/videos/TVBLW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a bag of chips'." + }, + { + "from": "gpt", + "value": "From <11.0> to <19.0>." + } + ] + }, + { + "video_id": "TVBLW", + "question_id": "TVBLW", + "video_file": "Charades/videos/TVBLW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person eats from a bottle at the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.0> to <19.0>." + } + ] + }, + { + "video_id": "TVBLW", + "question_id": "TVBLW", + "video_file": "Charades/videos/TVBLW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.0> to <19.0>." + } + ] + }, + { + "video_id": "JXKPA", + "question_id": "JXKPA", + "video_file": "Charades/videos/JXKPA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pours the contents of the bottle into the glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.1> to <23.9>." + } + ] + }, + { + "video_id": "JXKPA", + "question_id": "JXKPA", + "video_file": "Charades/videos/JXKPA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person poured it into a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.1> to <23.9>." + } + ] + }, + { + "video_id": "S2FLE", + "question_id": "S2FLE", + "video_file": "Charades/videos/S2FLE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put it on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <17.2>." + } + ] + }, + { + "video_id": "S2FLE", + "question_id": "S2FLE", + "video_file": "Charades/videos/S2FLE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is putting homework on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.9> to <17.2>." + } + ] + }, + { + "video_id": "S2FLE", + "question_id": "S2FLE", + "video_file": "Charades/videos/S2FLE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.2> to <25.5>." + } + ] + }, + { + "video_id": "NALYS", + "question_id": "NALYS", + "video_file": "Charades/videos/NALYS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts it in the box'." + }, + { + "from": "gpt", + "value": "From <17.6> to <28.1>." + } + ] + }, + { + "video_id": "NALYS", + "question_id": "NALYS", + "video_file": "Charades/videos/NALYS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens box of clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.1>." + } + ] + }, + { + "video_id": "NALYS", + "question_id": "NALYS", + "video_file": "Charades/videos/NALYS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts smiling' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.3> to <21.0>." + } + ] + }, + { + "video_id": "NALYS", + "question_id": "NALYS", + "video_file": "Charades/videos/NALYS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens box'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.1>." + } + ] + }, + { + "video_id": "NALYS", + "question_id": "NALYS", + "video_file": "Charades/videos/NALYS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person carries box out of closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.4> to <17.9>." + } + ] + }, + { + "video_id": "0BX9N", + "question_id": "0BX9N", + "video_file": "Charades/videos/0BX9N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.8> to <21.4>." + } + ] + }, + { + "video_id": "0BX9N", + "question_id": "0BX9N", + "video_file": "Charades/videos/0BX9N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes out books' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.8> to <24.0>." + } + ] + }, + { + "video_id": "0BX9N", + "question_id": "0BX9N", + "video_file": "Charades/videos/0BX9N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they open the bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.8> to <21.4>." + } + ] + }, + { + "video_id": "0BX9N", + "question_id": "0BX9N", + "video_file": "Charades/videos/0BX9N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes the homework back out of the bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.8>." + } + ] + }, + { + "video_id": "0BX9N", + "question_id": "0BX9N", + "video_file": "Charades/videos/0BX9N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts down a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <10.2>." + } + ] + }, + { + "video_id": "JIOD0", + "question_id": "JIOD0", + "video_file": "Charades/videos/JIOD0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person tidying up the blanket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <19.4>." + } + ] + }, + { + "video_id": "GZMJK", + "question_id": "GZMJK", + "video_file": "Charades/videos/GZMJK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sneezing with a towel over their heads' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.7> to <15.1>." + } + ] + }, + { + "video_id": "GZMJK", + "question_id": "GZMJK", + "video_file": "Charades/videos/GZMJK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person eats a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.1> to <33.0>." + } + ] + }, + { + "video_id": "CW4AU", + "question_id": "CW4AU", + "video_file": "Charades/videos/CW4AU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person run around the room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.8> to <35.2>." + } + ] + }, + { + "video_id": "CW4AU", + "question_id": "CW4AU", + "video_file": "Charades/videos/CW4AU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs around the room once' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.8> to <35.2>." + } + ] + }, + { + "video_id": "CW4AU", + "question_id": "CW4AU", + "video_file": "Charades/videos/CW4AU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they put their shoes on' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.5> to <27.9>." + } + ] + }, + { + "video_id": "CW4AU", + "question_id": "CW4AU", + "video_file": "Charades/videos/CW4AU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts on shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.5> to <27.9>." + } + ] + }, + { + "video_id": "KC863", + "question_id": "KC863", + "video_file": "Charades/videos/KC863.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts it on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.9> to <17.8>." + } + ] + }, + { + "video_id": "TPHRR", + "question_id": "TPHRR", + "video_file": "Charades/videos/TPHRR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <8.0>." + } + ] + }, + { + "video_id": "TPHRR", + "question_id": "TPHRR", + "video_file": "Charades/videos/TPHRR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting in a chair'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.0>." + } + ] + }, + { + "video_id": "TPHRR", + "question_id": "TPHRR", + "video_file": "Charades/videos/TPHRR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sitting on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.0>." + } + ] + }, + { + "video_id": "TW3RS", + "question_id": "TW3RS", + "video_file": "Charades/videos/TW3RS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person smiling in a mirror' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.6> to <10.2>." + } + ] + }, + { + "video_id": "19ZJB", + "question_id": "19ZJB", + "video_file": "Charades/videos/19ZJB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person is putting books by the television' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.9> to <27.9>." + } + ] + }, + { + "video_id": "19ZJB", + "question_id": "19ZJB", + "video_file": "Charades/videos/19ZJB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'this person opens a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.1> to <11.9>." + } + ] + }, + { + "video_id": "E1VFZ", + "question_id": "E1VFZ", + "video_file": "Charades/videos/E1VFZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person puts a box on a desk' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.5> to <6.0>." + } + ] + }, + { + "video_id": "VCU4P", + "question_id": "VCU4P", + "video_file": "Charades/videos/VCU4P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a woman is smiling at a young guy' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "VCU4P", + "question_id": "VCU4P", + "video_file": "Charades/videos/VCU4P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is smiling' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "OMFOJ", + "question_id": "OMFOJ", + "video_file": "Charades/videos/OMFOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <21.1>." + } + ] + }, + { + "video_id": "OMFOJ", + "question_id": "OMFOJ", + "video_file": "Charades/videos/OMFOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts their cleaning supplies down on a nearby shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.7> to <25.2>." + } + ] + }, + { + "video_id": "8JT6S", + "question_id": "8JT6S", + "video_file": "Charades/videos/8JT6S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks to a sofa holding a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "0YOXV", + "question_id": "0YOXV", + "video_file": "Charades/videos/0YOXV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking bottle of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "0YOXV", + "question_id": "0YOXV", + "video_file": "Charades/videos/0YOXV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is drinking water from a bottle' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "0YOXV", + "question_id": "0YOXV", + "video_file": "Charades/videos/0YOXV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person is smiling at picture on phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <10.0>." + } + ] + }, + { + "video_id": "HSH5R", + "question_id": "HSH5R", + "video_file": "Charades/videos/HSH5R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.7> to <15.8>." + } + ] + }, + { + "video_id": "HSH5R", + "question_id": "HSH5R", + "video_file": "Charades/videos/HSH5R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person someone is sitting at a table writing on something' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.0>." + } + ] + }, + { + "video_id": "LZEWE", + "question_id": "LZEWE", + "video_file": "Charades/videos/LZEWE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person fixed their hair'." + }, + { + "from": "gpt", + "value": "From <23.0> to <28.0>." + } + ] + }, + { + "video_id": "LZEWE", + "question_id": "LZEWE", + "video_file": "Charades/videos/LZEWE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting clothes into a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "2JR26", + "question_id": "2JR26", + "video_file": "Charades/videos/2JR26.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they start drinking some coffee from a glass'." + }, + { + "from": "gpt", + "value": "From <29.4> to <38.0>." + } + ] + }, + { + "video_id": "1TW98", + "question_id": "1TW98", + "video_file": "Charades/videos/1TW98.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closing a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <21.5>." + } + ] + }, + { + "video_id": "X8A6H", + "question_id": "X8A6H", + "video_file": "Charades/videos/X8A6H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts the homework on the shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <24.0>." + } + ] + }, + { + "video_id": "X8A6H", + "question_id": "X8A6H", + "video_file": "Charades/videos/X8A6H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakens on the floor of the closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.2> to <7.8>." + } + ] + }, + { + "video_id": "8Y7P5", + "question_id": "8Y7P5", + "video_file": "Charades/videos/8Y7P5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person runs out of the garage' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "8Y7P5", + "question_id": "8Y7P5", + "video_file": "Charades/videos/8Y7P5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person another is running out a doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "WPZQ7", + "question_id": "WPZQ7", + "video_file": "Charades/videos/WPZQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off his shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <20.0>." + } + ] + }, + { + "video_id": "WPZQ7", + "question_id": "WPZQ7", + "video_file": "Charades/videos/WPZQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.0> to <20.0>." + } + ] + }, + { + "video_id": "XF93D", + "question_id": "XF93D", + "video_file": "Charades/videos/XF93D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they close the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.0> to <11.0>." + } + ] + }, + { + "video_id": "XF93D", + "question_id": "XF93D", + "video_file": "Charades/videos/XF93D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts away a broom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <8.0>." + } + ] + }, + { + "video_id": "XF93D", + "question_id": "XF93D", + "video_file": "Charades/videos/XF93D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts broom in the closet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <8.0>." + } + ] + }, + { + "video_id": "XF93D", + "question_id": "XF93D", + "video_file": "Charades/videos/XF93D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a door'." + }, + { + "from": "gpt", + "value": "From <22.3> to <25.0>." + } + ] + }, + { + "video_id": "Y9JWG", + "question_id": "Y9JWG", + "video_file": "Charades/videos/Y9JWG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person washing dishes at the sink'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.3>." + } + ] + }, + { + "video_id": "Y9JWG", + "question_id": "Y9JWG", + "video_file": "Charades/videos/Y9JWG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looking out the window'." + }, + { + "from": "gpt", + "value": "From <1.5> to <8.0>." + } + ] + }, + { + "video_id": "GK15U", + "question_id": "GK15U", + "video_file": "Charades/videos/GK15U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sits on the sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.7> to <6.9>." + } + ] + }, + { + "video_id": "GK15U", + "question_id": "GK15U", + "video_file": "Charades/videos/GK15U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person playing on their phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.9> to <33.0>." + } + ] + }, + { + "video_id": "GK15U", + "question_id": "GK15U", + "video_file": "Charades/videos/GK15U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating a bag of chips' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.9> to <33.0>." + } + ] + }, + { + "video_id": "LFU36", + "question_id": "LFU36", + "video_file": "Charades/videos/LFU36.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person holding a laptop puts it on the stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "LFU36", + "question_id": "LFU36", + "video_file": "Charades/videos/LFU36.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person holding a laptop opens the refrigerator door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "LFU36", + "question_id": "LFU36", + "video_file": "Charades/videos/LFU36.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.7> to <9.1>." + } + ] + }, + { + "video_id": "VTEWM", + "question_id": "VTEWM", + "video_file": "Charades/videos/VTEWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.3> to <29.4>." + } + ] + }, + { + "video_id": "VTEWM", + "question_id": "VTEWM", + "video_file": "Charades/videos/VTEWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits on the edge of the bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.0> to <6.8>." + } + ] + }, + { + "video_id": "P3EW1", + "question_id": "P3EW1", + "video_file": "Charades/videos/P3EW1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes off their shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.1> to <14.7>." + } + ] + }, + { + "video_id": "SSP8C", + "question_id": "SSP8C", + "video_file": "Charades/videos/SSP8C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person awakens in bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.7> to <17.4>." + } + ] + }, + { + "video_id": "J1YJ6", + "question_id": "J1YJ6", + "video_file": "Charades/videos/J1YJ6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a window'." + }, + { + "from": "gpt", + "value": "From <3.6> to <13.3>." + } + ] + }, + { + "video_id": "J1YJ6", + "question_id": "J1YJ6", + "video_file": "Charades/videos/J1YJ6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a window sits on a sofa'." + }, + { + "from": "gpt", + "value": "From <3.6> to <13.3>." + } + ] + }, + { + "video_id": "J1YJ6", + "question_id": "J1YJ6", + "video_file": "Charades/videos/J1YJ6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is looking out a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <15.8>." + } + ] + }, + { + "video_id": "TD6CM", + "question_id": "TD6CM", + "video_file": "Charades/videos/TD6CM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person started dressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <14.0>." + } + ] + }, + { + "video_id": "ZCULX", + "question_id": "ZCULX", + "video_file": "Charades/videos/ZCULX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person starts smiling tries to read a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.6> to <26.0>." + } + ] + }, + { + "video_id": "ZCULX", + "question_id": "ZCULX", + "video_file": "Charades/videos/ZCULX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person reading a book at a desk'." + }, + { + "from": "gpt", + "value": "From <15.6> to <26.0>." + } + ] + }, + { + "video_id": "ZCULX", + "question_id": "ZCULX", + "video_file": "Charades/videos/ZCULX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <5.8> to <14.0>." + } + ] + }, + { + "video_id": "9X208", + "question_id": "9X208", + "video_file": "Charades/videos/9X208.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens in bed'." + }, + { + "from": "gpt", + "value": "From <14.0> to <25.6>." + } + ] + }, + { + "video_id": "9X208", + "question_id": "9X208", + "video_file": "Charades/videos/9X208.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person leaves by opening a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.0> to <30.4>." + } + ] + }, + { + "video_id": "9X208", + "question_id": "9X208", + "video_file": "Charades/videos/9X208.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a towel off of a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <26.8>." + } + ] + }, + { + "video_id": "9X208", + "question_id": "9X208", + "video_file": "Charades/videos/9X208.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walking through the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.2> to <30.9>." + } + ] + }, + { + "video_id": "9X208", + "question_id": "9X208", + "video_file": "Charades/videos/9X208.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a towel' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <26.8>." + } + ] + }, + { + "video_id": "7VSIV", + "question_id": "7VSIV", + "video_file": "Charades/videos/7VSIV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is standing on a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.3>." + } + ] + }, + { + "video_id": "7VSIV", + "question_id": "7VSIV", + "video_file": "Charades/videos/7VSIV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person stands on a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.3>." + } + ] + }, + { + "video_id": "EJJIO", + "question_id": "EJJIO", + "video_file": "Charades/videos/EJJIO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.8> to <10.5>." + } + ] + }, + { + "video_id": "EJJIO", + "question_id": "EJJIO", + "video_file": "Charades/videos/EJJIO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.8> to <10.5>." + } + ] + }, + { + "video_id": "EJJIO", + "question_id": "EJJIO", + "video_file": "Charades/videos/EJJIO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <23.2>." + } + ] + }, + { + "video_id": "NOLQ7", + "question_id": "NOLQ7", + "video_file": "Charades/videos/NOLQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put the dishes down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.8> to <26.8>." + } + ] + }, + { + "video_id": "NOLQ7", + "question_id": "NOLQ7", + "video_file": "Charades/videos/NOLQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts away a dish' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.8> to <26.8>." + } + ] + }, + { + "video_id": "YMM1Z", + "question_id": "YMM1Z", + "video_file": "Charades/videos/YMM1Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person runs to the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.4> to <22.2>." + } + ] + }, + { + "video_id": "ZKOCS", + "question_id": "ZKOCS", + "video_file": "Charades/videos/ZKOCS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person begins eating from the plate' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.8> to <20.6>." + } + ] + }, + { + "video_id": "ZKOCS", + "question_id": "ZKOCS", + "video_file": "Charades/videos/ZKOCS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins to eat' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.8> to <20.6>." + } + ] + }, + { + "video_id": "ZKOCS", + "question_id": "ZKOCS", + "video_file": "Charades/videos/ZKOCS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sitting down at the table'." + }, + { + "from": "gpt", + "value": "From <9.0> to <17.5>." + } + ] + }, + { + "video_id": "VY1GQ", + "question_id": "VY1GQ", + "video_file": "Charades/videos/VY1GQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person starts laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.8> to <29.8>." + } + ] + }, + { + "video_id": "VY1GQ", + "question_id": "VY1GQ", + "video_file": "Charades/videos/VY1GQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person looking out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <29.9>." + } + ] + }, + { + "video_id": "VY1GQ", + "question_id": "VY1GQ", + "video_file": "Charades/videos/VY1GQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person smiling at their reflection' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.7> to <20.8>." + } + ] + }, + { + "video_id": "VY1GQ", + "question_id": "VY1GQ", + "video_file": "Charades/videos/VY1GQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person wash hands at sink'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "6U082", + "question_id": "6U082", + "video_file": "Charades/videos/6U082.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person run off'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.4> to <25.0>." + } + ] + }, + { + "video_id": "6U082", + "question_id": "6U082", + "video_file": "Charades/videos/6U082.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person run out of the room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.4> to <25.0>." + } + ] + }, + { + "video_id": "PIMO4", + "question_id": "PIMO4", + "video_file": "Charades/videos/PIMO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is running into the room holding a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.8> to <10.3>." + } + ] + }, + { + "video_id": "PIMO4", + "question_id": "PIMO4", + "video_file": "Charades/videos/PIMO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begins playing on their phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.1> to <31.0>." + } + ] + }, + { + "video_id": "Q9H04", + "question_id": "Q9H04", + "video_file": "Charades/videos/Q9H04.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is in the rec room undressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <6.9>." + } + ] + }, + { + "video_id": "Q9H04", + "question_id": "Q9H04", + "video_file": "Charades/videos/Q9H04.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throwing their clothes about' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <10.2>." + } + ] + }, + { + "video_id": "Q9H04", + "question_id": "Q9H04", + "video_file": "Charades/videos/Q9H04.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person b is eating from a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.9> to <14.7>." + } + ] + }, + { + "video_id": "Q9H04", + "question_id": "Q9H04", + "video_file": "Charades/videos/Q9H04.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'is approached by another person eating a snack' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.9> to <14.7>." + } + ] + }, + { + "video_id": "Q9H04", + "question_id": "Q9H04", + "video_file": "Charades/videos/Q9H04.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person walks in eating a cracker' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.9> to <14.7>." + } + ] + }, + { + "video_id": "Q9H04", + "question_id": "Q9H04", + "video_file": "Charades/videos/Q9H04.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <6.9>." + } + ] + }, + { + "video_id": "WFUK2", + "question_id": "WFUK2", + "video_file": "Charades/videos/WFUK2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they are smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <22.0>." + } + ] + }, + { + "video_id": "WR2Q2", + "question_id": "WR2Q2", + "video_file": "Charades/videos/WR2Q2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is eating a sandwich from a dish'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.6>." + } + ] + }, + { + "video_id": "GC1Q1", + "question_id": "GC1Q1", + "video_file": "Charades/videos/GC1Q1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opening a door is carrying a towel' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "GC1Q1", + "question_id": "GC1Q1", + "video_file": "Charades/videos/GC1Q1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the towel'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <11.6>." + } + ] + }, + { + "video_id": "O45BC", + "question_id": "O45BC", + "video_file": "Charades/videos/O45BC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws some clothes on the floor'." + }, + { + "from": "gpt", + "value": "From <11.1> to <16.1>." + } + ] + }, + { + "video_id": "O45BC", + "question_id": "O45BC", + "video_file": "Charades/videos/O45BC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person puts down a laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.6> to <11.0>." + } + ] + }, + { + "video_id": "O45BC", + "question_id": "O45BC", + "video_file": "Charades/videos/O45BC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws clothing on floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <16.1>." + } + ] + }, + { + "video_id": "FEWUR", + "question_id": "FEWUR", + "video_file": "Charades/videos/FEWUR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throw their shoes'." + }, + { + "from": "gpt", + "value": "From <21.2> to <31.4>." + } + ] + }, + { + "video_id": "FEWUR", + "question_id": "FEWUR", + "video_file": "Charades/videos/FEWUR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.4> to <31.2>." + } + ] + }, + { + "video_id": "1U9TF", + "question_id": "1U9TF", + "video_file": "Charades/videos/1U9TF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person begins sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.0> to <19.4>." + } + ] + }, + { + "video_id": "1U9TF", + "question_id": "1U9TF", + "video_file": "Charades/videos/1U9TF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <19.4>." + } + ] + }, + { + "video_id": "IZZ3Y", + "question_id": "IZZ3Y", + "video_file": "Charades/videos/IZZ3Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opens up a laptop computer'." + }, + { + "from": "gpt", + "value": "From <10.4> to <18.1>." + } + ] + }, + { + "video_id": "IZZ3Y", + "question_id": "IZZ3Y", + "video_file": "Charades/videos/IZZ3Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take it with a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.3> to <11.6>." + } + ] + }, + { + "video_id": "IZZ3Y", + "question_id": "IZZ3Y", + "video_file": "Charades/videos/IZZ3Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening a laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <18.1>." + } + ] + }, + { + "video_id": "IZZ3Y", + "question_id": "IZZ3Y", + "video_file": "Charades/videos/IZZ3Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes some more medicine with another drink of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <12.3>." + } + ] + }, + { + "video_id": "IZZ3Y", + "question_id": "IZZ3Y", + "video_file": "Charades/videos/IZZ3Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes some medicine with a drink of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <12.3>." + } + ] + }, + { + "video_id": "U5B14", + "question_id": "U5B14", + "video_file": "Charades/videos/U5B14.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person someone sitting on a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.0>." + } + ] + }, + { + "video_id": "U5B14", + "question_id": "U5B14", + "video_file": "Charades/videos/U5B14.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting in a chair in the bathroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.0>." + } + ] + }, + { + "video_id": "WZ351", + "question_id": "WZ351", + "video_file": "Charades/videos/WZ351.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is smiling' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.4> to <16.2>." + } + ] + }, + { + "video_id": "U03TA", + "question_id": "U03TA", + "video_file": "Charades/videos/U03TA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the pillow down on a bench' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.1> to <15.2>." + } + ] + }, + { + "video_id": "U03TA", + "question_id": "U03TA", + "video_file": "Charades/videos/U03TA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a man holds a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.4>." + } + ] + }, + { + "video_id": "JQHIC", + "question_id": "JQHIC", + "video_file": "Charades/videos/JQHIC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is opening a closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <14.8>." + } + ] + }, + { + "video_id": "JQHIC", + "question_id": "JQHIC", + "video_file": "Charades/videos/JQHIC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person opens the door to the closet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <13.4>." + } + ] + }, + { + "video_id": "ZDX0Y", + "question_id": "ZDX0Y", + "video_file": "Charades/videos/ZDX0Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person is looking at a book'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "ZDX0Y", + "question_id": "ZDX0Y", + "video_file": "Charades/videos/ZDX0Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts book in a box'." + }, + { + "from": "gpt", + "value": "From <5.9> to <17.3>." + } + ] + }, + { + "video_id": "XYGU1", + "question_id": "XYGU1", + "video_file": "Charades/videos/XYGU1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is standing a dining room holding a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.9>." + } + ] + }, + { + "video_id": "XYGU1", + "question_id": "XYGU1", + "video_file": "Charades/videos/XYGU1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is holding a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.9>." + } + ] + }, + { + "video_id": "XYGU1", + "question_id": "XYGU1", + "video_file": "Charades/videos/XYGU1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.8> to <20.5>." + } + ] + }, + { + "video_id": "XYGU1", + "question_id": "XYGU1", + "video_file": "Charades/videos/XYGU1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a drink from a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <20.9>." + } + ] + }, + { + "video_id": "Q9TIN", + "question_id": "Q9TIN", + "video_file": "Charades/videos/Q9TIN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is closing a door by holding the doorknob' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.8> to <24.8>." + } + ] + }, + { + "video_id": "IZCB0", + "question_id": "IZCB0", + "video_file": "Charades/videos/IZCB0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.3>." + } + ] + }, + { + "video_id": "IZCB0", + "question_id": "IZCB0", + "video_file": "Charades/videos/IZCB0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closing closet door'." + }, + { + "from": "gpt", + "value": "From <2.2> to <12.2>." + } + ] + }, + { + "video_id": "YL612", + "question_id": "YL612", + "video_file": "Charades/videos/YL612.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws them on to the floor'." + }, + { + "from": "gpt", + "value": "From <12.4> to <18.3>." + } + ] + }, + { + "video_id": "YL612", + "question_id": "YL612", + "video_file": "Charades/videos/YL612.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throwing a pile of clothes onto the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <18.3>." + } + ] + }, + { + "video_id": "DU7H1", + "question_id": "DU7H1", + "video_file": "Charades/videos/DU7H1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins tidying the dining room table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.4> to <32.1>." + } + ] + }, + { + "video_id": "DU7H1", + "question_id": "DU7H1", + "video_file": "Charades/videos/DU7H1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person talking on the phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <13.2>." + } + ] + }, + { + "video_id": "DU7H1", + "question_id": "DU7H1", + "video_file": "Charades/videos/DU7H1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person is smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <39.7> to <48.0>." + } + ] + }, + { + "video_id": "39LQ8", + "question_id": "39LQ8", + "video_file": "Charades/videos/39LQ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person runs away'." + }, + { + "from": "gpt", + "value": "From <13.3> to <18.9>." + } + ] + }, + { + "video_id": "39LQ8", + "question_id": "39LQ8", + "video_file": "Charades/videos/39LQ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a cardboard box'." + }, + { + "from": "gpt", + "value": "From <7.4> to <16.4>." + } + ] + }, + { + "video_id": "39LQ8", + "question_id": "39LQ8", + "video_file": "Charades/videos/39LQ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs out of the room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.3> to <18.9>." + } + ] + }, + { + "video_id": "GU43L", + "question_id": "GU43L", + "video_file": "Charades/videos/GU43L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes their shoes off'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "SLMQF", + "question_id": "SLMQF", + "video_file": "Charades/videos/SLMQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the clothes down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.7> to <9.4>." + } + ] + }, + { + "video_id": "SLMQF", + "question_id": "SLMQF", + "video_file": "Charades/videos/SLMQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person takes off some clothes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.9> to <19.9>." + } + ] + }, + { + "video_id": "SLMQF", + "question_id": "SLMQF", + "video_file": "Charades/videos/SLMQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a book from a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.1> to <27.1>." + } + ] + }, + { + "video_id": "SLMQF", + "question_id": "SLMQF", + "video_file": "Charades/videos/SLMQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person gets redressed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.7> to <20.7>." + } + ] + }, + { + "video_id": "SLMQF", + "question_id": "SLMQF", + "video_file": "Charades/videos/SLMQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.7> to <20.7>." + } + ] + }, + { + "video_id": "SLMQF", + "question_id": "SLMQF", + "video_file": "Charades/videos/SLMQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person than puts clothes back on'." + }, + { + "from": "gpt", + "value": "From <4.7> to <9.4>." + } + ] + }, + { + "video_id": "U3NKR", + "question_id": "U3NKR", + "video_file": "Charades/videos/U3NKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the towel on the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <11.4>." + } + ] + }, + { + "video_id": "TLSUX", + "question_id": "TLSUX", + "video_file": "Charades/videos/TLSUX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person runs into the laundry room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "TLSUX", + "question_id": "TLSUX", + "video_file": "Charades/videos/TLSUX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person is running though doorway'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "GVX7E", + "question_id": "GVX7E", + "video_file": "Charades/videos/GVX7E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks through a doorway into a kitchen' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "WV9FZ", + "question_id": "WV9FZ", + "video_file": "Charades/videos/WV9FZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating potato chips' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.4> to <11.2>." + } + ] + }, + { + "video_id": "WV9FZ", + "question_id": "WV9FZ", + "video_file": "Charades/videos/WV9FZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats the food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.4> to <11.2>." + } + ] + }, + { + "video_id": "WV9FZ", + "question_id": "WV9FZ", + "video_file": "Charades/videos/WV9FZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes food from a package' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.0> to <8.2>." + } + ] + }, + { + "video_id": "6KDQO", + "question_id": "6KDQO", + "video_file": "Charades/videos/6KDQO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person stand up' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.3> to <27.5>." + } + ] + }, + { + "video_id": "6KDQO", + "question_id": "6KDQO", + "video_file": "Charades/videos/6KDQO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the glass on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.6> to <28.6>." + } + ] + }, + { + "video_id": "S2EXT", + "question_id": "S2EXT", + "video_file": "Charades/videos/S2EXT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <12.1>." + } + ] + }, + { + "video_id": "RRDAM", + "question_id": "RRDAM", + "video_file": "Charades/videos/RRDAM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person run from the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.7> to <24.0>." + } + ] + }, + { + "video_id": "RRDAM", + "question_id": "RRDAM", + "video_file": "Charades/videos/RRDAM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <15.7> to <23.0>." + } + ] + }, + { + "video_id": "RRDAM", + "question_id": "RRDAM", + "video_file": "Charades/videos/RRDAM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs out of the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.7> to <24.0>." + } + ] + }, + { + "video_id": "RRDAM", + "question_id": "RRDAM", + "video_file": "Charades/videos/RRDAM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks into the laundry room from the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <8.5>." + } + ] + }, + { + "video_id": "RRDAM", + "question_id": "RRDAM", + "video_file": "Charades/videos/RRDAM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <12.7>." + } + ] + }, + { + "video_id": "41CZS", + "question_id": "41CZS", + "video_file": "Charades/videos/41CZS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes off their jacket they pick a blanket up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.5> to <19.6>." + } + ] + }, + { + "video_id": "41CZS", + "question_id": "41CZS", + "video_file": "Charades/videos/41CZS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person next the open a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.3> to <23.4>." + } + ] + }, + { + "video_id": "41CZS", + "question_id": "41CZS", + "video_file": "Charades/videos/41CZS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person undresses by taking off a jacket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <15.7>." + } + ] + }, + { + "video_id": "V34GF", + "question_id": "V34GF", + "video_file": "Charades/videos/V34GF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put the glass on a counter'." + }, + { + "from": "gpt", + "value": "From <12.9> to <20.4>." + } + ] + }, + { + "video_id": "V34GF", + "question_id": "V34GF", + "video_file": "Charades/videos/V34GF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts the empty glass down on a desk' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <20.4>." + } + ] + }, + { + "video_id": "01KML", + "question_id": "01KML", + "video_file": "Charades/videos/01KML.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person begins watching their laptop intently'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.0> to <35.0>." + } + ] + }, + { + "video_id": "FERH5", + "question_id": "FERH5", + "video_file": "Charades/videos/FERH5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person is opening medicine cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.3> to <9.4>." + } + ] + }, + { + "video_id": "FERH5", + "question_id": "FERH5", + "video_file": "Charades/videos/FERH5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is opening their medicine cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.3> to <9.4>." + } + ] + }, + { + "video_id": "FERH5", + "question_id": "FERH5", + "video_file": "Charades/videos/FERH5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.3> to <22.2>." + } + ] + }, + { + "video_id": "FERH5", + "question_id": "FERH5", + "video_file": "Charades/videos/FERH5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they sneeze as they close the medicine cabinet'." + }, + { + "from": "gpt", + "value": "From <26.0> to <31.0>." + } + ] + }, + { + "video_id": "NOJK1", + "question_id": "NOJK1", + "video_file": "Charades/videos/NOJK1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take a phone from a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.7> to <29.0>." + } + ] + }, + { + "video_id": "NOJK1", + "question_id": "NOJK1", + "video_file": "Charades/videos/NOJK1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person play with the phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.3> to <33.0>." + } + ] + }, + { + "video_id": "NOJK1", + "question_id": "NOJK1", + "video_file": "Charades/videos/NOJK1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakens in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.5> to <16.4>." + } + ] + }, + { + "video_id": "NOJK1", + "question_id": "NOJK1", + "video_file": "Charades/videos/NOJK1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes their phone from the bed'." + }, + { + "from": "gpt", + "value": "From <23.7> to <29.0>." + } + ] + }, + { + "video_id": "V3RAX", + "question_id": "V3RAX", + "video_file": "Charades/videos/V3RAX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is undressing under a blanket'." + }, + { + "from": "gpt", + "value": "From <3.5> to <10.4>." + } + ] + }, + { + "video_id": "V3RAX", + "question_id": "V3RAX", + "video_file": "Charades/videos/V3RAX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person tidying up dirty clothes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.8> to <24.4>." + } + ] + }, + { + "video_id": "V3RAX", + "question_id": "V3RAX", + "video_file": "Charades/videos/V3RAX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person straightens clothes than takes a bite of a sandwich'." + }, + { + "from": "gpt", + "value": "From <9.6> to <21.2>." + } + ] + }, + { + "video_id": "V3RAX", + "question_id": "V3RAX", + "video_file": "Charades/videos/V3RAX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person finally eating a sandwich in a basement'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.8> to <29.0>." + } + ] + }, + { + "video_id": "QNKDI", + "question_id": "QNKDI", + "video_file": "Charades/videos/QNKDI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person reading a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.3> to <33.0>." + } + ] + }, + { + "video_id": "QNKDI", + "question_id": "QNKDI", + "video_file": "Charades/videos/QNKDI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person holding the book decides to stand up'." + }, + { + "from": "gpt", + "value": "From <24.9> to <33.0>." + } + ] + }, + { + "video_id": "GIC6A", + "question_id": "GIC6A", + "video_file": "Charades/videos/GIC6A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "GIC6A", + "question_id": "GIC6A", + "video_file": "Charades/videos/GIC6A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person enters through a doorway holding a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "GIC6A", + "question_id": "GIC6A", + "video_file": "Charades/videos/GIC6A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person walks back to the doorway'." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.7>." + } + ] + }, + { + "video_id": "NCDE3", + "question_id": "NCDE3", + "video_file": "Charades/videos/NCDE3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person is eating by the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.7>." + } + ] + }, + { + "video_id": "NCDE3", + "question_id": "NCDE3", + "video_file": "Charades/videos/NCDE3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks into a kitchen eating some food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.7>." + } + ] + }, + { + "video_id": "NCDE3", + "question_id": "NCDE3", + "video_file": "Charades/videos/NCDE3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating something'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.7>." + } + ] + }, + { + "video_id": "UB3KO", + "question_id": "UB3KO", + "video_file": "Charades/videos/UB3KO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person cooking a piece of meat'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <11.2>." + } + ] + }, + { + "video_id": "UB3KO", + "question_id": "UB3KO", + "video_file": "Charades/videos/UB3KO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens their laptop'." + }, + { + "from": "gpt", + "value": "From <20.8> to <26.3>." + } + ] + }, + { + "video_id": "UB3KO", + "question_id": "UB3KO", + "video_file": "Charades/videos/UB3KO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person cooks something' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.2> to <11.2>." + } + ] + }, + { + "video_id": "FIV2P", + "question_id": "FIV2P", + "video_file": "Charades/videos/FIV2P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is putting small items into a bag'." + }, + { + "from": "gpt", + "value": "From <14.5> to <19.4>." + } + ] + }, + { + "video_id": "AGELP", + "question_id": "AGELP", + "video_file": "Charades/videos/AGELP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person opens a window'." + }, + { + "from": "gpt", + "value": "From <5.9> to <13.1>." + } + ] + }, + { + "video_id": "AGELP", + "question_id": "AGELP", + "video_file": "Charades/videos/AGELP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person takes a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <27.4>." + } + ] + }, + { + "video_id": "AGELP", + "question_id": "AGELP", + "video_file": "Charades/videos/AGELP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins to open the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.9> to <13.1>." + } + ] + }, + { + "video_id": "AGELP", + "question_id": "AGELP", + "video_file": "Charades/videos/AGELP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <15.4>." + } + ] + }, + { + "video_id": "X29NF", + "question_id": "X29NF", + "video_file": "Charades/videos/X29NF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <15.9>." + } + ] + }, + { + "video_id": "X29NF", + "question_id": "X29NF", + "video_file": "Charades/videos/X29NF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person watches themselves undress in front of a mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.1> to <31.3>." + } + ] + }, + { + "video_id": "XHGV7", + "question_id": "XHGV7", + "video_file": "Charades/videos/XHGV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <10.0>." + } + ] + }, + { + "video_id": "X89FW", + "question_id": "X89FW", + "video_file": "Charades/videos/X89FW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person looks out a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <18.2>." + } + ] + }, + { + "video_id": "X89FW", + "question_id": "X89FW", + "video_file": "Charades/videos/X89FW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is looking out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <18.2>." + } + ] + }, + { + "video_id": "X89FW", + "question_id": "X89FW", + "video_file": "Charades/videos/X89FW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is laughing at something outside the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.3> to <6.3>." + } + ] + }, + { + "video_id": "X89FW", + "question_id": "X89FW", + "video_file": "Charades/videos/X89FW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person laughing they move from the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <6.3>." + } + ] + }, + { + "video_id": "A4SR3", + "question_id": "A4SR3", + "video_file": "Charades/videos/A4SR3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts the food aside' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.1> to <23.8>." + } + ] + }, + { + "video_id": "N7ZBM", + "question_id": "N7ZBM", + "video_file": "Charades/videos/N7ZBM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person can throw a broom in the living room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <13.6>." + } + ] + }, + { + "video_id": "N7ZBM", + "question_id": "N7ZBM", + "video_file": "Charades/videos/N7ZBM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throw the broom on the floor'." + }, + { + "from": "gpt", + "value": "From <8.8> to <13.6>." + } + ] + }, + { + "video_id": "56XKK", + "question_id": "56XKK", + "video_file": "Charades/videos/56XKK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the bag of groceries down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.3> to <12.3>." + } + ] + }, + { + "video_id": "56XKK", + "question_id": "56XKK", + "video_file": "Charades/videos/56XKK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts away a few groceries' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.3> to <12.3>." + } + ] + }, + { + "video_id": "EIB97", + "question_id": "EIB97", + "video_file": "Charades/videos/EIB97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is dressing in clothes from a hamper'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "EIB97", + "question_id": "EIB97", + "video_file": "Charades/videos/EIB97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put the bag on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.6> to <31.3>." + } + ] + }, + { + "video_id": "EIB97", + "question_id": "EIB97", + "video_file": "Charades/videos/EIB97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put it on the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.6> to <31.3>." + } + ] + }, + { + "video_id": "ZXHLY", + "question_id": "ZXHLY", + "video_file": "Charades/videos/ZXHLY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person washing the glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.0> to <16.0>." + } + ] + }, + { + "video_id": "ZXHLY", + "question_id": "ZXHLY", + "video_file": "Charades/videos/ZXHLY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person washes cup with water that is in the bucket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <16.0>." + } + ] + }, + { + "video_id": "P9D7P", + "question_id": "P9D7P", + "video_file": "Charades/videos/P9D7P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person looking out the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.2> to <27.3>." + } + ] + }, + { + "video_id": "P9D7P", + "question_id": "P9D7P", + "video_file": "Charades/videos/P9D7P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looks out window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <27.3>." + } + ] + }, + { + "video_id": "B6S6T", + "question_id": "B6S6T", + "video_file": "Charades/videos/B6S6T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looked out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.1> to <12.8>." + } + ] + }, + { + "video_id": "B6S6T", + "question_id": "B6S6T", + "video_file": "Charades/videos/B6S6T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person looks out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <12.8>." + } + ] + }, + { + "video_id": "B6S6T", + "question_id": "B6S6T", + "video_file": "Charades/videos/B6S6T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts a pillow in a clothes dryer'." + }, + { + "from": "gpt", + "value": "From <1.3> to <7.5>." + } + ] + }, + { + "video_id": "GFFJM", + "question_id": "GFFJM", + "video_file": "Charades/videos/GFFJM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.2> to <32.0>." + } + ] + }, + { + "video_id": "GFFJM", + "question_id": "GFFJM", + "video_file": "Charades/videos/GFFJM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person stands up to stretch turns off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.4> to <32.0>." + } + ] + }, + { + "video_id": "WBATR", + "question_id": "WBATR", + "video_file": "Charades/videos/WBATR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person undressing in a hallway'." + }, + { + "from": "gpt", + "value": "From <27.2> to <34.0>." + } + ] + }, + { + "video_id": "WBATR", + "question_id": "WBATR", + "video_file": "Charades/videos/WBATR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.6> to <32.5>." + } + ] + }, + { + "video_id": "WBATR", + "question_id": "WBATR", + "video_file": "Charades/videos/WBATR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.6> to <32.5>." + } + ] + }, + { + "video_id": "WBATR", + "question_id": "WBATR", + "video_file": "Charades/videos/WBATR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is drinking a glass of wine' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "DAS12", + "question_id": "DAS12", + "video_file": "Charades/videos/DAS12.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts them on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.8> to <35.0>." + } + ] + }, + { + "video_id": "DAS12", + "question_id": "DAS12", + "video_file": "Charades/videos/DAS12.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts shoes into a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.6> to <35.0>." + } + ] + }, + { + "video_id": "DAS12", + "question_id": "DAS12", + "video_file": "Charades/videos/DAS12.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the bag on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.6> to <35.0>." + } + ] + }, + { + "video_id": "DAS12", + "question_id": "DAS12", + "video_file": "Charades/videos/DAS12.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts some shoes in it'." + }, + { + "from": "gpt", + "value": "From <17.8> to <30.3>." + } + ] + }, + { + "video_id": "MUK89", + "question_id": "MUK89", + "video_file": "Charades/videos/MUK89.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns off a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.1> to <26.9>." + } + ] + }, + { + "video_id": "W2LM5", + "question_id": "W2LM5", + "video_file": "Charades/videos/W2LM5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person talking on a phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "W2LM5", + "question_id": "W2LM5", + "video_file": "Charades/videos/W2LM5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person talking on the phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "JJON5", + "question_id": "JJON5", + "video_file": "Charades/videos/JJON5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the refrigerator'." + }, + { + "from": "gpt", + "value": "From <6.7> to <14.7>." + } + ] + }, + { + "video_id": "JJON5", + "question_id": "JJON5", + "video_file": "Charades/videos/JJON5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person walks through a doorway to the kitchen' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <8.3>." + } + ] + }, + { + "video_id": "73X8Q", + "question_id": "73X8Q", + "video_file": "Charades/videos/73X8Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person drinks from a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <11.2>." + } + ] + }, + { + "video_id": "73X8Q", + "question_id": "73X8Q", + "video_file": "Charades/videos/73X8Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drink from their glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.6> to <11.2>." + } + ] + }, + { + "video_id": "73X8Q", + "question_id": "73X8Q", + "video_file": "Charades/videos/73X8Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person walks through a doorway'." + }, + { + "from": "gpt", + "value": "From <13.2> to <22.3>." + } + ] + }, + { + "video_id": "73X8Q", + "question_id": "73X8Q", + "video_file": "Charades/videos/73X8Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes a phone off the dryer' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.6> to <6.2>." + } + ] + }, + { + "video_id": "O5N93", + "question_id": "O5N93", + "video_file": "Charades/videos/O5N93.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person he puts the sandwich down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.6> to <29.0>." + } + ] + }, + { + "video_id": "O5N93", + "question_id": "O5N93", + "video_file": "Charades/videos/O5N93.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a smiling person is in their pantry eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.8>." + } + ] + }, + { + "video_id": "NDX21", + "question_id": "NDX21", + "video_file": "Charades/videos/NDX21.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding their phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <21.6>." + } + ] + }, + { + "video_id": "U0ACD", + "question_id": "U0ACD", + "video_file": "Charades/videos/U0ACD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes some medicine' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.0> to <28.3>." + } + ] + }, + { + "video_id": "WPR8G", + "question_id": "WPR8G", + "video_file": "Charades/videos/WPR8G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person bend down to put their shoes on' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.0> to <26.9>." + } + ] + }, + { + "video_id": "WPR8G", + "question_id": "WPR8G", + "video_file": "Charades/videos/WPR8G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begin putting on some shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.0> to <26.9>." + } + ] + }, + { + "video_id": "869NM", + "question_id": "869NM", + "video_file": "Charades/videos/869NM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opens a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.0> to <10.3>." + } + ] + }, + { + "video_id": "869NM", + "question_id": "869NM", + "video_file": "Charades/videos/869NM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they begin to open a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <10.3>." + } + ] + }, + { + "video_id": "99SGB", + "question_id": "99SGB", + "video_file": "Charades/videos/99SGB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opened the cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "9R4AQ", + "question_id": "9R4AQ", + "video_file": "Charades/videos/9R4AQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person standing drinking a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "9R4AQ", + "question_id": "9R4AQ", + "video_file": "Charades/videos/9R4AQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person drinks from the glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "N00HR", + "question_id": "N00HR", + "video_file": "Charades/videos/N00HR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looks out the window'." + }, + { + "from": "gpt", + "value": "From <19.1> to <32.0>." + } + ] + }, + { + "video_id": "N00HR", + "question_id": "N00HR", + "video_file": "Charades/videos/N00HR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person stand up' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.3> to <24.9>." + } + ] + }, + { + "video_id": "N00HR", + "question_id": "N00HR", + "video_file": "Charades/videos/N00HR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person look out a window'." + }, + { + "from": "gpt", + "value": "From <19.1> to <32.0>." + } + ] + }, + { + "video_id": "N00HR", + "question_id": "N00HR", + "video_file": "Charades/videos/N00HR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting in a chair smiling' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.1> to <10.3>." + } + ] + }, + { + "video_id": "N00HR", + "question_id": "N00HR", + "video_file": "Charades/videos/N00HR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks to look out a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.1> to <32.0>." + } + ] + }, + { + "video_id": "WIOCA", + "question_id": "WIOCA", + "video_file": "Charades/videos/WIOCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person talking on a phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.9> to <19.5>." + } + ] + }, + { + "video_id": "WIOCA", + "question_id": "WIOCA", + "video_file": "Charades/videos/WIOCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'encounters person b talking on the phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.9> to <19.5>." + } + ] + }, + { + "video_id": "S7RXC", + "question_id": "S7RXC", + "video_file": "Charades/videos/S7RXC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs into their hallway holding a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "S7RXC", + "question_id": "S7RXC", + "video_file": "Charades/videos/S7RXC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they turn on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <12.3>." + } + ] + }, + { + "video_id": "S7RXC", + "question_id": "S7RXC", + "video_file": "Charades/videos/S7RXC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they turn on a light'." + }, + { + "from": "gpt", + "value": "From <3.9> to <12.3>." + } + ] + }, + { + "video_id": "7HWK2", + "question_id": "7HWK2", + "video_file": "Charades/videos/7HWK2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person a book they put on their sweater'." + }, + { + "from": "gpt", + "value": "From <37.5> to <40.0>." + } + ] + }, + { + "video_id": "7HWK2", + "question_id": "7HWK2", + "video_file": "Charades/videos/7HWK2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the cup down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.9> to <15.0>." + } + ] + }, + { + "video_id": "7HWK2", + "question_id": "7HWK2", + "video_file": "Charades/videos/7HWK2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pouring coffee from a pot into a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.5>." + } + ] + }, + { + "video_id": "7HWK2", + "question_id": "7HWK2", + "video_file": "Charades/videos/7HWK2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person look at a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <37.5> to <40.0>." + } + ] + }, + { + "video_id": "7HWK2", + "question_id": "7HWK2", + "video_file": "Charades/videos/7HWK2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looking at a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <37.5> to <40.0>." + } + ] + }, + { + "video_id": "7HWK2", + "question_id": "7HWK2", + "video_file": "Charades/videos/7HWK2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person pours coffee into a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.5>." + } + ] + }, + { + "video_id": "WZA37", + "question_id": "WZA37", + "video_file": "Charades/videos/WZA37.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating food off a dish' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.4> to <14.8>." + } + ] + }, + { + "video_id": "WZA37", + "question_id": "WZA37", + "video_file": "Charades/videos/WZA37.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'dressing when another person opens the door holding a camera'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "NK9RX", + "question_id": "NK9RX", + "video_file": "Charades/videos/NK9RX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking something from a glass'." + }, + { + "from": "gpt", + "value": "From <13.2> to <28.2>." + } + ] + }, + { + "video_id": "NK9RX", + "question_id": "NK9RX", + "video_file": "Charades/videos/NK9RX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they begin drinking from a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <28.2>." + } + ] + }, + { + "video_id": "NK9RX", + "question_id": "NK9RX", + "video_file": "Charades/videos/NK9RX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person takes a camera' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.1> to <10.5>." + } + ] + }, + { + "video_id": "NK9RX", + "question_id": "NK9RX", + "video_file": "Charades/videos/NK9RX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts it by a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.0> to <9.1>." + } + ] + }, + { + "video_id": "9RA4P", + "question_id": "9RA4P", + "video_file": "Charades/videos/9RA4P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens the door to the rec room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <7.9>." + } + ] + }, + { + "video_id": "78FJX", + "question_id": "78FJX", + "video_file": "Charades/videos/78FJX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "78FJX", + "question_id": "78FJX", + "video_file": "Charades/videos/78FJX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes out a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.2> to <17.8>." + } + ] + }, + { + "video_id": "78FJX", + "question_id": "78FJX", + "video_file": "Charades/videos/78FJX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes a book out of a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.2> to <17.8>." + } + ] + }, + { + "video_id": "78FJX", + "question_id": "78FJX", + "video_file": "Charades/videos/78FJX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.4> to <22.5>." + } + ] + }, + { + "video_id": "Z1LVD", + "question_id": "Z1LVD", + "video_file": "Charades/videos/Z1LVD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person close the window'." + }, + { + "from": "gpt", + "value": "From <9.6> to <17.8>." + } + ] + }, + { + "video_id": "Z1LVD", + "question_id": "Z1LVD", + "video_file": "Charades/videos/Z1LVD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opens the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.1> to <9.3>." + } + ] + }, + { + "video_id": "Z1LVD", + "question_id": "Z1LVD", + "video_file": "Charades/videos/Z1LVD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.1> to <9.3>." + } + ] + }, + { + "video_id": "Z1LVD", + "question_id": "Z1LVD", + "video_file": "Charades/videos/Z1LVD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.6> to <17.8>." + } + ] + }, + { + "video_id": "Z1LVD", + "question_id": "Z1LVD", + "video_file": "Charades/videos/Z1LVD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looks out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.3> to <12.1>." + } + ] + }, + { + "video_id": "V4YCA", + "question_id": "V4YCA", + "video_file": "Charades/videos/V4YCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person throws the food in the garbage' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.9> to <21.8>." + } + ] + }, + { + "video_id": "V4YCA", + "question_id": "V4YCA", + "video_file": "Charades/videos/V4YCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a laughing person opens a refrigerator'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "RRZU3", + "question_id": "RRZU3", + "video_file": "Charades/videos/RRZU3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs into the room holding a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "RRZU3", + "question_id": "RRZU3", + "video_file": "Charades/videos/RRZU3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs into the living room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "RRZU3", + "question_id": "RRZU3", + "video_file": "Charades/videos/RRZU3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a drink from a nearby cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <21.6>." + } + ] + }, + { + "video_id": "RRZU3", + "question_id": "RRZU3", + "video_file": "Charades/videos/RRZU3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opens the laptop'." + }, + { + "from": "gpt", + "value": "From <3.3> to <11.0>." + } + ] + }, + { + "video_id": "5ZGRN", + "question_id": "5ZGRN", + "video_file": "Charades/videos/5ZGRN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "5ZGRN", + "question_id": "5ZGRN", + "video_file": "Charades/videos/5ZGRN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.6> to <15.8>." + } + ] + }, + { + "video_id": "5ZGRN", + "question_id": "5ZGRN", + "video_file": "Charades/videos/5ZGRN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is putting food in a cupboard'." + }, + { + "from": "gpt", + "value": "From <1.2> to <13.0>." + } + ] + }, + { + "video_id": "NGXV2", + "question_id": "NGXV2", + "video_file": "Charades/videos/NGXV2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening the book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.4> to <32.0>." + } + ] + }, + { + "video_id": "NGXV2", + "question_id": "NGXV2", + "video_file": "Charades/videos/NGXV2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <24.3>." + } + ] + }, + { + "video_id": "5C837", + "question_id": "5C837", + "video_file": "Charades/videos/5C837.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person smiling at a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <9.2>." + } + ] + }, + { + "video_id": "5C837", + "question_id": "5C837", + "video_file": "Charades/videos/5C837.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting at a desk smiling' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.9> to <9.2>." + } + ] + }, + { + "video_id": "G4MDY", + "question_id": "G4MDY", + "video_file": "Charades/videos/G4MDY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person stands by door of kitchen holding a grocery bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.2>." + } + ] + }, + { + "video_id": "G4MDY", + "question_id": "G4MDY", + "video_file": "Charades/videos/G4MDY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person turns on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.7> to <5.7>." + } + ] + }, + { + "video_id": "G4MDY", + "question_id": "G4MDY", + "video_file": "Charades/videos/G4MDY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person flips on the light'." + }, + { + "from": "gpt", + "value": "From <0.7> to <5.7>." + } + ] + }, + { + "video_id": "9LONJ", + "question_id": "9LONJ", + "video_file": "Charades/videos/9LONJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting the cup on top of the toilet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <10.6>." + } + ] + }, + { + "video_id": "9LONJ", + "question_id": "9LONJ", + "video_file": "Charades/videos/9LONJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sitting down on chair to unlace boots'." + }, + { + "from": "gpt", + "value": "From <8.2> to <15.0>." + } + ] + }, + { + "video_id": "JYC6G", + "question_id": "JYC6G", + "video_file": "Charades/videos/JYC6G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'as the person was dressing'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.9>." + } + ] + }, + { + "video_id": "LK1AE", + "question_id": "LK1AE", + "video_file": "Charades/videos/LK1AE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns the light off'." + }, + { + "from": "gpt", + "value": "From <21.2> to <26.8>." + } + ] + }, + { + "video_id": "LK1AE", + "question_id": "LK1AE", + "video_file": "Charades/videos/LK1AE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.2> to <26.8>." + } + ] + }, + { + "video_id": "13MWT", + "question_id": "13MWT", + "video_file": "Charades/videos/13MWT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turning off a light with a remote'." + }, + { + "from": "gpt", + "value": "From <15.4> to <22.8>." + } + ] + }, + { + "video_id": "13MWT", + "question_id": "13MWT", + "video_file": "Charades/videos/13MWT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes off their shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.6> to <11.3>." + } + ] + }, + { + "video_id": "IACBH", + "question_id": "IACBH", + "video_file": "Charades/videos/IACBH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person throws a blanket on the bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.0> to <12.2>." + } + ] + }, + { + "video_id": "IACBH", + "question_id": "IACBH", + "video_file": "Charades/videos/IACBH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person throws a blanket onto a bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.0> to <12.2>." + } + ] + }, + { + "video_id": "7JTEK", + "question_id": "7JTEK", + "video_file": "Charades/videos/7JTEK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person laughing as they cook at the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.3> to <24.8>." + } + ] + }, + { + "video_id": "9UDMH", + "question_id": "9UDMH", + "video_file": "Charades/videos/9UDMH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person watching the television' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.5> to <23.0>." + } + ] + }, + { + "video_id": "VE5DN", + "question_id": "VE5DN", + "video_file": "Charades/videos/VE5DN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <28.6>." + } + ] + }, + { + "video_id": "VE5DN", + "question_id": "VE5DN", + "video_file": "Charades/videos/VE5DN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put the pillow away in a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.0> to <29.2>." + } + ] + }, + { + "video_id": "VE5DN", + "question_id": "VE5DN", + "video_file": "Charades/videos/VE5DN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begin sneezing by the stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <28.2> to <34.0>." + } + ] + }, + { + "video_id": "VE5DN", + "question_id": "VE5DN", + "video_file": "Charades/videos/VE5DN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put it on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.4> to <31.0>." + } + ] + }, + { + "video_id": "VE5DN", + "question_id": "VE5DN", + "video_file": "Charades/videos/VE5DN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakens on a floor lying on a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.4> to <11.8>." + } + ] + }, + { + "video_id": "W74ZY", + "question_id": "W74ZY", + "video_file": "Charades/videos/W74ZY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits on couch' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <15.8>." + } + ] + }, + { + "video_id": "W74ZY", + "question_id": "W74ZY", + "video_file": "Charades/videos/W74ZY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person sits down on the sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <15.8>." + } + ] + }, + { + "video_id": "6E6GF", + "question_id": "6E6GF", + "video_file": "Charades/videos/6E6GF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put dishes in a sink' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.6> to <45.1>." + } + ] + }, + { + "video_id": "6E6GF", + "question_id": "6E6GF", + "video_file": "Charades/videos/6E6GF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begins cooking' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.1> to <25.7>." + } + ] + }, + { + "video_id": "6E6GF", + "question_id": "6E6GF", + "video_file": "Charades/videos/6E6GF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person someone is in the kitchen cooking something' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.1> to <25.7>." + } + ] + }, + { + "video_id": "6E6GF", + "question_id": "6E6GF", + "video_file": "Charades/videos/6E6GF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person turns the light on over the stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.2> to <10.1>." + } + ] + }, + { + "video_id": "6E6GF", + "question_id": "6E6GF", + "video_file": "Charades/videos/6E6GF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they turn on a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <10.1>." + } + ] + }, + { + "video_id": "14TJT", + "question_id": "14TJT", + "video_file": "Charades/videos/14TJT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <9.1>." + } + ] + }, + { + "video_id": "8GP10", + "question_id": "8GP10", + "video_file": "Charades/videos/8GP10.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they are eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <17.5>." + } + ] + }, + { + "video_id": "OD1A5", + "question_id": "OD1A5", + "video_file": "Charades/videos/OD1A5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is standing in the bathroom washing their hands'." + }, + { + "from": "gpt", + "value": "From <0.1> to <11.8>." + } + ] + }, + { + "video_id": "OD1A5", + "question_id": "OD1A5", + "video_file": "Charades/videos/OD1A5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person switched off the lights'." + }, + { + "from": "gpt", + "value": "From <28.2> to <32.7>." + } + ] + }, + { + "video_id": "OD1A5", + "question_id": "OD1A5", + "video_file": "Charades/videos/OD1A5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person washed his hands seen in the mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <11.8>." + } + ] + }, + { + "video_id": "OD1A5", + "question_id": "OD1A5", + "video_file": "Charades/videos/OD1A5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person washes their hands' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.1> to <11.8>." + } + ] + }, + { + "video_id": "OD1A5", + "question_id": "OD1A5", + "video_file": "Charades/videos/OD1A5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a picture of themselves with a camera' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <23.6>." + } + ] + }, + { + "video_id": "OD1A5", + "question_id": "OD1A5", + "video_file": "Charades/videos/OD1A5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person turns off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.2> to <32.7>." + } + ] + }, + { + "video_id": "OD1A5", + "question_id": "OD1A5", + "video_file": "Charades/videos/OD1A5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person cuts the light off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.2> to <32.7>." + } + ] + }, + { + "video_id": "PQBGG", + "question_id": "PQBGG", + "video_file": "Charades/videos/PQBGG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looks out a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.1> to <10.1>." + } + ] + }, + { + "video_id": "PQBGG", + "question_id": "PQBGG", + "video_file": "Charades/videos/PQBGG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes off their shoes'." + }, + { + "from": "gpt", + "value": "From <23.5> to <41.0>." + } + ] + }, + { + "video_id": "PQBGG", + "question_id": "PQBGG", + "video_file": "Charades/videos/PQBGG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.7> to <41.0>." + } + ] + }, + { + "video_id": "7JSAI", + "question_id": "7JSAI", + "video_file": "Charades/videos/7JSAI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person standing near by starts smiling' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.5> to <10.9>." + } + ] + }, + { + "video_id": "7JSAI", + "question_id": "7JSAI", + "video_file": "Charades/videos/7JSAI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person than puts their bag on a glass table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.5> to <22.9>." + } + ] + }, + { + "video_id": "KJK56", + "question_id": "KJK56", + "video_file": "Charades/videos/KJK56.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting on the sofa holding a phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.6>." + } + ] + }, + { + "video_id": "KJK56", + "question_id": "KJK56", + "video_file": "Charades/videos/KJK56.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes off their shoes,and goes to the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.8> to <13.9>." + } + ] + }, + { + "video_id": "XP305", + "question_id": "XP305", + "video_file": "Charades/videos/XP305.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting down on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <19.0>." + } + ] + }, + { + "video_id": "XP305", + "question_id": "XP305", + "video_file": "Charades/videos/XP305.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throwing a pillow on the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <17.3>." + } + ] + }, + { + "video_id": "XP305", + "question_id": "XP305", + "video_file": "Charades/videos/XP305.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <19.0>." + } + ] + }, + { + "video_id": "XP305", + "question_id": "XP305", + "video_file": "Charades/videos/XP305.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throw it on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <17.8>." + } + ] + }, + { + "video_id": "IZTHW", + "question_id": "IZTHW", + "video_file": "Charades/videos/IZTHW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person laughing at what is on the screen' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <29.4> to <34.0>." + } + ] + }, + { + "video_id": "IGVLE", + "question_id": "IGVLE", + "video_file": "Charades/videos/IGVLE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take a picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.9> to <33.0>." + } + ] + }, + { + "video_id": "IGVLE", + "question_id": "IGVLE", + "video_file": "Charades/videos/IGVLE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts undressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <27.1>." + } + ] + }, + { + "video_id": "IGVLE", + "question_id": "IGVLE", + "video_file": "Charades/videos/IGVLE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes out a cell phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.0> to <33.0>." + } + ] + }, + { + "video_id": "VAIVN", + "question_id": "VAIVN", + "video_file": "Charades/videos/VAIVN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is laying on a sofa reading a book'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "VAIVN", + "question_id": "VAIVN", + "video_file": "Charades/videos/VAIVN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins to snuggle with a blanket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.8> to <24.8>." + } + ] + }, + { + "video_id": "VAIVN", + "question_id": "VAIVN", + "video_file": "Charades/videos/VAIVN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes off their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.4> to <30.9>." + } + ] + }, + { + "video_id": "K6DF2", + "question_id": "K6DF2", + "video_file": "Charades/videos/K6DF2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting it on a table'." + }, + { + "from": "gpt", + "value": "From <4.2> to <10.4>." + } + ] + }, + { + "video_id": "K6DF2", + "question_id": "K6DF2", + "video_file": "Charades/videos/K6DF2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking from the glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <24.3>." + } + ] + }, + { + "video_id": "K6DF2", + "question_id": "K6DF2", + "video_file": "Charades/videos/K6DF2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is seen putting their water on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <10.4>." + } + ] + }, + { + "video_id": "K6DF2", + "question_id": "K6DF2", + "video_file": "Charades/videos/K6DF2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is holding a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.3>." + } + ] + }, + { + "video_id": "K6DF2", + "question_id": "K6DF2", + "video_file": "Charades/videos/K6DF2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turn on a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.4> to <27.5>." + } + ] + }, + { + "video_id": "K6DF2", + "question_id": "K6DF2", + "video_file": "Charades/videos/K6DF2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns on a light'." + }, + { + "from": "gpt", + "value": "From <21.4> to <27.5>." + } + ] + }, + { + "video_id": "DY9AE", + "question_id": "DY9AE", + "video_file": "Charades/videos/DY9AE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person eats food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "DY9AE", + "question_id": "DY9AE", + "video_file": "Charades/videos/DY9AE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.1> to <27.0>." + } + ] + }, + { + "video_id": "BOKLM", + "question_id": "BOKLM", + "video_file": "Charades/videos/BOKLM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits to put on shoes'." + }, + { + "from": "gpt", + "value": "From <7.2> to <16.3>." + } + ] + }, + { + "video_id": "BOKLM", + "question_id": "BOKLM", + "video_file": "Charades/videos/BOKLM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person open a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.3> to <32.0>." + } + ] + }, + { + "video_id": "BOKLM", + "question_id": "BOKLM", + "video_file": "Charades/videos/BOKLM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts their shoes on'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.2> to <16.3>." + } + ] + }, + { + "video_id": "Q166J", + "question_id": "Q166J", + "video_file": "Charades/videos/Q166J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person begins opening a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.6> to <15.0>." + } + ] + }, + { + "video_id": "A7OCU", + "question_id": "A7OCU", + "video_file": "Charades/videos/A7OCU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they begin dressing by putting a jacket on'." + }, + { + "from": "gpt", + "value": "From <6.2> to <19.9>." + } + ] + }, + { + "video_id": "A7OCU", + "question_id": "A7OCU", + "video_file": "Charades/videos/A7OCU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person putting a dish on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.1> to <25.0>." + } + ] + }, + { + "video_id": "D32A4", + "question_id": "D32A4", + "video_file": "Charades/videos/D32A4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting the picture of his daughter on his desk' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.9> to <17.4>." + } + ] + }, + { + "video_id": "D32A4", + "question_id": "D32A4", + "video_file": "Charades/videos/D32A4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person was smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.1> to <15.0>." + } + ] + }, + { + "video_id": "M91YM", + "question_id": "M91YM", + "video_file": "Charades/videos/M91YM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <24.5>." + } + ] + }, + { + "video_id": "M91YM", + "question_id": "M91YM", + "video_file": "Charades/videos/M91YM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.2> to <8.3>." + } + ] + }, + { + "video_id": "M91YM", + "question_id": "M91YM", + "video_file": "Charades/videos/M91YM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person turns on the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.2> to <8.3>." + } + ] + }, + { + "video_id": "YPIKO", + "question_id": "YPIKO", + "video_file": "Charades/videos/YPIKO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pours soda into a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.7> to <27.3>." + } + ] + }, + { + "video_id": "YPIKO", + "question_id": "YPIKO", + "video_file": "Charades/videos/YPIKO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person watching television' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.6> to <35.5>." + } + ] + }, + { + "video_id": "YPIKO", + "question_id": "YPIKO", + "video_file": "Charades/videos/YPIKO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting in a chair eating food'." + }, + { + "from": "gpt", + "value": "From <3.0> to <9.1>." + } + ] + }, + { + "video_id": "YPIKO", + "question_id": "YPIKO", + "video_file": "Charades/videos/YPIKO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating a watching television' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <9.1>." + } + ] + }, + { + "video_id": "VCV6A", + "question_id": "VCV6A", + "video_file": "Charades/videos/VCV6A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they also take a sandwich out of the bag'." + }, + { + "from": "gpt", + "value": "From <9.4> to <14.0>." + } + ] + }, + { + "video_id": "VCV6A", + "question_id": "VCV6A", + "video_file": "Charades/videos/VCV6A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eat it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.9> to <21.1>." + } + ] + }, + { + "video_id": "VCV6A", + "question_id": "VCV6A", + "video_file": "Charades/videos/VCV6A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <4.3>." + } + ] + }, + { + "video_id": "VCV6A", + "question_id": "VCV6A", + "video_file": "Charades/videos/VCV6A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person take a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <14.7>." + } + ] + }, + { + "video_id": "VCV6A", + "question_id": "VCV6A", + "video_file": "Charades/videos/VCV6A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person takes a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.4> to <14.0>." + } + ] + }, + { + "video_id": "RTXE6", + "question_id": "RTXE6", + "video_file": "Charades/videos/RTXE6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they are both smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "RTXE6", + "question_id": "RTXE6", + "video_file": "Charades/videos/RTXE6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person are laughing at something funny on their phone'." + }, + { + "from": "gpt", + "value": "From <10.0> to <16.5>." + } + ] + }, + { + "video_id": "DPDER", + "question_id": "DPDER", + "video_file": "Charades/videos/DPDER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person closes the laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <21.0>." + } + ] + }, + { + "video_id": "DPDER", + "question_id": "DPDER", + "video_file": "Charades/videos/DPDER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person close the laptop'." + }, + { + "from": "gpt", + "value": "From <15.0> to <21.0>." + } + ] + }, + { + "video_id": "DPDER", + "question_id": "DPDER", + "video_file": "Charades/videos/DPDER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they get dressed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <24.1>." + } + ] + }, + { + "video_id": "5GITI", + "question_id": "5GITI", + "video_file": "Charades/videos/5GITI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the towel down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <6.4>." + } + ] + }, + { + "video_id": "F8UU2", + "question_id": "F8UU2", + "video_file": "Charades/videos/F8UU2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person is putting up groceries'." + }, + { + "from": "gpt", + "value": "From <3.5> to <15.1>." + } + ] + }, + { + "video_id": "F8UU2", + "question_id": "F8UU2", + "video_file": "Charades/videos/F8UU2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.7> to <16.2>." + } + ] + }, + { + "video_id": "H94MQ", + "question_id": "H94MQ", + "video_file": "Charades/videos/H94MQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "H94MQ", + "question_id": "H94MQ", + "video_file": "Charades/videos/H94MQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "C4MP2", + "question_id": "C4MP2", + "video_file": "Charades/videos/C4MP2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes a sandwich from their bag'." + }, + { + "from": "gpt", + "value": "From <24.4> to <35.1>." + } + ] + }, + { + "video_id": "2K755", + "question_id": "2K755", + "video_file": "Charades/videos/2K755.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person is undressing from some clothes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.1> to <20.7>." + } + ] + }, + { + "video_id": "2K755", + "question_id": "2K755", + "video_file": "Charades/videos/2K755.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person is undressing by the wardrobe' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.1> to <20.7>." + } + ] + }, + { + "video_id": "5H1P1", + "question_id": "5H1P1", + "video_file": "Charades/videos/5H1P1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person laughing at a show on the laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "5H1P1", + "question_id": "5H1P1", + "video_file": "Charades/videos/5H1P1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person laughs at something on a laptop screen'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "NOFSE", + "question_id": "NOFSE", + "video_file": "Charades/videos/NOFSE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person laying down on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.1> to <28.0>." + } + ] + }, + { + "video_id": "NOFSE", + "question_id": "NOFSE", + "video_file": "Charades/videos/NOFSE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes out a pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.7> to <16.4>." + } + ] + }, + { + "video_id": "WJFGC", + "question_id": "WJFGC", + "video_file": "Charades/videos/WJFGC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they open the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <15.6>." + } + ] + }, + { + "video_id": "WJFGC", + "question_id": "WJFGC", + "video_file": "Charades/videos/WJFGC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person stand as they take the laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.4> to <17.4>." + } + ] + }, + { + "video_id": "YBO6N", + "question_id": "YBO6N", + "video_file": "Charades/videos/YBO6N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes a cabinet door in another room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.2> to <28.0>." + } + ] + }, + { + "video_id": "YBO6N", + "question_id": "YBO6N", + "video_file": "Charades/videos/YBO6N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns on a light,opens' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.2> to <20.8>." + } + ] + }, + { + "video_id": "7W66P", + "question_id": "7W66P", + "video_file": "Charades/videos/7W66P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits on the sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "7W66P", + "question_id": "7W66P", + "video_file": "Charades/videos/7W66P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is watching television' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <12.7>." + } + ] + }, + { + "video_id": "7W66P", + "question_id": "7W66P", + "video_file": "Charades/videos/7W66P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits back on the couch' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "7W66P", + "question_id": "7W66P", + "video_file": "Charades/videos/7W66P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person watches television' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <12.7>." + } + ] + }, + { + "video_id": "7W66P", + "question_id": "7W66P", + "video_file": "Charades/videos/7W66P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sit down on a sofa' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "N79WJ", + "question_id": "N79WJ", + "video_file": "Charades/videos/N79WJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put it on a shelf'." + }, + { + "from": "gpt", + "value": "From <15.7> to <24.8>." + } + ] + }, + { + "video_id": "6JKD6", + "question_id": "6JKD6", + "video_file": "Charades/videos/6JKD6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is standing in the bathroom opening a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <16.4>." + } + ] + }, + { + "video_id": "6JKD6", + "question_id": "6JKD6", + "video_file": "Charades/videos/6JKD6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting a mirror into the box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.8> to <11.2>." + } + ] + }, + { + "video_id": "6JKD6", + "question_id": "6JKD6", + "video_file": "Charades/videos/6JKD6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes something out of a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.3> to <17.5>." + } + ] + }, + { + "video_id": "6JKD6", + "question_id": "6JKD6", + "video_file": "Charades/videos/6JKD6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put something back in the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <11.2>." + } + ] + }, + { + "video_id": "Q4TKG", + "question_id": "Q4TKG", + "video_file": "Charades/videos/Q4TKG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'that same person throws the pillow on the bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.2> to <10.7>." + } + ] + }, + { + "video_id": "Q4TKG", + "question_id": "Q4TKG", + "video_file": "Charades/videos/Q4TKG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they throw the pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.2> to <10.7>." + } + ] + }, + { + "video_id": "DJG7A", + "question_id": "DJG7A", + "video_file": "Charades/videos/DJG7A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts some clothes on it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.9> to <11.9>." + } + ] + }, + { + "video_id": "Q5ZIL", + "question_id": "Q5ZIL", + "video_file": "Charades/videos/Q5ZIL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes the medicine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.7> to <34.0>." + } + ] + }, + { + "video_id": "Q5ZIL", + "question_id": "Q5ZIL", + "video_file": "Charades/videos/Q5ZIL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person take some medicine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.7> to <34.0>." + } + ] + }, + { + "video_id": "Q5ZIL", + "question_id": "Q5ZIL", + "video_file": "Charades/videos/Q5ZIL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they begin sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "8VSV6", + "question_id": "8VSV6", + "video_file": "Charades/videos/8VSV6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.4> to <17.0>." + } + ] + }, + { + "video_id": "EHYXP", + "question_id": "EHYXP", + "video_file": "Charades/videos/EHYXP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person smiling the whole time'." + }, + { + "from": "gpt", + "value": "From <2.2> to <7.7>." + } + ] + }, + { + "video_id": "C9VPX", + "question_id": "C9VPX", + "video_file": "Charades/videos/C9VPX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes a pillow out of the washing machine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.8> to <13.4>." + } + ] + }, + { + "video_id": "C9VPX", + "question_id": "C9VPX", + "video_file": "Charades/videos/C9VPX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes a pillow from a washing machine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.8> to <13.4>." + } + ] + }, + { + "video_id": "5AR9B", + "question_id": "5AR9B", + "video_file": "Charades/videos/5AR9B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking out of a cup'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "5AR9B", + "question_id": "5AR9B", + "video_file": "Charades/videos/5AR9B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinks from a coffee cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "F082Z", + "question_id": "F082Z", + "video_file": "Charades/videos/F082Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.9> to <33.0>." + } + ] + }, + { + "video_id": "1JGRO", + "question_id": "1JGRO", + "video_file": "Charades/videos/1JGRO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person also eating a burrito' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.7> to <21.2>." + } + ] + }, + { + "video_id": "81YUE", + "question_id": "81YUE", + "video_file": "Charades/videos/81YUE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts watching television' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.6> to <25.7>." + } + ] + }, + { + "video_id": "81YUE", + "question_id": "81YUE", + "video_file": "Charades/videos/81YUE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins undressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.9> to <29.7>." + } + ] + }, + { + "video_id": "81YUE", + "question_id": "81YUE", + "video_file": "Charades/videos/81YUE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is lying on a bed watching television' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.6> to <25.7>." + } + ] + }, + { + "video_id": "WE2PF", + "question_id": "WE2PF", + "video_file": "Charades/videos/WE2PF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person sits in a chair holding a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.6>." + } + ] + }, + { + "video_id": "WE2PF", + "question_id": "WE2PF", + "video_file": "Charades/videos/WE2PF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they started to try to eat it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.6> to <22.6>." + } + ] + }, + { + "video_id": "EG1XK", + "question_id": "EG1XK", + "video_file": "Charades/videos/EG1XK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.9> to <35.0>." + } + ] + }, + { + "video_id": "EG1XK", + "question_id": "EG1XK", + "video_file": "Charades/videos/EG1XK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours a cup of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.9> to <16.4>." + } + ] + }, + { + "video_id": "EG1XK", + "question_id": "EG1XK", + "video_file": "Charades/videos/EG1XK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours water into the glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.9> to <16.4>." + } + ] + }, + { + "video_id": "EG1XK", + "question_id": "EG1XK", + "video_file": "Charades/videos/EG1XK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinks from the cup two times'." + }, + { + "from": "gpt", + "value": "From <13.9> to <22.1>." + } + ] + }, + { + "video_id": "7XR13", + "question_id": "7XR13", + "video_file": "Charades/videos/7XR13.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put had put on the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.0> to <29.0>." + } + ] + }, + { + "video_id": "XSS0J", + "question_id": "XSS0J", + "video_file": "Charades/videos/XSS0J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open a laptop'." + }, + { + "from": "gpt", + "value": "From <22.4> to <31.7>." + } + ] + }, + { + "video_id": "XSS0J", + "question_id": "XSS0J", + "video_file": "Charades/videos/XSS0J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opens their laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.4> to <31.7>." + } + ] + }, + { + "video_id": "P8UT3", + "question_id": "P8UT3", + "video_file": "Charades/videos/P8UT3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person close the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.6> to <27.6>." + } + ] + }, + { + "video_id": "P8UT3", + "question_id": "P8UT3", + "video_file": "Charades/videos/P8UT3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.6> to <27.6>." + } + ] + }, + { + "video_id": "OLEWM", + "question_id": "OLEWM", + "video_file": "Charades/videos/OLEWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person is eating' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "OLEWM", + "question_id": "OLEWM", + "video_file": "Charades/videos/OLEWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting in a chair eating' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "5U70R", + "question_id": "5U70R", + "video_file": "Charades/videos/5U70R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person sitting closes the laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.8> to <34.0>." + } + ] + }, + { + "video_id": "5U70R", + "question_id": "5U70R", + "video_file": "Charades/videos/5U70R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person closes the laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.8> to <34.0>." + } + ] + }, + { + "video_id": "HKJUR", + "question_id": "HKJUR", + "video_file": "Charades/videos/HKJUR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person is closing the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.2> to <27.3>." + } + ] + }, + { + "video_id": "HKJUR", + "question_id": "HKJUR", + "video_file": "Charades/videos/HKJUR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening a refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <10.9>." + } + ] + }, + { + "video_id": "YQCYJ", + "question_id": "YQCYJ", + "video_file": "Charades/videos/YQCYJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person is washing dishes in a sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.5> to <13.0>." + } + ] + }, + { + "video_id": "CK3H7", + "question_id": "CK3H7", + "video_file": "Charades/videos/CK3H7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put clothes away into a wardrobe'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.0> to <30.0>." + } + ] + }, + { + "video_id": "X3KFE", + "question_id": "X3KFE", + "video_file": "Charades/videos/X3KFE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinks the water in the glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.4> to <21.3>." + } + ] + }, + { + "video_id": "X3KFE", + "question_id": "X3KFE", + "video_file": "Charades/videos/X3KFE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holds a glass up high'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.8> to <17.3>." + } + ] + }, + { + "video_id": "X3KFE", + "question_id": "X3KFE", + "video_file": "Charades/videos/X3KFE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they hold up a glass of water'." + }, + { + "from": "gpt", + "value": "From <10.8> to <17.3>." + } + ] + }, + { + "video_id": "X3KFE", + "question_id": "X3KFE", + "video_file": "Charades/videos/X3KFE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a picture of theirself with a black camera' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.9>." + } + ] + }, + { + "video_id": "X3KFE", + "question_id": "X3KFE", + "video_file": "Charades/videos/X3KFE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person hold a cup in the air' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.8> to <17.3>." + } + ] + }, + { + "video_id": "PFCRB", + "question_id": "PFCRB", + "video_file": "Charades/videos/PFCRB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person looking out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "PFCRB", + "question_id": "PFCRB", + "video_file": "Charades/videos/PFCRB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is looking out the window holding a phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "PFCRB", + "question_id": "PFCRB", + "video_file": "Charades/videos/PFCRB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person start sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.3> to <22.3>." + } + ] + }, + { + "video_id": "4CHXK", + "question_id": "4CHXK", + "video_file": "Charades/videos/4CHXK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.8> to <12.1>." + } + ] + }, + { + "video_id": "4CHXK", + "question_id": "4CHXK", + "video_file": "Charades/videos/4CHXK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throw the glass on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.8> to <25.0>." + } + ] + }, + { + "video_id": "4QSRS", + "question_id": "4QSRS", + "video_file": "Charades/videos/4QSRS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person is standing holding a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.7>." + } + ] + }, + { + "video_id": "4QSRS", + "question_id": "4QSRS", + "video_file": "Charades/videos/4QSRS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person puts a book down onto the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.6>." + } + ] + }, + { + "video_id": "4QSRS", + "question_id": "4QSRS", + "video_file": "Charades/videos/4QSRS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting it on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.6>." + } + ] + }, + { + "video_id": "PW3GG", + "question_id": "PW3GG", + "video_file": "Charades/videos/PW3GG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they open the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <23.6>." + } + ] + }, + { + "video_id": "27SS2", + "question_id": "27SS2", + "video_file": "Charades/videos/27SS2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person proceeds to his window to look out'." + }, + { + "from": "gpt", + "value": "From <24.7> to <31.0>." + } + ] + }, + { + "video_id": "27SS2", + "question_id": "27SS2", + "video_file": "Charades/videos/27SS2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they take a drink from a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <17.3>." + } + ] + }, + { + "video_id": "7OXNH", + "question_id": "7OXNH", + "video_file": "Charades/videos/7OXNH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinks from a white cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.1> to <12.3>." + } + ] + }, + { + "video_id": "YKT8I", + "question_id": "YKT8I", + "video_file": "Charades/videos/YKT8I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the door to the bathroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "YKT8I", + "question_id": "YKT8I", + "video_file": "Charades/videos/YKT8I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "82ZNR", + "question_id": "82ZNR", + "video_file": "Charades/videos/82ZNR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is cooking by the stove'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.6>." + } + ] + }, + { + "video_id": "KFI7N", + "question_id": "KFI7N", + "video_file": "Charades/videos/KFI7N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws a broom onto the ground' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <10.0>." + } + ] + }, + { + "video_id": "KFI7N", + "question_id": "KFI7N", + "video_file": "Charades/videos/KFI7N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they throw the broom on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <10.0>." + } + ] + }, + { + "video_id": "KFI7N", + "question_id": "KFI7N", + "video_file": "Charades/videos/KFI7N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks through a doorway into an office'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "EWY6I", + "question_id": "EWY6I", + "video_file": "Charades/videos/EWY6I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person drinks a cup of coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.2>." + } + ] + }, + { + "video_id": "AAS9S", + "question_id": "AAS9S", + "video_file": "Charades/videos/AAS9S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts a small towel on the doorknob'." + }, + { + "from": "gpt", + "value": "From <19.4> to <29.8>." + } + ] + }, + { + "video_id": "AAS9S", + "question_id": "AAS9S", + "video_file": "Charades/videos/AAS9S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take a towel' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.9> to <25.6>." + } + ] + }, + { + "video_id": "O1ILB", + "question_id": "O1ILB", + "video_file": "Charades/videos/O1ILB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person runs out'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <18.0>." + } + ] + }, + { + "video_id": "O1ILB", + "question_id": "O1ILB", + "video_file": "Charades/videos/O1ILB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs into the garage' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <18.0>." + } + ] + }, + { + "video_id": "O1ILB", + "question_id": "O1ILB", + "video_file": "Charades/videos/O1ILB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person comes running into the garage' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.4> to <18.0>." + } + ] + }, + { + "video_id": "O1ILB", + "question_id": "O1ILB", + "video_file": "Charades/videos/O1ILB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person runs back out' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <18.0>." + } + ] + }, + { + "video_id": "8M0ZV", + "question_id": "8M0ZV", + "video_file": "Charades/videos/8M0ZV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a second person walks in drinking a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.5> to <31.0>." + } + ] + }, + { + "video_id": "8M0ZV", + "question_id": "8M0ZV", + "video_file": "Charades/videos/8M0ZV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person walks in drinking a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <27.5> to <31.0>." + } + ] + }, + { + "video_id": "94KP4", + "question_id": "94KP4", + "video_file": "Charades/videos/94KP4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.9> to <19.3>." + } + ] + }, + { + "video_id": "94KP4", + "question_id": "94KP4", + "video_file": "Charades/videos/94KP4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person stands in a room eating some food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.5>." + } + ] + }, + { + "video_id": "94KP4", + "question_id": "94KP4", + "video_file": "Charades/videos/94KP4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put the dish down on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.9> to <12.9>." + } + ] + }, + { + "video_id": "FQDD1", + "question_id": "FQDD1", + "video_file": "Charades/videos/FQDD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is undressing in the doorway of the bathroom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <17.0>." + } + ] + }, + { + "video_id": "FQDD1", + "question_id": "FQDD1", + "video_file": "Charades/videos/FQDD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is undressing in a doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <17.0>." + } + ] + }, + { + "video_id": "HB4AB", + "question_id": "HB4AB", + "video_file": "Charades/videos/HB4AB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.7> to <10.6>." + } + ] + }, + { + "video_id": "HB4AB", + "question_id": "HB4AB", + "video_file": "Charades/videos/HB4AB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person laugh when they drop something' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.5> to <24.0>." + } + ] + }, + { + "video_id": "0FO58", + "question_id": "0FO58", + "video_file": "Charades/videos/0FO58.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is walking through a hallway sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <13.1>." + } + ] + }, + { + "video_id": "0FO58", + "question_id": "0FO58", + "video_file": "Charades/videos/0FO58.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <13.1>." + } + ] + }, + { + "video_id": "0FO58", + "question_id": "0FO58", + "video_file": "Charades/videos/0FO58.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sneezing in the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <13.1>." + } + ] + }, + { + "video_id": "9335E", + "question_id": "9335E", + "video_file": "Charades/videos/9335E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin cooking on a stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.8> to <35.4>." + } + ] + }, + { + "video_id": "9335E", + "question_id": "9335E", + "video_file": "Charades/videos/9335E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person stands by the stove to do some cooking' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.8> to <35.4>." + } + ] + }, + { + "video_id": "9335E", + "question_id": "9335E", + "video_file": "Charades/videos/9335E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <37.5> to <43.0>." + } + ] + }, + { + "video_id": "9335E", + "question_id": "9335E", + "video_file": "Charades/videos/9335E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting in a chair making a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <23.3>." + } + ] + }, + { + "video_id": "9335E", + "question_id": "9335E", + "video_file": "Charades/videos/9335E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sits in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <23.3>." + } + ] + }, + { + "video_id": "9335E", + "question_id": "9335E", + "video_file": "Charades/videos/9335E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person cooking on the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.8> to <35.4>." + } + ] + }, + { + "video_id": "9335E", + "question_id": "9335E", + "video_file": "Charades/videos/9335E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person makes a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.1> to <21.4>." + } + ] + }, + { + "video_id": "6KUUU", + "question_id": "6KUUU", + "video_file": "Charades/videos/6KUUU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put boots on a eat a sandwich'." + }, + { + "from": "gpt", + "value": "From <21.0> to <32.0>." + } + ] + }, + { + "video_id": "6KUUU", + "question_id": "6KUUU", + "video_file": "Charades/videos/6KUUU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts eating it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.8> to <32.0>." + } + ] + }, + { + "video_id": "QH3FK", + "question_id": "QH3FK", + "video_file": "Charades/videos/QH3FK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they drink water from the glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.9> to <32.0>." + } + ] + }, + { + "video_id": "N9D4C", + "question_id": "N9D4C", + "video_file": "Charades/videos/N9D4C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person snuggle their pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.3> to <16.2>." + } + ] + }, + { + "video_id": "N9D4C", + "question_id": "N9D4C", + "video_file": "Charades/videos/N9D4C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person snuggled with a pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <16.2>." + } + ] + }, + { + "video_id": "QYSER", + "question_id": "QYSER", + "video_file": "Charades/videos/QYSER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person lays down on a couch eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <7.8>." + } + ] + }, + { + "video_id": "QYSER", + "question_id": "QYSER", + "video_file": "Charades/videos/QYSER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person messily eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <1.7> to <7.8>." + } + ] + }, + { + "video_id": "5CT0V", + "question_id": "5CT0V", + "video_file": "Charades/videos/5CT0V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person playing on a phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.8>." + } + ] + }, + { + "video_id": "AIT3V", + "question_id": "AIT3V", + "video_file": "Charades/videos/AIT3V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.1> to <28.9>." + } + ] + }, + { + "video_id": "R1OUC", + "question_id": "R1OUC", + "video_file": "Charades/videos/R1OUC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person wards they begin washing down a mirror' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <29.9> to <41.0>." + } + ] + }, + { + "video_id": "R1OUC", + "question_id": "R1OUC", + "video_file": "Charades/videos/R1OUC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the washer door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.6> to <10.3>." + } + ] + }, + { + "video_id": "OZIJ7", + "question_id": "OZIJ7", + "video_file": "Charades/videos/OZIJ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person laughs at a glass sitting on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.3> to <23.8>." + } + ] + }, + { + "video_id": "OZIJ7", + "question_id": "OZIJ7", + "video_file": "Charades/videos/OZIJ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person in the bedroom is laughing'." + }, + { + "from": "gpt", + "value": "From <15.3> to <23.8>." + } + ] + }, + { + "video_id": "TDNV3", + "question_id": "TDNV3", + "video_file": "Charades/videos/TDNV3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting at the table eating food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.6> to <16.5>." + } + ] + }, + { + "video_id": "RJIEO", + "question_id": "RJIEO", + "video_file": "Charades/videos/RJIEO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits the bag of groceries on the bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <14.6>." + } + ] + }, + { + "video_id": "83J53", + "question_id": "83J53", + "video_file": "Charades/videos/83J53.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'second person walks down stairs holding a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.2> to <21.7>." + } + ] + }, + { + "video_id": "BKEYX", + "question_id": "BKEYX", + "video_file": "Charades/videos/BKEYX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts shoes onto a shelf'." + }, + { + "from": "gpt", + "value": "From <5.2> to <11.5>." + } + ] + }, + { + "video_id": "BKEYX", + "question_id": "BKEYX", + "video_file": "Charades/videos/BKEYX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a drink from a bottle'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.3> to <32.1>." + } + ] + }, + { + "video_id": "RO8N5", + "question_id": "RO8N5", + "video_file": "Charades/videos/RO8N5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinks from the glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.5> to <32.0>." + } + ] + }, + { + "video_id": "AEJ4T", + "question_id": "AEJ4T", + "video_file": "Charades/videos/AEJ4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person a opens a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <14.0>." + } + ] + }, + { + "video_id": "201W8", + "question_id": "201W8", + "video_file": "Charades/videos/201W8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person closes the book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <18.0>." + } + ] + }, + { + "video_id": "201W8", + "question_id": "201W8", + "video_file": "Charades/videos/201W8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a book in the bathroom'." + }, + { + "from": "gpt", + "value": "From <0.1> to <5.6>." + } + ] + }, + { + "video_id": "201W8", + "question_id": "201W8", + "video_file": "Charades/videos/201W8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <18.0>." + } + ] + }, + { + "video_id": "OXUKR", + "question_id": "OXUKR", + "video_file": "Charades/videos/OXUKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts eating' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.4> to <21.1>." + } + ] + }, + { + "video_id": "OXUKR", + "question_id": "OXUKR", + "video_file": "Charades/videos/OXUKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.4> to <21.1>." + } + ] + }, + { + "video_id": "OXUKR", + "question_id": "OXUKR", + "video_file": "Charades/videos/OXUKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person watching television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.5>." + } + ] + }, + { + "video_id": "ZSHV8", + "question_id": "ZSHV8", + "video_file": "Charades/videos/ZSHV8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is putting dishes on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.5> to <31.2>." + } + ] + }, + { + "video_id": "ZSHV8", + "question_id": "ZSHV8", + "video_file": "Charades/videos/ZSHV8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person walks through a doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <29.6> to <37.0>." + } + ] + }, + { + "video_id": "IXY95", + "question_id": "IXY95", + "video_file": "Charades/videos/IXY95.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <18.0>." + } + ] + }, + { + "video_id": "IXY95", + "question_id": "IXY95", + "video_file": "Charades/videos/IXY95.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is looking at a picture' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <10.0>." + } + ] + }, + { + "video_id": "1X765", + "question_id": "1X765", + "video_file": "Charades/videos/1X765.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person removes clothes from a washing machine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.4> to <38.0>." + } + ] + }, + { + "video_id": "1X765", + "question_id": "1X765", + "video_file": "Charades/videos/1X765.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "JLQLC", + "question_id": "JLQLC", + "video_file": "Charades/videos/JLQLC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a picture of the wall' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.3>." + } + ] + }, + { + "video_id": "JLQLC", + "question_id": "JLQLC", + "video_file": "Charades/videos/JLQLC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they take a picture' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.3>." + } + ] + }, + { + "video_id": "DHFA6", + "question_id": "DHFA6", + "video_file": "Charades/videos/DHFA6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person begins snuggling up with a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.3> to <38.0>." + } + ] + }, + { + "video_id": "DHFA6", + "question_id": "DHFA6", + "video_file": "Charades/videos/DHFA6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person sits on the sofa' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.8> to <39.2>." + } + ] + }, + { + "video_id": "DHFA6", + "question_id": "DHFA6", + "video_file": "Charades/videos/DHFA6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person putting their feet up on a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <22.6>." + } + ] + }, + { + "video_id": "URG8B", + "question_id": "URG8B", + "video_file": "Charades/videos/URG8B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.3> to <5.8>." + } + ] + }, + { + "video_id": "URG8B", + "question_id": "URG8B", + "video_file": "Charades/videos/URG8B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats a cookie' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.3> to <9.1>." + } + ] + }, + { + "video_id": "URG8B", + "question_id": "URG8B", + "video_file": "Charades/videos/URG8B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a snack' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <9.1>." + } + ] + }, + { + "video_id": "0CGMQ", + "question_id": "0CGMQ", + "video_file": "Charades/videos/0CGMQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person so they stand up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.5> to <27.9>." + } + ] + }, + { + "video_id": "T35WB", + "question_id": "T35WB", + "video_file": "Charades/videos/T35WB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they begin laughing at the movement of the food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.4> to <23.5>." + } + ] + }, + { + "video_id": "48IQL", + "question_id": "48IQL", + "video_file": "Charades/videos/48IQL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <13.4>." + } + ] + }, + { + "video_id": "48IQL", + "question_id": "48IQL", + "video_file": "Charades/videos/48IQL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they take a cup of coffee from the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <22.7>." + } + ] + }, + { + "video_id": "48IQL", + "question_id": "48IQL", + "video_file": "Charades/videos/48IQL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person walks to a chair to sit' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.5> to <30.1>." + } + ] + }, + { + "video_id": "3KZF7", + "question_id": "3KZF7", + "video_file": "Charades/videos/3KZF7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person dresses by putting on a jacket' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.6> to <24.7>." + } + ] + }, + { + "video_id": "3KZF7", + "question_id": "3KZF7", + "video_file": "Charades/videos/3KZF7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person throws some clothes'." + }, + { + "from": "gpt", + "value": "From <4.8> to <9.9>." + } + ] + }, + { + "video_id": "DHR5F", + "question_id": "DHR5F", + "video_file": "Charades/videos/DHR5F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "DHR5F", + "question_id": "DHR5F", + "video_file": "Charades/videos/DHR5F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes the phone in hand' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <16.9>." + } + ] + }, + { + "video_id": "DHR5F", + "question_id": "DHR5F", + "video_file": "Charades/videos/DHR5F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.3>." + } + ] + }, + { + "video_id": "DHR5F", + "question_id": "DHR5F", + "video_file": "Charades/videos/DHR5F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person enters a room turns a light on' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.3>." + } + ] + }, + { + "video_id": "DHR5F", + "question_id": "DHR5F", + "video_file": "Charades/videos/DHR5F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person turns on the lights' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.3>." + } + ] + }, + { + "video_id": "VP3WN", + "question_id": "VP3WN", + "video_file": "Charades/videos/VP3WN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person awakens at a desk'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "V10LX", + "question_id": "V10LX", + "video_file": "Charades/videos/V10LX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person drinks from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.4>." + } + ] + }, + { + "video_id": "V10LX", + "question_id": "V10LX", + "video_file": "Charades/videos/V10LX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts glass on floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.7> to <15.5>." + } + ] + }, + { + "video_id": "V10LX", + "question_id": "V10LX", + "video_file": "Charades/videos/V10LX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking from glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.4>." + } + ] + }, + { + "video_id": "38QA4", + "question_id": "38QA4", + "video_file": "Charades/videos/38QA4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person begins undressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.8>." + } + ] + }, + { + "video_id": "38QA4", + "question_id": "38QA4", + "video_file": "Charades/videos/38QA4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks out through a doorway carrying the dishes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.4> to <31.0>." + } + ] + }, + { + "video_id": "38QA4", + "question_id": "38QA4", + "video_file": "Charades/videos/38QA4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a glass of milk'." + }, + { + "from": "gpt", + "value": "From <21.7> to <30.2>." + } + ] + }, + { + "video_id": "38QA4", + "question_id": "38QA4", + "video_file": "Charades/videos/38QA4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walk through the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.4> to <31.0>." + } + ] + }, + { + "video_id": "38QA4", + "question_id": "38QA4", + "video_file": "Charades/videos/38QA4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing in their bathroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.8>." + } + ] + }, + { + "video_id": "TC3BF", + "question_id": "TC3BF", + "video_file": "Charades/videos/TC3BF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person laugh at something on the screen'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <15.5>." + } + ] + }, + { + "video_id": "HWL2J", + "question_id": "HWL2J", + "video_file": "Charades/videos/HWL2J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sneezing'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.4>." + } + ] + }, + { + "video_id": "9HI9D", + "question_id": "9HI9D", + "video_file": "Charades/videos/9HI9D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opens the cabinet door'." + }, + { + "from": "gpt", + "value": "From <4.8> to <15.3>." + } + ] + }, + { + "video_id": "9HI9D", + "question_id": "9HI9D", + "video_file": "Charades/videos/9HI9D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'eats some anti-acids the person has taken out'." + }, + { + "from": "gpt", + "value": "From <23.9> to <30.4>." + } + ] + }, + { + "video_id": "9HI9D", + "question_id": "9HI9D", + "video_file": "Charades/videos/9HI9D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting some shoes on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.9> to <7.6>." + } + ] + }, + { + "video_id": "9HI9D", + "question_id": "9HI9D", + "video_file": "Charades/videos/9HI9D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is holding shoes that they have taken off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "PRQSR", + "question_id": "PRQSR", + "video_file": "Charades/videos/PRQSR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.4> to <27.2>." + } + ] + }, + { + "video_id": "PRQSR", + "question_id": "PRQSR", + "video_file": "Charades/videos/PRQSR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes their shoes off' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.6> to <9.7>." + } + ] + }, + { + "video_id": "PRQSR", + "question_id": "PRQSR", + "video_file": "Charades/videos/PRQSR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person guy walking takes off shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <9.7>." + } + ] + }, + { + "video_id": "GD9BQ", + "question_id": "GD9BQ", + "video_file": "Charades/videos/GD9BQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs in' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <28.7>." + } + ] + }, + { + "video_id": "GD9BQ", + "question_id": "GD9BQ", + "video_file": "Charades/videos/GD9BQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs out the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.7> to <28.7>." + } + ] + }, + { + "video_id": "3OQ8M", + "question_id": "3OQ8M", + "video_file": "Charades/videos/3OQ8M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws the clothes on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <11.1>." + } + ] + }, + { + "video_id": "3OQ8M", + "question_id": "3OQ8M", + "video_file": "Charades/videos/3OQ8M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person was throwing the clothes'." + }, + { + "from": "gpt", + "value": "From <2.5> to <11.1>." + } + ] + }, + { + "video_id": "VX7P0", + "question_id": "VX7P0", + "video_file": "Charades/videos/VX7P0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they are snuggling with a pillow against their chest' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.0>." + } + ] + }, + { + "video_id": "VX7P0", + "question_id": "VX7P0", + "video_file": "Charades/videos/VX7P0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person snuggling with a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.0>." + } + ] + }, + { + "video_id": "GX4B1", + "question_id": "GX4B1", + "video_file": "Charades/videos/GX4B1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person throws pillows' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.9> to <16.4>." + } + ] + }, + { + "video_id": "170OQ", + "question_id": "170OQ", + "video_file": "Charades/videos/170OQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person he puts the sandwich on the desk'." + }, + { + "from": "gpt", + "value": "From <12.1> to <18.4>." + } + ] + }, + { + "video_id": "P3XT7", + "question_id": "P3XT7", + "video_file": "Charades/videos/P3XT7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the towel on a bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.8> to <34.0>." + } + ] + }, + { + "video_id": "P3XT7", + "question_id": "P3XT7", + "video_file": "Charades/videos/P3XT7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes a drink from a water bottle' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <18.5>." + } + ] + }, + { + "video_id": "X0JKJ", + "question_id": "X0JKJ", + "video_file": "Charades/videos/X0JKJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person picks up a camera to take a picture'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.8> to <34.0>." + } + ] + }, + { + "video_id": "X0JKJ", + "question_id": "X0JKJ", + "video_file": "Charades/videos/X0JKJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a picture' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.8> to <34.0>." + } + ] + }, + { + "video_id": "XAX61", + "question_id": "XAX61", + "video_file": "Charades/videos/XAX61.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is opening bags of groceries'." + }, + { + "from": "gpt", + "value": "From <5.5> to <10.5>." + } + ] + }, + { + "video_id": "XFS7Z", + "question_id": "XFS7Z", + "video_file": "Charades/videos/XFS7Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts them on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.9> to <41.6>." + } + ] + }, + { + "video_id": "XFS7Z", + "question_id": "XFS7Z", + "video_file": "Charades/videos/XFS7Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes some medicine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <37.1> to <46.7>." + } + ] + }, + { + "video_id": "IJ587", + "question_id": "IJ587", + "video_file": "Charades/videos/IJ587.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person taking a few drinks from a cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.5> to <29.0>." + } + ] + }, + { + "video_id": "C41G7", + "question_id": "C41G7", + "video_file": "Charades/videos/C41G7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is tidying the hallway with a broom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.4> to <29.2>." + } + ] + }, + { + "video_id": "KBPI3", + "question_id": "KBPI3", + "video_file": "Charades/videos/KBPI3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person is putting blanket on desk' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.1> to <10.7>." + } + ] + }, + { + "video_id": "6YZOL", + "question_id": "6YZOL", + "video_file": "Charades/videos/6YZOL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closing a doorway'." + }, + { + "from": "gpt", + "value": "From <20.4> to <29.1>." + } + ] + }, + { + "video_id": "6YZOL", + "question_id": "6YZOL", + "video_file": "Charades/videos/6YZOL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.4> to <29.1>." + } + ] + }, + { + "video_id": "ESM8H", + "question_id": "ESM8H", + "video_file": "Charades/videos/ESM8H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws his shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <22.0>." + } + ] + }, + { + "video_id": "ESM8H", + "question_id": "ESM8H", + "video_file": "Charades/videos/ESM8H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws a pair of shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.8> to <22.0>." + } + ] + }, + { + "video_id": "7E6AB", + "question_id": "7E6AB", + "video_file": "Charades/videos/7E6AB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the refrigerator door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.5> to <17.0>." + } + ] + }, + { + "video_id": "7E6AB", + "question_id": "7E6AB", + "video_file": "Charades/videos/7E6AB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closed the refrigerator door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.5> to <17.0>." + } + ] + }, + { + "video_id": "1W28T", + "question_id": "1W28T", + "video_file": "Charades/videos/1W28T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person dresses by putting a shirt on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <12.8>." + } + ] + }, + { + "video_id": "1W28T", + "question_id": "1W28T", + "video_file": "Charades/videos/1W28T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sneezing they begin to self groom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.8> to <14.2>." + } + ] + }, + { + "video_id": "1W28T", + "question_id": "1W28T", + "video_file": "Charades/videos/1W28T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is dressing in front of a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.5> to <12.8>." + } + ] + }, + { + "video_id": "0RJKT", + "question_id": "0RJKT", + "video_file": "Charades/videos/0RJKT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding their phone'." + }, + { + "from": "gpt", + "value": "From <15.4> to <28.0>." + } + ] + }, + { + "video_id": "AZTTC", + "question_id": "AZTTC", + "video_file": "Charades/videos/AZTTC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put the book in a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.7> to <31.0>." + } + ] + }, + { + "video_id": "AZTTC", + "question_id": "AZTTC", + "video_file": "Charades/videos/AZTTC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the book down on a desk' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.7> to <31.0>." + } + ] + }, + { + "video_id": "AZTTC", + "question_id": "AZTTC", + "video_file": "Charades/videos/AZTTC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they start closing the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <22.0>." + } + ] + }, + { + "video_id": "AZTTC", + "question_id": "AZTTC", + "video_file": "Charades/videos/AZTTC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person someone holding a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <14.5>." + } + ] + }, + { + "video_id": "AZTTC", + "question_id": "AZTTC", + "video_file": "Charades/videos/AZTTC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a woman holding a book walks into a room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <14.5>." + } + ] + }, + { + "video_id": "AZTTC", + "question_id": "AZTTC", + "video_file": "Charades/videos/AZTTC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing a door,' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <22.0>." + } + ] + }, + { + "video_id": "AZTTC", + "question_id": "AZTTC", + "video_file": "Charades/videos/AZTTC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door behind her' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.8> to <22.0>." + } + ] + }, + { + "video_id": "ISHJ4", + "question_id": "ISHJ4", + "video_file": "Charades/videos/ISHJ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <16.8>." + } + ] + }, + { + "video_id": "ISHJ4", + "question_id": "ISHJ4", + "video_file": "Charades/videos/ISHJ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person closes a door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "QI3JB", + "question_id": "QI3JB", + "video_file": "Charades/videos/QI3JB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person is laughing at the tv'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.9> to <19.7>." + } + ] + }, + { + "video_id": "MC50M", + "question_id": "MC50M", + "video_file": "Charades/videos/MC50M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person puts clothes on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.7> to <9.2>." + } + ] + }, + { + "video_id": "URWJL", + "question_id": "URWJL", + "video_file": "Charades/videos/URWJL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens the door to the bedroom'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "URWJL", + "question_id": "URWJL", + "video_file": "Charades/videos/URWJL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits in bed for a few seconds'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.5> to <14.2>." + } + ] + }, + { + "video_id": "URWJL", + "question_id": "URWJL", + "video_file": "Charades/videos/URWJL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits on the bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.5> to <14.2>." + } + ] + }, + { + "video_id": "I4K7B", + "question_id": "I4K7B", + "video_file": "Charades/videos/I4K7B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put the pillow on the shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.4> to <18.9>." + } + ] + }, + { + "video_id": "I4K7B", + "question_id": "I4K7B", + "video_file": "Charades/videos/I4K7B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts a folded blanket back on the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.4> to <18.9>." + } + ] + }, + { + "video_id": "N9TRF", + "question_id": "N9TRF", + "video_file": "Charades/videos/N9TRF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person fixing their hair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.0> to <23.8>." + } + ] + }, + { + "video_id": "N9TRF", + "question_id": "N9TRF", + "video_file": "Charades/videos/N9TRF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put it on top of the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.1> to <18.5>." + } + ] + }, + { + "video_id": "5JIGM", + "question_id": "5JIGM", + "video_file": "Charades/videos/5JIGM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is in a kitchen washing dishes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <16.4>." + } + ] + }, + { + "video_id": "5JIGM", + "question_id": "5JIGM", + "video_file": "Charades/videos/5JIGM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person washes a glass out at the sink'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.1> to <15.9>." + } + ] + }, + { + "video_id": "Y6419", + "question_id": "Y6419", + "video_file": "Charades/videos/Y6419.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person in the kitchen is pretending to wash dishes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.7> to <28.9>." + } + ] + }, + { + "video_id": "4HGD7", + "question_id": "4HGD7", + "video_file": "Charades/videos/4HGD7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person runs in with a camera' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.2> to <26.5>." + } + ] + }, + { + "video_id": "4HGD7", + "question_id": "4HGD7", + "video_file": "Charades/videos/4HGD7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person runs in holding shoes'." + }, + { + "from": "gpt", + "value": "From <19.2> to <26.5>." + } + ] + }, + { + "video_id": "4HGD7", + "question_id": "4HGD7", + "video_file": "Charades/videos/4HGD7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'some other person comes running into the bathroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.2> to <26.5>." + } + ] + }, + { + "video_id": "DLBRW", + "question_id": "DLBRW", + "video_file": "Charades/videos/DLBRW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walks out of the room closing the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.8> to <32.8>." + } + ] + }, + { + "video_id": "DLBRW", + "question_id": "DLBRW", + "video_file": "Charades/videos/DLBRW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closing the door'." + }, + { + "from": "gpt", + "value": "From <24.8> to <32.8>." + } + ] + }, + { + "video_id": "QO6RV", + "question_id": "QO6RV", + "video_file": "Charades/videos/QO6RV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks through a doorway eating food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "QO6RV", + "question_id": "QO6RV", + "video_file": "Charades/videos/QO6RV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pours a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.3> to <23.0>." + } + ] + }, + { + "video_id": "QO6RV", + "question_id": "QO6RV", + "video_file": "Charades/videos/QO6RV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turning off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.2> to <30.0>." + } + ] + }, + { + "video_id": "QO6RV", + "question_id": "QO6RV", + "video_file": "Charades/videos/QO6RV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.2> to <30.0>." + } + ] + }, + { + "video_id": "QO6RV", + "question_id": "QO6RV", + "video_file": "Charades/videos/QO6RV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <12.9>." + } + ] + }, + { + "video_id": "4WUNC", + "question_id": "4WUNC", + "video_file": "Charades/videos/4WUNC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.0> to <39.0>." + } + ] + }, + { + "video_id": "4WUNC", + "question_id": "4WUNC", + "video_file": "Charades/videos/4WUNC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opened the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "4WUNC", + "question_id": "4WUNC", + "video_file": "Charades/videos/4WUNC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person x opened the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "4WUNC", + "question_id": "4WUNC", + "video_file": "Charades/videos/4WUNC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating their lunch' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.0> to <39.0>." + } + ] + }, + { + "video_id": "4WUNC", + "question_id": "4WUNC", + "video_file": "Charades/videos/4WUNC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating the sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.9> to <39.0>." + } + ] + }, + { + "video_id": "ZX1R7", + "question_id": "ZX1R7", + "video_file": "Charades/videos/ZX1R7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person takes a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.6> to <20.0>." + } + ] + }, + { + "video_id": "CVPQR", + "question_id": "CVPQR", + "video_file": "Charades/videos/CVPQR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is in a basement opening their laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "44Y6S", + "question_id": "44Y6S", + "video_file": "Charades/videos/44Y6S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a camera out of a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.3> to <15.2>." + } + ] + }, + { + "video_id": "44Y6S", + "question_id": "44Y6S", + "video_file": "Charades/videos/44Y6S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the camera on the desk' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <35.4> to <41.3>." + } + ] + }, + { + "video_id": "UCFCB", + "question_id": "UCFCB", + "video_file": "Charades/videos/UCFCB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts laptop away in bag'." + }, + { + "from": "gpt", + "value": "From <27.2> to <33.0>." + } + ] + }, + { + "video_id": "UCFCB", + "question_id": "UCFCB", + "video_file": "Charades/videos/UCFCB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.2> to <26.7>." + } + ] + }, + { + "video_id": "UCFCB", + "question_id": "UCFCB", + "video_file": "Charades/videos/UCFCB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts a laptop in a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.2> to <33.0>." + } + ] + }, + { + "video_id": "K21RO", + "question_id": "K21RO", + "video_file": "Charades/videos/K21RO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the book back on the shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.2> to <29.1>." + } + ] + }, + { + "video_id": "K21RO", + "question_id": "K21RO", + "video_file": "Charades/videos/K21RO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person looked at the book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.1> to <31.0>." + } + ] + }, + { + "video_id": "K21RO", + "question_id": "K21RO", + "video_file": "Charades/videos/K21RO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closed the cupboard door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.1> to <31.0>." + } + ] + }, + { + "video_id": "GRJG1", + "question_id": "GRJG1", + "video_file": "Charades/videos/GRJG1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts undressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.5> to <17.6>." + } + ] + }, + { + "video_id": "CCCUJ", + "question_id": "CCCUJ", + "video_file": "Charades/videos/CCCUJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person close the door'." + }, + { + "from": "gpt", + "value": "From <11.6> to <18.9>." + } + ] + }, + { + "video_id": "CCCUJ", + "question_id": "CCCUJ", + "video_file": "Charades/videos/CCCUJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put their laptop down off to the side' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.6> to <32.2>." + } + ] + }, + { + "video_id": "CCCUJ", + "question_id": "CCCUJ", + "video_file": "Charades/videos/CCCUJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turned on a light'." + }, + { + "from": "gpt", + "value": "From <8.6> to <13.5>." + } + ] + }, + { + "video_id": "CCCUJ", + "question_id": "CCCUJ", + "video_file": "Charades/videos/CCCUJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they turn on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.6> to <13.5>." + } + ] + }, + { + "video_id": "29C6X", + "question_id": "29C6X", + "video_file": "Charades/videos/29C6X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throw it on the floor'." + }, + { + "from": "gpt", + "value": "From <13.6> to <23.0>." + } + ] + }, + { + "video_id": "LKSBL", + "question_id": "LKSBL", + "video_file": "Charades/videos/LKSBL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a man opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "LKSBL", + "question_id": "LKSBL", + "video_file": "Charades/videos/LKSBL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.0> to <9.7>." + } + ] + }, + { + "video_id": "LKSBL", + "question_id": "LKSBL", + "video_file": "Charades/videos/LKSBL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "LKSBL", + "question_id": "LKSBL", + "video_file": "Charades/videos/LKSBL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens another cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "3GY40", + "question_id": "3GY40", + "video_file": "Charades/videos/3GY40.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person is throwing a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.4> to <15.7>." + } + ] + }, + { + "video_id": "3GY40", + "question_id": "3GY40", + "video_file": "Charades/videos/3GY40.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sneezing on dishes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <11.0>." + } + ] + }, + { + "video_id": "3GY40", + "question_id": "3GY40", + "video_file": "Charades/videos/3GY40.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a second person throws a pillow at the first person' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <15.7>." + } + ] + }, + { + "video_id": "3GY40", + "question_id": "3GY40", + "video_file": "Charades/videos/3GY40.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting at a table with some food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.0> to <21.0>." + } + ] + }, + { + "video_id": "UR7C8", + "question_id": "UR7C8", + "video_file": "Charades/videos/UR7C8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes their shoes off' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.2> to <13.8>." + } + ] + }, + { + "video_id": "UR7C8", + "question_id": "UR7C8", + "video_file": "Charades/videos/UR7C8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.1> to <8.4>." + } + ] + }, + { + "video_id": "UR7C8", + "question_id": "UR7C8", + "video_file": "Charades/videos/UR7C8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door in the hallway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <8.4>." + } + ] + }, + { + "video_id": "8Y4YD", + "question_id": "8Y4YD", + "video_file": "Charades/videos/8Y4YD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person holds a camera'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <32.0>." + } + ] + }, + { + "video_id": "1TWH6", + "question_id": "1TWH6", + "video_file": "Charades/videos/1TWH6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes a box'." + }, + { + "from": "gpt", + "value": "From <5.0> to <10.0>." + } + ] + }, + { + "video_id": "1TWH6", + "question_id": "1TWH6", + "video_file": "Charades/videos/1TWH6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes something out of a box'." + }, + { + "from": "gpt", + "value": "From <5.0> to <10.0>." + } + ] + }, + { + "video_id": "1TWH6", + "question_id": "1TWH6", + "video_file": "Charades/videos/1TWH6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person look out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <31.0>." + } + ] + }, + { + "video_id": "1TWH6", + "question_id": "1TWH6", + "video_file": "Charades/videos/1TWH6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes a mug out of a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.0> to <10.0>." + } + ] + }, + { + "video_id": "1TWH6", + "question_id": "1TWH6", + "video_file": "Charades/videos/1TWH6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looks out the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <31.0>." + } + ] + }, + { + "video_id": "28D7L", + "question_id": "28D7L", + "video_file": "Charades/videos/28D7L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts the coffee on the shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.0> to <22.0>." + } + ] + }, + { + "video_id": "28D7L", + "question_id": "28D7L", + "video_file": "Charades/videos/28D7L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the cup onto a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <22.0>." + } + ] + }, + { + "video_id": "4X2JC", + "question_id": "4X2JC", + "video_file": "Charades/videos/4X2JC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person begins eating the food in the bowl' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <32.0>." + } + ] + }, + { + "video_id": "4X2JC", + "question_id": "4X2JC", + "video_file": "Charades/videos/4X2JC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person start eating it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <32.0>." + } + ] + }, + { + "video_id": "129SP", + "question_id": "129SP", + "video_file": "Charades/videos/129SP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <27.5> to <32.0>." + } + ] + }, + { + "video_id": "129SP", + "question_id": "129SP", + "video_file": "Charades/videos/129SP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'as the person was closing the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.5> to <32.0>." + } + ] + }, + { + "video_id": "5R83A", + "question_id": "5R83A", + "video_file": "Charades/videos/5R83A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts the towel on a stair'." + }, + { + "from": "gpt", + "value": "From <5.7> to <12.9>." + } + ] + }, + { + "video_id": "5R83A", + "question_id": "5R83A", + "video_file": "Charades/videos/5R83A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is standing by the stairs holding a towel' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.0>." + } + ] + }, + { + "video_id": "5R83A", + "question_id": "5R83A", + "video_file": "Charades/videos/5R83A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person a broom they put the towel on the stairs' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <12.9>." + } + ] + }, + { + "video_id": "ZWY3E", + "question_id": "ZWY3E", + "video_file": "Charades/videos/ZWY3E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sitting on their bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "ZWY3E", + "question_id": "ZWY3E", + "video_file": "Charades/videos/ZWY3E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person puts on a pair of shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "ZWY3E", + "question_id": "ZWY3E", + "video_file": "Charades/videos/ZWY3E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is putting on their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "AYXFY", + "question_id": "AYXFY", + "video_file": "Charades/videos/AYXFY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating cookie' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.8> to <22.1>." + } + ] + }, + { + "video_id": "AYXFY", + "question_id": "AYXFY", + "video_file": "Charades/videos/AYXFY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating something'." + }, + { + "from": "gpt", + "value": "From <15.8> to <22.1>." + } + ] + }, + { + "video_id": "STDCJ", + "question_id": "STDCJ", + "video_file": "Charades/videos/STDCJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.9> to <15.1>." + } + ] + }, + { + "video_id": "STDCJ", + "question_id": "STDCJ", + "video_file": "Charades/videos/STDCJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person watching a video that is playing on their laptop'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "STDCJ", + "question_id": "STDCJ", + "video_file": "Charades/videos/STDCJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person watching a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "L57L2", + "question_id": "L57L2", + "video_file": "Charades/videos/L57L2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting it in a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.7> to <26.2>." + } + ] + }, + { + "video_id": "L57L2", + "question_id": "L57L2", + "video_file": "Charades/videos/L57L2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they close the box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.8> to <31.8>." + } + ] + }, + { + "video_id": "L57L2", + "question_id": "L57L2", + "video_file": "Charades/videos/L57L2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.8> to <31.8>." + } + ] + }, + { + "video_id": "IQM7A", + "question_id": "IQM7A", + "video_file": "Charades/videos/IQM7A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'there is a person washing their hands in a sink'." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.6>." + } + ] + }, + { + "video_id": "IQM7A", + "question_id": "IQM7A", + "video_file": "Charades/videos/IQM7A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is standing in their bathroom washing their hands'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.6>." + } + ] + }, + { + "video_id": "IQM7A", + "question_id": "IQM7A", + "video_file": "Charades/videos/IQM7A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person washes their hands in the sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.6>." + } + ] + }, + { + "video_id": "X7R9N", + "question_id": "X7R9N", + "video_file": "Charades/videos/X7R9N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <12.7>." + } + ] + }, + { + "video_id": "8ZF5S", + "question_id": "8ZF5S", + "video_file": "Charades/videos/8ZF5S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person starts washing a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "YX3W8", + "question_id": "YX3W8", + "video_file": "Charades/videos/YX3W8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is throwing clothes on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <7.6>." + } + ] + }, + { + "video_id": "SQYTO", + "question_id": "SQYTO", + "video_file": "Charades/videos/SQYTO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns off the lights'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <4.6>." + } + ] + }, + { + "video_id": "SQYTO", + "question_id": "SQYTO", + "video_file": "Charades/videos/SQYTO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person watching television' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "SQYTO", + "question_id": "SQYTO", + "video_file": "Charades/videos/SQYTO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turning a light off as they exit the room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <4.6>." + } + ] + }, + { + "video_id": "SQYTO", + "question_id": "SQYTO", + "video_file": "Charades/videos/SQYTO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turning off a light on the way' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.5> to <4.6>." + } + ] + }, + { + "video_id": "SQYTO", + "question_id": "SQYTO", + "video_file": "Charades/videos/SQYTO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person sits on a sofa watching television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "SQYTO", + "question_id": "SQYTO", + "video_file": "Charades/videos/SQYTO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person watching television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "8V9IX", + "question_id": "8V9IX", + "video_file": "Charades/videos/8V9IX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put the dishes away in a cupboard' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <8.7>." + } + ] + }, + { + "video_id": "N8WK8", + "question_id": "N8WK8", + "video_file": "Charades/videos/N8WK8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they close the book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.4> to <22.0>." + } + ] + }, + { + "video_id": "AVA41", + "question_id": "AVA41", + "video_file": "Charades/videos/AVA41.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person closes the bathroom door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.5> to <9.3>." + } + ] + }, + { + "video_id": "AVA41", + "question_id": "AVA41", + "video_file": "Charades/videos/AVA41.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <1.5> to <9.3>." + } + ] + }, + { + "video_id": "AVA41", + "question_id": "AVA41", + "video_file": "Charades/videos/AVA41.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open the medicine cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.3> to <21.8>." + } + ] + }, + { + "video_id": "AVA41", + "question_id": "AVA41", + "video_file": "Charades/videos/AVA41.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes a bottle of medicine from the cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.3> to <33.0>." + } + ] + }, + { + "video_id": "STYTN", + "question_id": "STYTN", + "video_file": "Charades/videos/STYTN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sit on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.6> to <24.3>." + } + ] + }, + { + "video_id": "STYTN", + "question_id": "STYTN", + "video_file": "Charades/videos/STYTN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits on the floor'." + }, + { + "from": "gpt", + "value": "From <15.6> to <24.3>." + } + ] + }, + { + "video_id": "STYTN", + "question_id": "STYTN", + "video_file": "Charades/videos/STYTN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put the bag back in the box'." + }, + { + "from": "gpt", + "value": "From <10.0> to <16.3>." + } + ] + }, + { + "video_id": "STYTN", + "question_id": "STYTN", + "video_file": "Charades/videos/STYTN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sitting on the floor of a pantry'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.6> to <24.3>." + } + ] + }, + { + "video_id": "9KGIR", + "question_id": "9KGIR", + "video_file": "Charades/videos/9KGIR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.8> to <6.6>." + } + ] + }, + { + "video_id": "9KGIR", + "question_id": "9KGIR", + "video_file": "Charades/videos/9KGIR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a cabinet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.8> to <6.6>." + } + ] + }, + { + "video_id": "R5L98", + "question_id": "R5L98", + "video_file": "Charades/videos/R5L98.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.9> to <26.8>." + } + ] + }, + { + "video_id": "6JGXL", + "question_id": "6JGXL", + "video_file": "Charades/videos/6JGXL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take some medicine' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.0> to <20.9>." + } + ] + }, + { + "video_id": "Q3UAN", + "question_id": "Q3UAN", + "video_file": "Charades/videos/Q3UAN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is washing dishes in the sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.0>." + } + ] + }, + { + "video_id": "Q3UAN", + "question_id": "Q3UAN", + "video_file": "Charades/videos/Q3UAN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is washing dishes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.0>." + } + ] + }, + { + "video_id": "Q3UAN", + "question_id": "Q3UAN", + "video_file": "Charades/videos/Q3UAN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pours some of that into a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.2> to <32.0>." + } + ] + }, + { + "video_id": "463UL", + "question_id": "463UL", + "video_file": "Charades/videos/463UL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.9>." + } + ] + }, + { + "video_id": "AWITI", + "question_id": "AWITI", + "video_file": "Charades/videos/AWITI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take pillows off the bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <15.9>." + } + ] + }, + { + "video_id": "AWITI", + "question_id": "AWITI", + "video_file": "Charades/videos/AWITI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put them on a shelf'." + }, + { + "from": "gpt", + "value": "From <23.4> to <29.5>." + } + ] + }, + { + "video_id": "AWITI", + "question_id": "AWITI", + "video_file": "Charades/videos/AWITI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turn o the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.4>." + } + ] + }, + { + "video_id": "AWITI", + "question_id": "AWITI", + "video_file": "Charades/videos/AWITI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns on the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.4>." + } + ] + }, + { + "video_id": "1O6PH", + "question_id": "1O6PH", + "video_file": "Charades/videos/1O6PH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person he walks over to the closet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.8> to <22.4>." + } + ] + }, + { + "video_id": "1O6PH", + "question_id": "1O6PH", + "video_file": "Charades/videos/1O6PH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they close the closet door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.8> to <22.4>." + } + ] + }, + { + "video_id": "1O6PH", + "question_id": "1O6PH", + "video_file": "Charades/videos/1O6PH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is seen putting a pillow on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.9> to <14.5>." + } + ] + }, + { + "video_id": "1O6PH", + "question_id": "1O6PH", + "video_file": "Charades/videos/1O6PH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is holding a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "XBRO4", + "question_id": "XBRO4", + "video_file": "Charades/videos/XBRO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs out of the study with a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.7> to <30.1>." + } + ] + }, + { + "video_id": "XBRO4", + "question_id": "XBRO4", + "video_file": "Charades/videos/XBRO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person run out of the room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.7> to <30.1>." + } + ] + }, + { + "video_id": "7QWL3", + "question_id": "7QWL3", + "video_file": "Charades/videos/7QWL3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looks out the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.4> to <33.0>." + } + ] + }, + { + "video_id": "7QWL3", + "question_id": "7QWL3", + "video_file": "Charades/videos/7QWL3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is standing in the closet reading a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.1> to <20.4>." + } + ] + }, + { + "video_id": "7QWL3", + "question_id": "7QWL3", + "video_file": "Charades/videos/7QWL3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they close the book'." + }, + { + "from": "gpt", + "value": "From <15.1> to <20.4>." + } + ] + }, + { + "video_id": "7QWL3", + "question_id": "7QWL3", + "video_file": "Charades/videos/7QWL3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person closes the book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.1> to <20.4>." + } + ] + }, + { + "video_id": "7QWL3", + "question_id": "7QWL3", + "video_file": "Charades/videos/7QWL3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person look out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.4> to <33.0>." + } + ] + }, + { + "video_id": "Q57HC", + "question_id": "Q57HC", + "video_file": "Charades/videos/Q57HC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person open a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.0> to <39.2>." + } + ] + }, + { + "video_id": "Q57HC", + "question_id": "Q57HC", + "video_file": "Charades/videos/Q57HC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they open a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.0> to <39.2>." + } + ] + }, + { + "video_id": "79YH2", + "question_id": "79YH2", + "video_file": "Charades/videos/79YH2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.8> to <32.0>." + } + ] + }, + { + "video_id": "G3OUM", + "question_id": "G3OUM", + "video_file": "Charades/videos/G3OUM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws it on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.8> to <27.8>." + } + ] + }, + { + "video_id": "G3OUM", + "question_id": "G3OUM", + "video_file": "Charades/videos/G3OUM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is cooking food on the stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.4> to <17.7>." + } + ] + }, + { + "video_id": "G3OUM", + "question_id": "G3OUM", + "video_file": "Charades/videos/G3OUM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is cooking on the stove'." + }, + { + "from": "gpt", + "value": "From <3.4> to <17.7>." + } + ] + }, + { + "video_id": "G3OUM", + "question_id": "G3OUM", + "video_file": "Charades/videos/G3OUM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a broom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <23.1>." + } + ] + }, + { + "video_id": "P2UBC", + "question_id": "P2UBC", + "video_file": "Charades/videos/P2UBC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begin sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <13.7>." + } + ] + }, + { + "video_id": "P2UBC", + "question_id": "P2UBC", + "video_file": "Charades/videos/P2UBC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turn off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.8> to <33.5>." + } + ] + }, + { + "video_id": "P2UBC", + "question_id": "P2UBC", + "video_file": "Charades/videos/P2UBC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person proceed to put on a pair of shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.5> to <13.0>." + } + ] + }, + { + "video_id": "P2UBC", + "question_id": "P2UBC", + "video_file": "Charades/videos/P2UBC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put on their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <13.0>." + } + ] + }, + { + "video_id": "P2UBC", + "question_id": "P2UBC", + "video_file": "Charades/videos/P2UBC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turning off a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.8> to <33.5>." + } + ] + }, + { + "video_id": "GN4SJ", + "question_id": "GN4SJ", + "video_file": "Charades/videos/GN4SJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats a piece of food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "GN4SJ", + "question_id": "GN4SJ", + "video_file": "Charades/videos/GN4SJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats something'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "GN4SJ", + "question_id": "GN4SJ", + "video_file": "Charades/videos/GN4SJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the bag in a box'." + }, + { + "from": "gpt", + "value": "From <5.4> to <21.5>." + } + ] + }, + { + "video_id": "GN4SJ", + "question_id": "GN4SJ", + "video_file": "Charades/videos/GN4SJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person put their hand in the bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.4> to <21.5>." + } + ] + }, + { + "video_id": "GN4SJ", + "question_id": "GN4SJ", + "video_file": "Charades/videos/GN4SJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person eats from a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "GN4SJ", + "question_id": "GN4SJ", + "video_file": "Charades/videos/GN4SJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is holding a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.4> to <8.2>." + } + ] + }, + { + "video_id": "6MUM6", + "question_id": "6MUM6", + "video_file": "Charades/videos/6MUM6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person pours some coffee into the cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <12.0>." + } + ] + }, + { + "video_id": "D19IR", + "question_id": "D19IR", + "video_file": "Charades/videos/D19IR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person stand up with a bag of groceries' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <18.1>." + } + ] + }, + { + "video_id": "YQSN8", + "question_id": "YQSN8", + "video_file": "Charades/videos/YQSN8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person is closing a window'." + }, + { + "from": "gpt", + "value": "From <3.4> to <13.8>." + } + ] + }, + { + "video_id": "YQSN8", + "question_id": "YQSN8", + "video_file": "Charades/videos/YQSN8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person closses the window'." + }, + { + "from": "gpt", + "value": "From <3.4> to <13.8>." + } + ] + }, + { + "video_id": "YQSN8", + "question_id": "YQSN8", + "video_file": "Charades/videos/YQSN8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the other person looks at the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.4> to <32.0>." + } + ] + }, + { + "video_id": "5K0KJ", + "question_id": "5K0KJ", + "video_file": "Charades/videos/5K0KJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a phone in the laundry room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <30.0>." + } + ] + }, + { + "video_id": "1DNAX", + "question_id": "1DNAX", + "video_file": "Charades/videos/1DNAX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person is walking to open closet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.4> to <6.9>." + } + ] + }, + { + "video_id": "1DNAX", + "question_id": "1DNAX", + "video_file": "Charades/videos/1DNAX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person is opening the doors to the closet'." + }, + { + "from": "gpt", + "value": "From <0.4> to <6.9>." + } + ] + }, + { + "video_id": "1DNAX", + "question_id": "1DNAX", + "video_file": "Charades/videos/1DNAX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns on the light'." + }, + { + "from": "gpt", + "value": "From <6.9> to <12.0>." + } + ] + }, + { + "video_id": "1DNAX", + "question_id": "1DNAX", + "video_file": "Charades/videos/1DNAX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.2> to <20.9>." + } + ] + }, + { + "video_id": "1DNAX", + "question_id": "1DNAX", + "video_file": "Charades/videos/1DNAX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns light on grabs broom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.9> to <12.0>." + } + ] + }, + { + "video_id": "O76N2", + "question_id": "O76N2", + "video_file": "Charades/videos/O76N2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting the bag on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.7> to <24.2>." + } + ] + }, + { + "video_id": "O76N2", + "question_id": "O76N2", + "video_file": "Charades/videos/O76N2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a bag'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "8BCI8", + "question_id": "8BCI8", + "video_file": "Charades/videos/8BCI8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is cooking food on a stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.6>." + } + ] + }, + { + "video_id": "8BCI8", + "question_id": "8BCI8", + "video_file": "Charades/videos/8BCI8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opening a refrigerator in a kitchen'." + }, + { + "from": "gpt", + "value": "From <21.6> to <29.4>." + } + ] + }, + { + "video_id": "8BCI8", + "question_id": "8BCI8", + "video_file": "Charades/videos/8BCI8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take some food from the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <28.6> to <35.9>." + } + ] + }, + { + "video_id": "36VYU", + "question_id": "36VYU", + "video_file": "Charades/videos/36VYU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person puts a cup of coffee onto the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "ZD3GI", + "question_id": "ZD3GI", + "video_file": "Charades/videos/ZD3GI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a sandwich in a living room'." + }, + { + "from": "gpt", + "value": "From <25.0> to <44.0>." + } + ] + }, + { + "video_id": "ZD3GI", + "question_id": "ZD3GI", + "video_file": "Charades/videos/ZD3GI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person pours some wine into a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.9> to <21.2>." + } + ] + }, + { + "video_id": "ZD3GI", + "question_id": "ZD3GI", + "video_file": "Charades/videos/ZD3GI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person putting the bottle on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.9> to <27.9>." + } + ] + }, + { + "video_id": "ZD3GI", + "question_id": "ZD3GI", + "video_file": "Charades/videos/ZD3GI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pouring a glass of wine'." + }, + { + "from": "gpt", + "value": "From <9.9> to <21.2>." + } + ] + }, + { + "video_id": "OB9H3", + "question_id": "OB9H3", + "video_file": "Charades/videos/OB9H3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person playing with a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <13.9>." + } + ] + }, + { + "video_id": "30LGT", + "question_id": "30LGT", + "video_file": "Charades/videos/30LGT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person sat on the floor reading a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.6>." + } + ] + }, + { + "video_id": "30LGT", + "question_id": "30LGT", + "video_file": "Charades/videos/30LGT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person reading a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.6>." + } + ] + }, + { + "video_id": "UX8NQ", + "question_id": "UX8NQ", + "video_file": "Charades/videos/UX8NQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is standing in front of a sink laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.8> to <12.8>." + } + ] + }, + { + "video_id": "UX8NQ", + "question_id": "UX8NQ", + "video_file": "Charades/videos/UX8NQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is washing dishes in a sink'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <12.9>." + } + ] + }, + { + "video_id": "UX8NQ", + "question_id": "UX8NQ", + "video_file": "Charades/videos/UX8NQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person washing dishes'." + }, + { + "from": "gpt", + "value": "From <1.0> to <12.9>." + } + ] + }, + { + "video_id": "BQAUC", + "question_id": "BQAUC", + "video_file": "Charades/videos/BQAUC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person runs off' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.1> to <15.6>." + } + ] + }, + { + "video_id": "BQAUC", + "question_id": "BQAUC", + "video_file": "Charades/videos/BQAUC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts shoes on a shelf'." + }, + { + "from": "gpt", + "value": "From <2.1> to <8.0>." + } + ] + }, + { + "video_id": "BQAUC", + "question_id": "BQAUC", + "video_file": "Charades/videos/BQAUC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts the shoes away' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <8.0>." + } + ] + }, + { + "video_id": "VLZ40", + "question_id": "VLZ40", + "video_file": "Charades/videos/VLZ40.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes off shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.3>." + } + ] + }, + { + "video_id": "S47DI", + "question_id": "S47DI", + "video_file": "Charades/videos/S47DI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they close the fridge door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.3> to <31.0>." + } + ] + }, + { + "video_id": "RH2YW", + "question_id": "RH2YW", + "video_file": "Charades/videos/RH2YW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <18.3>." + } + ] + }, + { + "video_id": "RH2YW", + "question_id": "RH2YW", + "video_file": "Charades/videos/RH2YW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <8.5>." + } + ] + }, + { + "video_id": "RH2YW", + "question_id": "RH2YW", + "video_file": "Charades/videos/RH2YW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person standing next to a table eats a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.3> to <8.5>." + } + ] + }, + { + "video_id": "ZAJAJ", + "question_id": "ZAJAJ", + "video_file": "Charades/videos/ZAJAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person snuggles against a big pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <9.1>." + } + ] + }, + { + "video_id": "ZAJAJ", + "question_id": "ZAJAJ", + "video_file": "Charades/videos/ZAJAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.5> to <10.0>." + } + ] + }, + { + "video_id": "IGOQK", + "question_id": "IGOQK", + "video_file": "Charades/videos/IGOQK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing door behind them' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "OZCGO", + "question_id": "OZCGO", + "video_file": "Charades/videos/OZCGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person throws a pillow on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <7.2>." + } + ] + }, + { + "video_id": "OVHFT", + "question_id": "OVHFT", + "video_file": "Charades/videos/OVHFT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.2> to <17.5>." + } + ] + }, + { + "video_id": "OVHFT", + "question_id": "OVHFT", + "video_file": "Charades/videos/OVHFT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns on a light'." + }, + { + "from": "gpt", + "value": "From <6.3> to <12.9>." + } + ] + }, + { + "video_id": "WYYUD", + "question_id": "WYYUD", + "video_file": "Charades/videos/WYYUD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person awakens from bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <15.1>." + } + ] + }, + { + "video_id": "WYYUD", + "question_id": "WYYUD", + "video_file": "Charades/videos/WYYUD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person gets undressed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.7> to <35.0>." + } + ] + }, + { + "video_id": "WYYUD", + "question_id": "WYYUD", + "video_file": "Charades/videos/WYYUD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is undressing in bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.7> to <35.0>." + } + ] + }, + { + "video_id": "WYYUD", + "question_id": "WYYUD", + "video_file": "Charades/videos/WYYUD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws their clothes on a rack' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.1> to <31.9>." + } + ] + }, + { + "video_id": "706BT", + "question_id": "706BT", + "video_file": "Charades/videos/706BT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is seen eating a sandwich in the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "706BT", + "question_id": "706BT", + "video_file": "Charades/videos/706BT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person stands in a doorway eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "U45LK", + "question_id": "U45LK", + "video_file": "Charades/videos/U45LK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they close the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <29.3> to <36.0>." + } + ] + }, + { + "video_id": "U45LK", + "question_id": "U45LK", + "video_file": "Charades/videos/U45LK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is putting clothes away in a wardrobe'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <8.8>." + } + ] + }, + { + "video_id": "2Q5Y2", + "question_id": "2Q5Y2", + "video_file": "Charades/videos/2Q5Y2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person holding a blanket walks into the laundry room'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.6>." + } + ] + }, + { + "video_id": "2Q5Y2", + "question_id": "2Q5Y2", + "video_file": "Charades/videos/2Q5Y2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put their blanket into the washer for washing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.9> to <15.0>." + } + ] + }, + { + "video_id": "QOWXK", + "question_id": "QOWXK", + "video_file": "Charades/videos/QOWXK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they are dressing by putting on a jacket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.8> to <34.7>." + } + ] + }, + { + "video_id": "QOWXK", + "question_id": "QOWXK", + "video_file": "Charades/videos/QOWXK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person reaches for a picture'." + }, + { + "from": "gpt", + "value": "From <31.2> to <37.0>." + } + ] + }, + { + "video_id": "RBNLA", + "question_id": "RBNLA", + "video_file": "Charades/videos/RBNLA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person close the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <28.5>." + } + ] + }, + { + "video_id": "RBNLA", + "question_id": "RBNLA", + "video_file": "Charades/videos/RBNLA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turning off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.4> to <32.0>." + } + ] + }, + { + "video_id": "D0YWV", + "question_id": "D0YWV", + "video_file": "Charades/videos/D0YWV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.7> to <13.2>." + } + ] + }, + { + "video_id": "D0YWV", + "question_id": "D0YWV", + "video_file": "Charades/videos/D0YWV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is walking through a door the closing it'." + }, + { + "from": "gpt", + "value": "From <6.7> to <13.2>." + } + ] + }, + { + "video_id": "7R8ZU", + "question_id": "7R8ZU", + "video_file": "Charades/videos/7R8ZU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person closed the laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <20.8>." + } + ] + }, + { + "video_id": "9B93K", + "question_id": "9B93K", + "video_file": "Charades/videos/9B93K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts the sandwich back' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.0> to <25.7>." + } + ] + }, + { + "video_id": "9B93K", + "question_id": "9B93K", + "video_file": "Charades/videos/9B93K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turning off the light behind them'." + }, + { + "from": "gpt", + "value": "From <27.9> to <36.6>." + } + ] + }, + { + "video_id": "9B93K", + "question_id": "9B93K", + "video_file": "Charades/videos/9B93K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the sandwich down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.0> to <25.7>." + } + ] + }, + { + "video_id": "NTYC3", + "question_id": "NTYC3", + "video_file": "Charades/videos/NTYC3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.1> to <31.6>." + } + ] + }, + { + "video_id": "NTYC3", + "question_id": "NTYC3", + "video_file": "Charades/videos/NTYC3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person started to laugh'." + }, + { + "from": "gpt", + "value": "From <31.2> to <38.0>." + } + ] + }, + { + "video_id": "NTYC3", + "question_id": "NTYC3", + "video_file": "Charades/videos/NTYC3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.1> to <31.6>." + } + ] + }, + { + "video_id": "NTYC3", + "question_id": "NTYC3", + "video_file": "Charades/videos/NTYC3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they opened their laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.1> to <31.6>." + } + ] + }, + { + "video_id": "WOOYQ", + "question_id": "WOOYQ", + "video_file": "Charades/videos/WOOYQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person holding a towel walks through the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <11.7>." + } + ] + }, + { + "video_id": "WOOYQ", + "question_id": "WOOYQ", + "video_file": "Charades/videos/WOOYQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a broom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.5> to <23.8>." + } + ] + }, + { + "video_id": "RUOOY", + "question_id": "RUOOY", + "video_file": "Charades/videos/RUOOY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take a picture with their phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.1> to <19.8>." + } + ] + }, + { + "video_id": "RUOOY", + "question_id": "RUOOY", + "video_file": "Charades/videos/RUOOY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take a picture of themselves'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.1> to <19.8>." + } + ] + }, + { + "video_id": "E9EKR", + "question_id": "E9EKR", + "video_file": "Charades/videos/E9EKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person throws their shoes onto the bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <19.7>." + } + ] + }, + { + "video_id": "E9EKR", + "question_id": "E9EKR", + "video_file": "Charades/videos/E9EKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting on a bed'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.7>." + } + ] + }, + { + "video_id": "XAOUP", + "question_id": "XAOUP", + "video_file": "Charades/videos/XAOUP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closing the door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "J3V1D", + "question_id": "J3V1D", + "video_file": "Charades/videos/J3V1D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes off their shoes'." + }, + { + "from": "gpt", + "value": "From <19.6> to <25.6>." + } + ] + }, + { + "video_id": "TSBB8", + "question_id": "TSBB8", + "video_file": "Charades/videos/TSBB8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns the light off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <12.3>." + } + ] + }, + { + "video_id": "E2Q68", + "question_id": "E2Q68", + "video_file": "Charades/videos/E2Q68.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts opening a bag full of groceries' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.6> to <13.5>." + } + ] + }, + { + "video_id": "O6V4U", + "question_id": "O6V4U", + "video_file": "Charades/videos/O6V4U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person a is in a bedroom getting dressed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "O6V4U", + "question_id": "O6V4U", + "video_file": "Charades/videos/O6V4U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is dressing themselves' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "O6V4U", + "question_id": "O6V4U", + "video_file": "Charades/videos/O6V4U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws a pillow onto a bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.0> to <30.0>." + } + ] + }, + { + "video_id": "O6V4U", + "question_id": "O6V4U", + "video_file": "Charades/videos/O6V4U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.0> to <18.2>." + } + ] + }, + { + "video_id": "OOWJ9", + "question_id": "OOWJ9", + "video_file": "Charades/videos/OOWJ9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person reading a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "OOWJ9", + "question_id": "OOWJ9", + "video_file": "Charades/videos/OOWJ9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they pour a glass of water'." + }, + { + "from": "gpt", + "value": "From <15.2> to <25.9>." + } + ] + }, + { + "video_id": "OOWJ9", + "question_id": "OOWJ9", + "video_file": "Charades/videos/OOWJ9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they read a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "OOWJ9", + "question_id": "OOWJ9", + "video_file": "Charades/videos/OOWJ9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person the pour a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <25.9>." + } + ] + }, + { + "video_id": "0MDYC", + "question_id": "0MDYC", + "video_file": "Charades/videos/0MDYC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts the clothes on a chair'." + }, + { + "from": "gpt", + "value": "From <18.3> to <24.8>." + } + ] + }, + { + "video_id": "G81JB", + "question_id": "G81JB", + "video_file": "Charades/videos/G81JB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begin to tidy up the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.6> to <30.7>." + } + ] + }, + { + "video_id": "G81JB", + "question_id": "G81JB", + "video_file": "Charades/videos/G81JB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is cooking on a stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.7> to <11.3>." + } + ] + }, + { + "video_id": "G81JB", + "question_id": "G81JB", + "video_file": "Charades/videos/G81JB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is cooking in a kitchen' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.7> to <11.3>." + } + ] + }, + { + "video_id": "1YTD7", + "question_id": "1YTD7", + "video_file": "Charades/videos/1YTD7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.4> to <30.9>." + } + ] + }, + { + "video_id": "1YTD7", + "question_id": "1YTD7", + "video_file": "Charades/videos/1YTD7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.4> to <30.9>." + } + ] + }, + { + "video_id": "CMEA6", + "question_id": "CMEA6", + "video_file": "Charades/videos/CMEA6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person closes the door in the entryway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "ZCH7N", + "question_id": "ZCH7N", + "video_file": "Charades/videos/ZCH7N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they being playing with their phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <27.9>." + } + ] + }, + { + "video_id": "ZCH7N", + "question_id": "ZCH7N", + "video_file": "Charades/videos/ZCH7N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins eating it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <15.5>." + } + ] + }, + { + "video_id": "ZCH7N", + "question_id": "ZCH7N", + "video_file": "Charades/videos/ZCH7N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take out their phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <17.9>." + } + ] + }, + { + "video_id": "ZCH7N", + "question_id": "ZCH7N", + "video_file": "Charades/videos/ZCH7N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.1> to <15.5>." + } + ] + }, + { + "video_id": "QLKSV", + "question_id": "QLKSV", + "video_file": "Charades/videos/QLKSV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person is wrapped in a blanket eating' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.2> to <28.7>." + } + ] + }, + { + "video_id": "QLKSV", + "question_id": "QLKSV", + "video_file": "Charades/videos/QLKSV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating something' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.2> to <28.7>." + } + ] + }, + { + "video_id": "FUT86", + "question_id": "FUT86", + "video_file": "Charades/videos/FUT86.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.3> to <36.1>." + } + ] + }, + { + "video_id": "FUT86", + "question_id": "FUT86", + "video_file": "Charades/videos/FUT86.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person eats from a bag of groceries' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.3> to <36.1>." + } + ] + }, + { + "video_id": "RLS2I", + "question_id": "RLS2I", + "video_file": "Charades/videos/RLS2I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.8> to <8.4>." + } + ] + }, + { + "video_id": "UNOW9", + "question_id": "UNOW9", + "video_file": "Charades/videos/UNOW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes off shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <22.8>." + } + ] + }, + { + "video_id": "UNOW9", + "question_id": "UNOW9", + "video_file": "Charades/videos/UNOW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they take off their shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <22.8>." + } + ] + }, + { + "video_id": "UNOW9", + "question_id": "UNOW9", + "video_file": "Charades/videos/UNOW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.6> to <22.8>." + } + ] + }, + { + "video_id": "MU54W", + "question_id": "MU54W", + "video_file": "Charades/videos/MU54W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person snuggling a big fluffy blanket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <27.0>." + } + ] + }, + { + "video_id": "MU54W", + "question_id": "MU54W", + "video_file": "Charades/videos/MU54W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is snuggling with a blanket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.9> to <27.0>." + } + ] + }, + { + "video_id": "MU54W", + "question_id": "MU54W", + "video_file": "Charades/videos/MU54W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is holding a vacuum' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "FDOY9", + "question_id": "FDOY9", + "video_file": "Charades/videos/FDOY9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turning off a light in an entryway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.4> to <35.7>." + } + ] + }, + { + "video_id": "FDOY9", + "question_id": "FDOY9", + "video_file": "Charades/videos/FDOY9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is pouring a cup of coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.7> to <18.7>." + } + ] + }, + { + "video_id": "FDOY9", + "question_id": "FDOY9", + "video_file": "Charades/videos/FDOY9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.4> to <35.7>." + } + ] + }, + { + "video_id": "YMFWC", + "question_id": "YMFWC", + "video_file": "Charades/videos/YMFWC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts shoes outside the door on the floor'." + }, + { + "from": "gpt", + "value": "From <13.7> to <20.4>." + } + ] + }, + { + "video_id": "ZJ54N", + "question_id": "ZJ54N", + "video_file": "Charades/videos/ZJ54N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begin undressing by taking off a jacket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <27.6>." + } + ] + }, + { + "video_id": "QEZX9", + "question_id": "QEZX9", + "video_file": "Charades/videos/QEZX9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns off the light'." + }, + { + "from": "gpt", + "value": "From <11.0> to <16.1>." + } + ] + }, + { + "video_id": "9Q1N7", + "question_id": "9Q1N7", + "video_file": "Charades/videos/9Q1N7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is running to the bed'." + }, + { + "from": "gpt", + "value": "From <7.5> to <17.5>." + } + ] + }, + { + "video_id": "9Q1N7", + "question_id": "9Q1N7", + "video_file": "Charades/videos/9Q1N7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person run down the stairs' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <17.5>." + } + ] + }, + { + "video_id": "WKDPZ", + "question_id": "WKDPZ", + "video_file": "Charades/videos/WKDPZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the picture on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.5> to <26.5>." + } + ] + }, + { + "video_id": "6FYXS", + "question_id": "6FYXS", + "video_file": "Charades/videos/6FYXS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a closet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.0> to <13.8>." + } + ] + }, + { + "video_id": "6FYXS", + "question_id": "6FYXS", + "video_file": "Charades/videos/6FYXS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes out a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.1> to <12.5>." + } + ] + }, + { + "video_id": "6FYXS", + "question_id": "6FYXS", + "video_file": "Charades/videos/6FYXS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person snuggling with the pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.4> to <17.2>." + } + ] + }, + { + "video_id": "SQKR0", + "question_id": "SQKR0", + "video_file": "Charades/videos/SQKR0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs back down the hallway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.5>." + } + ] + }, + { + "video_id": "SQKR0", + "question_id": "SQKR0", + "video_file": "Charades/videos/SQKR0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'this person runs back down the hall way' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.5>." + } + ] + }, + { + "video_id": "SQKR0", + "question_id": "SQKR0", + "video_file": "Charades/videos/SQKR0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person running down the hall'." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.5>." + } + ] + }, + { + "video_id": "SQKR0", + "question_id": "SQKR0", + "video_file": "Charades/videos/SQKR0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person continues to run' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.5>." + } + ] + }, + { + "video_id": "SQKR0", + "question_id": "SQKR0", + "video_file": "Charades/videos/SQKR0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <18.2>." + } + ] + }, + { + "video_id": "CDNNO", + "question_id": "CDNNO", + "video_file": "Charades/videos/CDNNO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person the other to tidy up a shelf'." + }, + { + "from": "gpt", + "value": "From <10.9> to <20.2>." + } + ] + }, + { + "video_id": "UPWM5", + "question_id": "UPWM5", + "video_file": "Charades/videos/UPWM5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts down the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <26.5>." + } + ] + }, + { + "video_id": "UPWM5", + "question_id": "UPWM5", + "video_file": "Charades/videos/UPWM5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put the box on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <26.5>." + } + ] + }, + { + "video_id": "TVS1P", + "question_id": "TVS1P", + "video_file": "Charades/videos/TVS1P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is standing up to turn off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.4> to <25.9>." + } + ] + }, + { + "video_id": "TVS1P", + "question_id": "TVS1P", + "video_file": "Charades/videos/TVS1P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turning off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.4> to <25.9>." + } + ] + }, + { + "video_id": "VXRCZ", + "question_id": "VXRCZ", + "video_file": "Charades/videos/VXRCZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person is looking out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.6>." + } + ] + }, + { + "video_id": "VXRCZ", + "question_id": "VXRCZ", + "video_file": "Charades/videos/VXRCZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eat a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.9> to <15.6>." + } + ] + }, + { + "video_id": "VXRCZ", + "question_id": "VXRCZ", + "video_file": "Charades/videos/VXRCZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person laughs as they look out the window'." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.6>." + } + ] + }, + { + "video_id": "JUGS8", + "question_id": "JUGS8", + "video_file": "Charades/videos/JUGS8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person a is holding a bottle of medicine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.7> to <23.7>." + } + ] + }, + { + "video_id": "JUGS8", + "question_id": "JUGS8", + "video_file": "Charades/videos/JUGS8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person b is undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <14.8>." + } + ] + }, + { + "video_id": "JUGS8", + "question_id": "JUGS8", + "video_file": "Charades/videos/JUGS8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'watching another person undressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <14.8>." + } + ] + }, + { + "video_id": "L62J5", + "question_id": "L62J5", + "video_file": "Charades/videos/L62J5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes some food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <14.0>." + } + ] + }, + { + "video_id": "PHIIX", + "question_id": "PHIIX", + "video_file": "Charades/videos/PHIIX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the pillow on the shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <11.7>." + } + ] + }, + { + "video_id": "ETAFB", + "question_id": "ETAFB", + "video_file": "Charades/videos/ETAFB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens the refrigerator looking for food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.4> to <18.9>." + } + ] + }, + { + "video_id": "5AHQV", + "question_id": "5AHQV", + "video_file": "Charades/videos/5AHQV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person angrily turns on the lights' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.1> to <4.8>." + } + ] + }, + { + "video_id": "5AHQV", + "question_id": "5AHQV", + "video_file": "Charades/videos/5AHQV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turning on the light'." + }, + { + "from": "gpt", + "value": "From <0.1> to <4.8>." + } + ] + }, + { + "video_id": "5AHQV", + "question_id": "5AHQV", + "video_file": "Charades/videos/5AHQV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws the pillow on the floor'." + }, + { + "from": "gpt", + "value": "From <6.3> to <14.6>." + } + ] + }, + { + "video_id": "5AHQV", + "question_id": "5AHQV", + "video_file": "Charades/videos/5AHQV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person throws a pillow into the corner' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.3> to <14.6>." + } + ] + }, + { + "video_id": "OFNQV", + "question_id": "OFNQV", + "video_file": "Charades/videos/OFNQV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws the book on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.4> to <30.9>." + } + ] + }, + { + "video_id": "DGSBQ", + "question_id": "DGSBQ", + "video_file": "Charades/videos/DGSBQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the food away on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.1> to <20.0>." + } + ] + }, + { + "video_id": "TYKVA", + "question_id": "TYKVA", + "video_file": "Charades/videos/TYKVA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the pillows back on' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.0> to <25.0>." + } + ] + }, + { + "video_id": "TYKVA", + "question_id": "TYKVA", + "video_file": "Charades/videos/TYKVA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts a pillow on the bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <25.0>." + } + ] + }, + { + "video_id": "PZQIN", + "question_id": "PZQIN", + "video_file": "Charades/videos/PZQIN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the pantry door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.4>." + } + ] + }, + { + "video_id": "PZQIN", + "question_id": "PZQIN", + "video_file": "Charades/videos/PZQIN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closing the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.4> to <22.6>." + } + ] + }, + { + "video_id": "E7DKD", + "question_id": "E7DKD", + "video_file": "Charades/videos/E7DKD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is opening a bag'." + }, + { + "from": "gpt", + "value": "From <0.3> to <8.5>." + } + ] + }, + { + "video_id": "B53VP", + "question_id": "B53VP", + "video_file": "Charades/videos/B53VP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begin sneezing into their hands'." + }, + { + "from": "gpt", + "value": "From <13.2> to <18.8>." + } + ] + }, + { + "video_id": "B53VP", + "question_id": "B53VP", + "video_file": "Charades/videos/B53VP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person sits on the floor holding a pillow'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "B53VP", + "question_id": "B53VP", + "video_file": "Charades/videos/B53VP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take the pillow from the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "H8QM1", + "question_id": "H8QM1", + "video_file": "Charades/videos/H8QM1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person is sneezing in pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.3> to <5.1>." + } + ] + }, + { + "video_id": "HAR5P", + "question_id": "HAR5P", + "video_file": "Charades/videos/HAR5P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person running up the stairs' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.2> to <33.0>." + } + ] + }, + { + "video_id": "HAR5P", + "question_id": "HAR5P", + "video_file": "Charades/videos/HAR5P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person runs to the top of the steps'." + }, + { + "from": "gpt", + "value": "From <25.2> to <33.0>." + } + ] + }, + { + "video_id": "HAR5P", + "question_id": "HAR5P", + "video_file": "Charades/videos/HAR5P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakens on the stairs' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.1>." + } + ] + }, + { + "video_id": "HAR5P", + "question_id": "HAR5P", + "video_file": "Charades/videos/HAR5P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person fix thier hair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <23.5>." + } + ] + }, + { + "video_id": "T0PJG", + "question_id": "T0PJG", + "video_file": "Charades/videos/T0PJG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks through the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <10.9>." + } + ] + }, + { + "video_id": "T0PJG", + "question_id": "T0PJG", + "video_file": "Charades/videos/T0PJG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits on the couch'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.6> to <13.5>." + } + ] + }, + { + "video_id": "T0PJG", + "question_id": "T0PJG", + "video_file": "Charades/videos/T0PJG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts the folded towel on the shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <29.5> to <34.5>." + } + ] + }, + { + "video_id": "T0PJG", + "question_id": "T0PJG", + "video_file": "Charades/videos/T0PJG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'sits on the bed next to a person lying down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.6> to <13.5>." + } + ] + }, + { + "video_id": "T0PJG", + "question_id": "T0PJG", + "video_file": "Charades/videos/T0PJG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person who entered the room sits on a bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.6> to <13.5>." + } + ] + }, + { + "video_id": "T0PJG", + "question_id": "T0PJG", + "video_file": "Charades/videos/T0PJG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts it on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <29.5> to <34.5>." + } + ] + }, + { + "video_id": "V3CWF", + "question_id": "V3CWF", + "video_file": "Charades/videos/V3CWF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is seen putting a book on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <9.4>." + } + ] + }, + { + "video_id": "V3CWF", + "question_id": "V3CWF", + "video_file": "Charades/videos/V3CWF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts a book on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <9.4>." + } + ] + }, + { + "video_id": "RSRZ0", + "question_id": "RSRZ0", + "video_file": "Charades/videos/RSRZ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the first person opens the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.4> to <28.0>." + } + ] + }, + { + "video_id": "RSRZ0", + "question_id": "RSRZ0", + "video_file": "Charades/videos/RSRZ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a second person is holding a broom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "RSRZ0", + "question_id": "RSRZ0", + "video_file": "Charades/videos/RSRZ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the window'." + }, + { + "from": "gpt", + "value": "From <19.4> to <28.0>." + } + ] + }, + { + "video_id": "WZVHJ", + "question_id": "WZVHJ", + "video_file": "Charades/videos/WZVHJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <16.8> to <24.9>." + } + ] + }, + { + "video_id": "WZVHJ", + "question_id": "WZVHJ", + "video_file": "Charades/videos/WZVHJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.2> to <23.4>." + } + ] + }, + { + "video_id": "TEPWB", + "question_id": "TEPWB", + "video_file": "Charades/videos/TEPWB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put some clothes in it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <16.1>." + } + ] + }, + { + "video_id": "E6A0Y", + "question_id": "E6A0Y", + "video_file": "Charades/videos/E6A0Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put some clothes in the washer' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <11.0>." + } + ] + }, + { + "video_id": "DSXEN", + "question_id": "DSXEN", + "video_file": "Charades/videos/DSXEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person make a sandwich'." + }, + { + "from": "gpt", + "value": "From <19.1> to <32.0>." + } + ] + }, + { + "video_id": "DSXEN", + "question_id": "DSXEN", + "video_file": "Charades/videos/DSXEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes out some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "DSXEN", + "question_id": "DSXEN", + "video_file": "Charades/videos/DSXEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person takes food from a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "RAQNI", + "question_id": "RAQNI", + "video_file": "Charades/videos/RAQNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a second person is looking out a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.6> to <33.0>." + } + ] + }, + { + "video_id": "RAQNI", + "question_id": "RAQNI", + "video_file": "Charades/videos/RAQNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person is watching out a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.6> to <33.0>." + } + ] + }, + { + "video_id": "RAQNI", + "question_id": "RAQNI", + "video_file": "Charades/videos/RAQNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person stands looking out of a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.6> to <33.0>." + } + ] + }, + { + "video_id": "RAQNI", + "question_id": "RAQNI", + "video_file": "Charades/videos/RAQNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is tidying up a pile of clothes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.7>." + } + ] + }, + { + "video_id": "RIUEI", + "question_id": "RIUEI", + "video_file": "Charades/videos/RIUEI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <12.0>." + } + ] + }, + { + "video_id": "RIUEI", + "question_id": "RIUEI", + "video_file": "Charades/videos/RIUEI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is drink walking eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <12.0>." + } + ] + }, + { + "video_id": "RIUEI", + "question_id": "RIUEI", + "video_file": "Charades/videos/RIUEI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they turn on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.1> to <19.6>." + } + ] + }, + { + "video_id": "RIUEI", + "question_id": "RIUEI", + "video_file": "Charades/videos/RIUEI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throw it in the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.1> to <29.8>." + } + ] + }, + { + "video_id": "2I871", + "question_id": "2I871", + "video_file": "Charades/videos/2I871.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs away dancing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.2> to <32.0>." + } + ] + }, + { + "video_id": "2I871", + "question_id": "2I871", + "video_file": "Charades/videos/2I871.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person running in to the house' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.2> to <32.0>." + } + ] + }, + { + "video_id": "2I871", + "question_id": "2I871", + "video_file": "Charades/videos/2I871.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person running out of the house' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.2> to <32.0>." + } + ] + }, + { + "video_id": "LSVHK", + "question_id": "LSVHK", + "video_file": "Charades/videos/LSVHK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes a towel away from the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.3> to <20.7>." + } + ] + }, + { + "video_id": "LSVHK", + "question_id": "LSVHK", + "video_file": "Charades/videos/LSVHK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes a towel off a door'." + }, + { + "from": "gpt", + "value": "From <15.3> to <20.7>." + } + ] + }, + { + "video_id": "AXKNF", + "question_id": "AXKNF", + "video_file": "Charades/videos/AXKNF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours it into a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <12.5>." + } + ] + }, + { + "video_id": "AXKNF", + "question_id": "AXKNF", + "video_file": "Charades/videos/AXKNF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is pouring some water into a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.4> to <12.5>." + } + ] + }, + { + "video_id": "OM5R4", + "question_id": "OM5R4", + "video_file": "Charades/videos/OM5R4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens closet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <19.0>." + } + ] + }, + { + "video_id": "OM5R4", + "question_id": "OM5R4", + "video_file": "Charades/videos/OM5R4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person after awakening'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <11.0>." + } + ] + }, + { + "video_id": "DNXBJ", + "question_id": "DNXBJ", + "video_file": "Charades/videos/DNXBJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is closing the cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.7> to <22.5>." + } + ] + }, + { + "video_id": "R5MFX", + "question_id": "R5MFX", + "video_file": "Charades/videos/R5MFX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts a dish on the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.6> to <14.4>." + } + ] + }, + { + "video_id": "E3S4O", + "question_id": "E3S4O", + "video_file": "Charades/videos/E3S4O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns on light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.8> to <32.0>." + } + ] + }, + { + "video_id": "E3S4O", + "question_id": "E3S4O", + "video_file": "Charades/videos/E3S4O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person plays with the light switch' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.8> to <32.0>." + } + ] + }, + { + "video_id": "3NE5P", + "question_id": "3NE5P", + "video_file": "Charades/videos/3NE5P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is cooking on the stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.0>." + } + ] + }, + { + "video_id": "3NE5P", + "question_id": "3NE5P", + "video_file": "Charades/videos/3NE5P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is cooking food at a stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.0>." + } + ] + }, + { + "video_id": "H1XBH", + "question_id": "H1XBH", + "video_file": "Charades/videos/H1XBH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'that same person eats some chips'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.7>." + } + ] + }, + { + "video_id": "H1XBH", + "question_id": "H1XBH", + "video_file": "Charades/videos/H1XBH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins to eat some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.7>." + } + ] + }, + { + "video_id": "H1XBH", + "question_id": "H1XBH", + "video_file": "Charades/videos/H1XBH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is holding a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "H1XBH", + "question_id": "H1XBH", + "video_file": "Charades/videos/H1XBH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is putting a pillow on a sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.2> to <7.2>." + } + ] + }, + { + "video_id": "5Y3VW", + "question_id": "5Y3VW", + "video_file": "Charades/videos/5Y3VW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts the towel over the shower curtain' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <14.3>." + } + ] + }, + { + "video_id": "O8BH6", + "question_id": "O8BH6", + "video_file": "Charades/videos/O8BH6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person laughing at it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.2> to <12.8>." + } + ] + }, + { + "video_id": "3ZHEX", + "question_id": "3ZHEX", + "video_file": "Charades/videos/3ZHEX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "3ZHEX", + "question_id": "3ZHEX", + "video_file": "Charades/videos/3ZHEX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person a door opens'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "3ZHEX", + "question_id": "3ZHEX", + "video_file": "Charades/videos/3ZHEX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <9.4>." + } + ] + }, + { + "video_id": "3ZHEX", + "question_id": "3ZHEX", + "video_file": "Charades/videos/3ZHEX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <14.5>." + } + ] + }, + { + "video_id": "3ZHEX", + "question_id": "3ZHEX", + "video_file": "Charades/videos/3ZHEX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking a cup of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <7.1>." + } + ] + }, + { + "video_id": "3ZHEX", + "question_id": "3ZHEX", + "video_file": "Charades/videos/3ZHEX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they eat a bite of their sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <14.5>." + } + ] + }, + { + "video_id": "UG0TA", + "question_id": "UG0TA", + "video_file": "Charades/videos/UG0TA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person open their pantry door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "UG0TA", + "question_id": "UG0TA", + "video_file": "Charades/videos/UG0TA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <22.8>." + } + ] + }, + { + "video_id": "B4F0T", + "question_id": "B4F0T", + "video_file": "Charades/videos/B4F0T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits down on floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <28.6>." + } + ] + }, + { + "video_id": "GYB9U", + "question_id": "GYB9U", + "video_file": "Charades/videos/GYB9U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is playing with their phone in bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "PW6CO", + "question_id": "PW6CO", + "video_file": "Charades/videos/PW6CO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person awakens on the floor of the hallway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <29.0>." + } + ] + }, + { + "video_id": "AXKNP", + "question_id": "AXKNP", + "video_file": "Charades/videos/AXKNP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "2WGSN", + "question_id": "2WGSN", + "video_file": "Charades/videos/2WGSN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person drinks from a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.1>." + } + ] + }, + { + "video_id": "2WGSN", + "question_id": "2WGSN", + "video_file": "Charades/videos/2WGSN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person drinks from a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.1>." + } + ] + }, + { + "video_id": "CPF9H", + "question_id": "CPF9H", + "video_file": "Charades/videos/CPF9H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person laughing as they flips the page' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <9.8>." + } + ] + }, + { + "video_id": "DQEC3", + "question_id": "DQEC3", + "video_file": "Charades/videos/DQEC3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts groceries away in the pantry' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.8> to <31.0>." + } + ] + }, + { + "video_id": "DQEC3", + "question_id": "DQEC3", + "video_file": "Charades/videos/DQEC3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts them on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.1>." + } + ] + }, + { + "video_id": "DQEC3", + "question_id": "DQEC3", + "video_file": "Charades/videos/DQEC3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens bag of chips'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.7> to <17.2>." + } + ] + }, + { + "video_id": "DQEC3", + "question_id": "DQEC3", + "video_file": "Charades/videos/DQEC3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins food taken from it'." + }, + { + "from": "gpt", + "value": "From <27.8> to <31.0>." + } + ] + }, + { + "video_id": "DQEC3", + "question_id": "DQEC3", + "video_file": "Charades/videos/DQEC3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.7> to <17.2>." + } + ] + }, + { + "video_id": "Y9SMX", + "question_id": "Y9SMX", + "video_file": "Charades/videos/Y9SMX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looking outside window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.2> to <44.1>." + } + ] + }, + { + "video_id": "Y9SMX", + "question_id": "Y9SMX", + "video_file": "Charades/videos/Y9SMX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens refrigerator's door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <43.1> to <50.0>." + } + ] + }, + { + "video_id": "Y9SMX", + "question_id": "Y9SMX", + "video_file": "Charades/videos/Y9SMX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is pouring a cup of coffee' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <19.1>." + } + ] + }, + { + "video_id": "Y9SMX", + "question_id": "Y9SMX", + "video_file": "Charades/videos/Y9SMX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening the door of the refrigerator in a kitchen' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <43.1> to <50.0>." + } + ] + }, + { + "video_id": "Y9SMX", + "question_id": "Y9SMX", + "video_file": "Charades/videos/Y9SMX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pours coffee into cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <19.1>." + } + ] + }, + { + "video_id": "CTOQR", + "question_id": "CTOQR", + "video_file": "Charades/videos/CTOQR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person turns off the light as they leave' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.3> to <25.8>." + } + ] + }, + { + "video_id": "CTOQR", + "question_id": "CTOQR", + "video_file": "Charades/videos/CTOQR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person smiling in the mirror'." + }, + { + "from": "gpt", + "value": "From <15.4> to <21.8>." + } + ] + }, + { + "video_id": "CTOQR", + "question_id": "CTOQR", + "video_file": "Charades/videos/CTOQR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they turn off the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.3> to <25.8>." + } + ] + }, + { + "video_id": "NODDU", + "question_id": "NODDU", + "video_file": "Charades/videos/NODDU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.9> to <31.0>." + } + ] + }, + { + "video_id": "PXYN8", + "question_id": "PXYN8", + "video_file": "Charades/videos/PXYN8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person runs around holding a pile of clothes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.2>." + } + ] + }, + { + "video_id": "N67PL", + "question_id": "N67PL", + "video_file": "Charades/videos/N67PL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person the are laughing'." + }, + { + "from": "gpt", + "value": "From <15.4> to <21.6>." + } + ] + }, + { + "video_id": "N67PL", + "question_id": "N67PL", + "video_file": "Charades/videos/N67PL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.4> to <21.6>." + } + ] + }, + { + "video_id": "71QKB", + "question_id": "71QKB", + "video_file": "Charades/videos/71QKB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person eats a sandwich in the kitchen' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "71QKB", + "question_id": "71QKB", + "video_file": "Charades/videos/71QKB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "RYDUK", + "question_id": "RYDUK", + "video_file": "Charades/videos/RYDUK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person begins undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <8.3>." + } + ] + }, + { + "video_id": "RYDUK", + "question_id": "RYDUK", + "video_file": "Charades/videos/RYDUK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the clothes in the closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.3> to <9.1>." + } + ] + }, + { + "video_id": "J39ZC", + "question_id": "J39ZC", + "video_file": "Charades/videos/J39ZC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is eating something from a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.8> to <36.0>." + } + ] + }, + { + "video_id": "J39ZC", + "question_id": "J39ZC", + "video_file": "Charades/videos/J39ZC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person someone is standing by a door eating something' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.8> to <36.0>." + } + ] + }, + { + "video_id": "75AX5", + "question_id": "75AX5", + "video_file": "Charades/videos/75AX5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs into the kitchen'." + }, + { + "from": "gpt", + "value": "From <0.4> to <7.6>." + } + ] + }, + { + "video_id": "75AX5", + "question_id": "75AX5", + "video_file": "Charades/videos/75AX5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins cooking food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.2> to <12.3>." + } + ] + }, + { + "video_id": "75AX5", + "question_id": "75AX5", + "video_file": "Charades/videos/75AX5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <42.3> to <47.6>." + } + ] + }, + { + "video_id": "75AX5", + "question_id": "75AX5", + "video_file": "Charades/videos/75AX5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person opens the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <36.7> to <47.9>." + } + ] + }, + { + "video_id": "75AX5", + "question_id": "75AX5", + "video_file": "Charades/videos/75AX5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a man running into the kitchen' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <7.6>." + } + ] + }, + { + "video_id": "DML20", + "question_id": "DML20", + "video_file": "Charades/videos/DML20.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person takes off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "ABK63", + "question_id": "ABK63", + "video_file": "Charades/videos/ABK63.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person drinks from a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.7> to <26.3>." + } + ] + }, + { + "video_id": "TBU3U", + "question_id": "TBU3U", + "video_file": "Charades/videos/TBU3U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a pair of shoes from the chair'." + }, + { + "from": "gpt", + "value": "From <7.5> to <14.6>." + } + ] + }, + { + "video_id": "67RSR", + "question_id": "67RSR", + "video_file": "Charades/videos/67RSR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person after awakening' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <16.4>." + } + ] + }, + { + "video_id": "TE5P7", + "question_id": "TE5P7", + "video_file": "Charades/videos/TE5P7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is opening the refrigerator'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "TE5P7", + "question_id": "TE5P7", + "video_file": "Charades/videos/TE5P7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pouring a drink in a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.7> to <13.8>." + } + ] + }, + { + "video_id": "AKALB", + "question_id": "AKALB", + "video_file": "Charades/videos/AKALB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.4> to <20.4>." + } + ] + }, + { + "video_id": "AKALB", + "question_id": "AKALB", + "video_file": "Charades/videos/AKALB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins throwing their pillows at the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.4> to <20.4>." + } + ] + }, + { + "video_id": "BD4P9", + "question_id": "BD4P9", + "video_file": "Charades/videos/BD4P9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.8> to <24.3>." + } + ] + }, + { + "video_id": "BD4P9", + "question_id": "BD4P9", + "video_file": "Charades/videos/BD4P9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turning on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.7> to <28.4>." + } + ] + }, + { + "video_id": "BD4P9", + "question_id": "BD4P9", + "video_file": "Charades/videos/BD4P9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turn on a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.7> to <28.4>." + } + ] + }, + { + "video_id": "BD4P9", + "question_id": "BD4P9", + "video_file": "Charades/videos/BD4P9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns on a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.7> to <28.4>." + } + ] + }, + { + "video_id": "J6TVB", + "question_id": "J6TVB", + "video_file": "Charades/videos/J6TVB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'next the person begins to play with a phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <31.0>." + } + ] + }, + { + "video_id": "J6TVB", + "question_id": "J6TVB", + "video_file": "Charades/videos/J6TVB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'that same person puts the glass down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.8> to <21.4>." + } + ] + }, + { + "video_id": "J6TVB", + "question_id": "J6TVB", + "video_file": "Charades/videos/J6TVB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person takes a cup of the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "7CXYB", + "question_id": "7CXYB", + "video_file": "Charades/videos/7CXYB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is holding a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <20.9>." + } + ] + }, + { + "video_id": "X1RBM", + "question_id": "X1RBM", + "video_file": "Charades/videos/X1RBM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person gets up ant eats something over the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.9> to <35.0>." + } + ] + }, + { + "video_id": "X1RBM", + "question_id": "X1RBM", + "video_file": "Charades/videos/X1RBM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begings to eat on top of the sink'." + }, + { + "from": "gpt", + "value": "From <20.9> to <35.0>." + } + ] + }, + { + "video_id": "X1RBM", + "question_id": "X1RBM", + "video_file": "Charades/videos/X1RBM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats food off a sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.9> to <35.0>." + } + ] + }, + { + "video_id": "WT2C0", + "question_id": "WT2C0", + "video_file": "Charades/videos/WT2C0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <32.0>." + } + ] + }, + { + "video_id": "WT2C0", + "question_id": "WT2C0", + "video_file": "Charades/videos/WT2C0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person use it to wipe the sweat from their forehead' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.6> to <32.0>." + } + ] + }, + { + "video_id": "PSPMA", + "question_id": "PSPMA", + "video_file": "Charades/videos/PSPMA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take a bit out of their food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.4> to <28.8>." + } + ] + }, + { + "video_id": "PSPMA", + "question_id": "PSPMA", + "video_file": "Charades/videos/PSPMA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.9> to <32.0>." + } + ] + }, + { + "video_id": "7BL9D", + "question_id": "7BL9D", + "video_file": "Charades/videos/7BL9D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <11.5>." + } + ] + }, + { + "video_id": "7BL9D", + "question_id": "7BL9D", + "video_file": "Charades/videos/7BL9D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating something'." + }, + { + "from": "gpt", + "value": "From <4.2> to <11.5>." + } + ] + }, + { + "video_id": "AIOTI", + "question_id": "AIOTI", + "video_file": "Charades/videos/AIOTI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person was reading a book in the shower' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.0>." + } + ] + }, + { + "video_id": "AIOTI", + "question_id": "AIOTI", + "video_file": "Charades/videos/AIOTI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is reading a book in the bathroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.0>." + } + ] + }, + { + "video_id": "FMJZQ", + "question_id": "FMJZQ", + "video_file": "Charades/videos/FMJZQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.7> to <9.6>." + } + ] + }, + { + "video_id": "A5PL0", + "question_id": "A5PL0", + "video_file": "Charades/videos/A5PL0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens up a bag of groceries to look inside'." + }, + { + "from": "gpt", + "value": "From <17.6> to <31.0>." + } + ] + }, + { + "video_id": "MV9N4", + "question_id": "MV9N4", + "video_file": "Charades/videos/MV9N4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <8.8>." + } + ] + }, + { + "video_id": "KN0TK", + "question_id": "KN0TK", + "video_file": "Charades/videos/KN0TK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.7> to <7.6>." + } + ] + }, + { + "video_id": "NMC3H", + "question_id": "NMC3H", + "video_file": "Charades/videos/NMC3H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is undressing in front of a mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "NMC3H", + "question_id": "NMC3H", + "video_file": "Charades/videos/NMC3H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is undressing in the kitchen' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "Q7RK6", + "question_id": "Q7RK6", + "video_file": "Charades/videos/Q7RK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person fixes his hair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <15.8>." + } + ] + }, + { + "video_id": "VDXXV", + "question_id": "VDXXV", + "video_file": "Charades/videos/VDXXV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.3> to <23.8>." + } + ] + }, + { + "video_id": "VDXXV", + "question_id": "VDXXV", + "video_file": "Charades/videos/VDXXV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person lies down on the bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.9> to <21.9>." + } + ] + }, + { + "video_id": "VDXXV", + "question_id": "VDXXV", + "video_file": "Charades/videos/VDXXV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening the closet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.5> to <28.0>." + } + ] + }, + { + "video_id": "OJLFZ", + "question_id": "OJLFZ", + "video_file": "Charades/videos/OJLFZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking from a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.4> to <19.2>." + } + ] + }, + { + "video_id": "OJLFZ", + "question_id": "OJLFZ", + "video_file": "Charades/videos/OJLFZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sits on the sofa' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.4> to <8.1>." + } + ] + }, + { + "video_id": "OJLFZ", + "question_id": "OJLFZ", + "video_file": "Charades/videos/OJLFZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open a box with clothes in it'." + }, + { + "from": "gpt", + "value": "From <18.2> to <25.9>." + } + ] + }, + { + "video_id": "OLV1L", + "question_id": "OLV1L", + "video_file": "Charades/videos/OLV1L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins to fix a vacuum'." + }, + { + "from": "gpt", + "value": "From <4.9> to <18.8>." + } + ] + }, + { + "video_id": "830OE", + "question_id": "830OE", + "video_file": "Charades/videos/830OE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person throws a pillow at a person drinking coffee'." + }, + { + "from": "gpt", + "value": "From <2.6> to <7.7>." + } + ] + }, + { + "video_id": "830OE", + "question_id": "830OE", + "video_file": "Charades/videos/830OE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they both laugh'." + }, + { + "from": "gpt", + "value": "From <6.3> to <12.0>." + } + ] + }, + { + "video_id": "4RLYA", + "question_id": "4RLYA", + "video_file": "Charades/videos/4RLYA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person puts some dishes on a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.7> to <11.7>." + } + ] + }, + { + "video_id": "4RLYA", + "question_id": "4RLYA", + "video_file": "Charades/videos/4RLYA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person washes a mirror'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.8> to <32.0>." + } + ] + }, + { + "video_id": "POLVK", + "question_id": "POLVK", + "video_file": "Charades/videos/POLVK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person takes clothes out of the washing machine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <11.2>." + } + ] + }, + { + "video_id": "CBPJF", + "question_id": "CBPJF", + "video_file": "Charades/videos/CBPJF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts a phone on the stove'." + }, + { + "from": "gpt", + "value": "From <9.8> to <16.8>." + } + ] + }, + { + "video_id": "CBPJF", + "question_id": "CBPJF", + "video_file": "Charades/videos/CBPJF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is playing on the phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.8>." + } + ] + }, + { + "video_id": "V5PK0", + "question_id": "V5PK0", + "video_file": "Charades/videos/V5PK0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "2FECZ", + "question_id": "2FECZ", + "video_file": "Charades/videos/2FECZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drink from a cup,' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.2> to <35.0>." + } + ] + }, + { + "video_id": "2FECZ", + "question_id": "2FECZ", + "video_file": "Charades/videos/2FECZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put a picture frame into a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.4> to <25.7>." + } + ] + }, + { + "video_id": "2FECZ", + "question_id": "2FECZ", + "video_file": "Charades/videos/2FECZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person take a drink of water from a nearby glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.7> to <32.6>." + } + ] + }, + { + "video_id": "2FECZ", + "question_id": "2FECZ", + "video_file": "Charades/videos/2FECZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person start drinking from a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.2> to <35.0>." + } + ] + }, + { + "video_id": "A5CRZ", + "question_id": "A5CRZ", + "video_file": "Charades/videos/A5CRZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits at a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.5> to <10.7>." + } + ] + }, + { + "video_id": "A5CRZ", + "question_id": "A5CRZ", + "video_file": "Charades/videos/A5CRZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person someone sitting at a table with food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.5> to <10.7>." + } + ] + }, + { + "video_id": "A5CRZ", + "question_id": "A5CRZ", + "video_file": "Charades/videos/A5CRZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits at table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <10.7>." + } + ] + }, + { + "video_id": "0NVG2", + "question_id": "0NVG2", + "video_file": "Charades/videos/0NVG2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eat some'." + }, + { + "from": "gpt", + "value": "From <22.6> to <31.0>." + } + ] + }, + { + "video_id": "0NVG2", + "question_id": "0NVG2", + "video_file": "Charades/videos/0NVG2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person took a drink from the bottle of water'." + }, + { + "from": "gpt", + "value": "From <10.2> to <20.7>." + } + ] + }, + { + "video_id": "0NVG2", + "question_id": "0NVG2", + "video_file": "Charades/videos/0NVG2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins to eat' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.6> to <31.0>." + } + ] + }, + { + "video_id": "0NVG2", + "question_id": "0NVG2", + "video_file": "Charades/videos/0NVG2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes their shoes off'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.6>." + } + ] + }, + { + "video_id": "Z1A0Q", + "question_id": "Z1A0Q", + "video_file": "Charades/videos/Z1A0Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person stand up'." + }, + { + "from": "gpt", + "value": "From <15.0> to <22.8>." + } + ] + }, + { + "video_id": "Z1A0Q", + "question_id": "Z1A0Q", + "video_file": "Charades/videos/Z1A0Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they begin sneezing uncontrollably' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.0> to <19.2>." + } + ] + }, + { + "video_id": "W3DYV", + "question_id": "W3DYV", + "video_file": "Charades/videos/W3DYV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is holding the doorknob' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.3>." + } + ] + }, + { + "video_id": "W3DYV", + "question_id": "W3DYV", + "video_file": "Charades/videos/W3DYV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "W3DYV", + "question_id": "W3DYV", + "video_file": "Charades/videos/W3DYV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holds doorknob' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.3>." + } + ] + }, + { + "video_id": "W3DYV", + "question_id": "W3DYV", + "video_file": "Charades/videos/W3DYV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person closes a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "W3DYV", + "question_id": "W3DYV", + "video_file": "Charades/videos/W3DYV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "40309", + "question_id": "40309", + "video_file": "Charades/videos/40309.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.4> to <29.2>." + } + ] + }, + { + "video_id": "ZWVO3", + "question_id": "ZWVO3", + "video_file": "Charades/videos/ZWVO3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "ZWVO3", + "question_id": "ZWVO3", + "video_file": "Charades/videos/ZWVO3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person is smiling from the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "92Q8F", + "question_id": "92Q8F", + "video_file": "Charades/videos/92Q8F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens on a bed'." + }, + { + "from": "gpt", + "value": "From <1.7> to <7.1>." + } + ] + }, + { + "video_id": "5KQL2", + "question_id": "5KQL2", + "video_file": "Charades/videos/5KQL2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts them back on the shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.2> to <16.1>." + } + ] + }, + { + "video_id": "5KQL2", + "question_id": "5KQL2", + "video_file": "Charades/videos/5KQL2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person is holding a bag'." + }, + { + "from": "gpt", + "value": "From <0.5> to <5.6>." + } + ] + }, + { + "video_id": "5KQL2", + "question_id": "5KQL2", + "video_file": "Charades/videos/5KQL2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a bag of groceries' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.5> to <5.6>." + } + ] + }, + { + "video_id": "5KQL2", + "question_id": "5KQL2", + "video_file": "Charades/videos/5KQL2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person moves to opening a cabinet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.8> to <31.8>." + } + ] + }, + { + "video_id": "WSS3R", + "question_id": "WSS3R", + "video_file": "Charades/videos/WSS3R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person in a towel starts dressing in the closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.8> to <23.0>." + } + ] + }, + { + "video_id": "WSS3R", + "question_id": "WSS3R", + "video_file": "Charades/videos/WSS3R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begins dressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <23.0>." + } + ] + }, + { + "video_id": "C5044", + "question_id": "C5044", + "video_file": "Charades/videos/C5044.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.1>." + } + ] + }, + { + "video_id": "7JDU4", + "question_id": "7JDU4", + "video_file": "Charades/videos/7JDU4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes out a camera' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.1> to <23.0>." + } + ] + }, + { + "video_id": "7JDU4", + "question_id": "7JDU4", + "video_file": "Charades/videos/7JDU4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a picture' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <29.7>." + } + ] + }, + { + "video_id": "7JDU4", + "question_id": "7JDU4", + "video_file": "Charades/videos/7JDU4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person wards he takes a picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.5> to <29.7>." + } + ] + }, + { + "video_id": "7JDU4", + "question_id": "7JDU4", + "video_file": "Charades/videos/7JDU4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs into the basement' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "S67WR", + "question_id": "S67WR", + "video_file": "Charades/videos/S67WR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the door to leave'." + }, + { + "from": "gpt", + "value": "From <24.1> to <30.0>." + } + ] + }, + { + "video_id": "S67WR", + "question_id": "S67WR", + "video_file": "Charades/videos/S67WR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open a door'." + }, + { + "from": "gpt", + "value": "From <24.1> to <30.0>." + } + ] + }, + { + "video_id": "5DRVF", + "question_id": "5DRVF", + "video_file": "Charades/videos/5DRVF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person pours some water into a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.2>." + } + ] + }, + { + "video_id": "EPKTX", + "question_id": "EPKTX", + "video_file": "Charades/videos/EPKTX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person watch television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.1> to <24.0>." + } + ] + }, + { + "video_id": "DRPO3", + "question_id": "DRPO3", + "video_file": "Charades/videos/DRPO3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting it on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.0> to <13.7>." + } + ] + }, + { + "video_id": "DRPO3", + "question_id": "DRPO3", + "video_file": "Charades/videos/DRPO3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person picks up bottle of water & drinks'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.7> to <7.0>." + } + ] + }, + { + "video_id": "DRPO3", + "question_id": "DRPO3", + "video_file": "Charades/videos/DRPO3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens up bag'." + }, + { + "from": "gpt", + "value": "From <13.7> to <25.9>." + } + ] + }, + { + "video_id": "DRPO3", + "question_id": "DRPO3", + "video_file": "Charades/videos/DRPO3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts to open door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.3> to <32.0>." + } + ] + }, + { + "video_id": "MOT57", + "question_id": "MOT57", + "video_file": "Charades/videos/MOT57.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the glass down on the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.7> to <17.7>." + } + ] + }, + { + "video_id": "YAXGX", + "question_id": "YAXGX", + "video_file": "Charades/videos/YAXGX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they close the book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.2> to <26.6>." + } + ] + }, + { + "video_id": "B55HG", + "question_id": "B55HG", + "video_file": "Charades/videos/B55HG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <9.1>." + } + ] + }, + { + "video_id": "52QLH", + "question_id": "52QLH", + "video_file": "Charades/videos/52QLH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is opening a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <10.1>." + } + ] + }, + { + "video_id": "3AEF5", + "question_id": "3AEF5", + "video_file": "Charades/videos/3AEF5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts it in a box'." + }, + { + "from": "gpt", + "value": "From <20.7> to <31.0>." + } + ] + }, + { + "video_id": "3AEF5", + "question_id": "3AEF5", + "video_file": "Charades/videos/3AEF5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person close the box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <8.4>." + } + ] + }, + { + "video_id": "3AEF5", + "question_id": "3AEF5", + "video_file": "Charades/videos/3AEF5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person puts a blanket in a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.7> to <31.0>." + } + ] + }, + { + "video_id": "SNU9T", + "question_id": "SNU9T", + "video_file": "Charades/videos/SNU9T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting holding some medicine' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.6>." + } + ] + }, + { + "video_id": "BB25Z", + "question_id": "BB25Z", + "video_file": "Charades/videos/BB25Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person lying on a bed stand up' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.6> to <12.7>." + } + ] + }, + { + "video_id": "CWUDQ", + "question_id": "CWUDQ", + "video_file": "Charades/videos/CWUDQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.0> to <31.0>." + } + ] + }, + { + "video_id": "CWUDQ", + "question_id": "CWUDQ", + "video_file": "Charades/videos/CWUDQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person undresses in front of the television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <19.4>." + } + ] + }, + { + "video_id": "CWUDQ", + "question_id": "CWUDQ", + "video_file": "Charades/videos/CWUDQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens the door of the study' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.0> to <31.0>." + } + ] + }, + { + "video_id": "JQY4L", + "question_id": "JQY4L", + "video_file": "Charades/videos/JQY4L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person runs through holding a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.2> to <12.8>." + } + ] + }, + { + "video_id": "O441U", + "question_id": "O441U", + "video_file": "Charades/videos/O441U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts it in a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.6> to <21.4>." + } + ] + }, + { + "video_id": "B32MT", + "question_id": "B32MT", + "video_file": "Charades/videos/B32MT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turn off the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.6> to <31.0>." + } + ] + }, + { + "video_id": "B32MT", + "question_id": "B32MT", + "video_file": "Charades/videos/B32MT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open a cabinet remove a item' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <16.9>." + } + ] + }, + { + "video_id": "B32MT", + "question_id": "B32MT", + "video_file": "Charades/videos/B32MT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns off the light'." + }, + { + "from": "gpt", + "value": "From <21.6> to <31.0>." + } + ] + }, + { + "video_id": "B32MT", + "question_id": "B32MT", + "video_file": "Charades/videos/B32MT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person opens the cabinet mirror'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <16.9>." + } + ] + }, + { + "video_id": "4FXUI", + "question_id": "4FXUI", + "video_file": "Charades/videos/4FXUI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person closes the laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <35.0> to <42.3>." + } + ] + }, + { + "video_id": "BMVXJ", + "question_id": "BMVXJ", + "video_file": "Charades/videos/BMVXJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens up the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <9.8>." + } + ] + }, + { + "video_id": "BMVXJ", + "question_id": "BMVXJ", + "video_file": "Charades/videos/BMVXJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person smile as they put it down on a table'." + }, + { + "from": "gpt", + "value": "From <13.6> to <18.4>." + } + ] + }, + { + "video_id": "VHOYB", + "question_id": "VHOYB", + "video_file": "Charades/videos/VHOYB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <27.4>." + } + ] + }, + { + "video_id": "VHOYB", + "question_id": "VHOYB", + "video_file": "Charades/videos/VHOYB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a picture'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.7> to <14.6>." + } + ] + }, + { + "video_id": "VHOYB", + "question_id": "VHOYB", + "video_file": "Charades/videos/VHOYB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person a puts down the camera onto a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <20.3>." + } + ] + }, + { + "video_id": "49QXY", + "question_id": "49QXY", + "video_file": "Charades/videos/49QXY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person standing holding a mirror looking at them self' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.0>." + } + ] + }, + { + "video_id": "49QXY", + "question_id": "49QXY", + "video_file": "Charades/videos/49QXY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'smiling.the person lays the mirror down'." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.5>." + } + ] + }, + { + "video_id": "49QXY", + "question_id": "49QXY", + "video_file": "Charades/videos/49QXY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person smiles as they look in the mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.7>." + } + ] + }, + { + "video_id": "8FH2F", + "question_id": "8FH2F", + "video_file": "Charades/videos/8FH2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a door with a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <9.3>." + } + ] + }, + { + "video_id": "8FH2F", + "question_id": "8FH2F", + "video_file": "Charades/videos/8FH2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a smiling person opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.4> to <9.3>." + } + ] + }, + { + "video_id": "8FH2F", + "question_id": "8FH2F", + "video_file": "Charades/videos/8FH2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.4> to <9.3>." + } + ] + }, + { + "video_id": "8FH2F", + "question_id": "8FH2F", + "video_file": "Charades/videos/8FH2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sitting on a sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <22.5>." + } + ] + }, + { + "video_id": "8FH2F", + "question_id": "8FH2F", + "video_file": "Charades/videos/8FH2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person he sits on a chair'." + }, + { + "from": "gpt", + "value": "From <14.6> to <19.8>." + } + ] + }, + { + "video_id": "8FH2F", + "question_id": "8FH2F", + "video_file": "Charades/videos/8FH2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinks from the glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.3> to <14.6>." + } + ] + }, + { + "video_id": "8FH2F", + "question_id": "8FH2F", + "video_file": "Charades/videos/8FH2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person enters the room smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <20.6>." + } + ] + }, + { + "video_id": "8FH2F", + "question_id": "8FH2F", + "video_file": "Charades/videos/8FH2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.4> to <17.6>." + } + ] + }, + { + "video_id": "8FH2F", + "question_id": "8FH2F", + "video_file": "Charades/videos/8FH2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinks from the cup'." + }, + { + "from": "gpt", + "value": "From <4.3> to <14.6>." + } + ] + }, + { + "video_id": "8FH2F", + "question_id": "8FH2F", + "video_file": "Charades/videos/8FH2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits on the chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <19.8>." + } + ] + }, + { + "video_id": "TXAG9", + "question_id": "TXAG9", + "video_file": "Charades/videos/TXAG9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person the walk to tub takes shoes off' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <23.4>." + } + ] + }, + { + "video_id": "TXAG9", + "question_id": "TXAG9", + "video_file": "Charades/videos/TXAG9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes off their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <23.4>." + } + ] + }, + { + "video_id": "7TJ60", + "question_id": "7TJ60", + "video_file": "Charades/videos/7TJ60.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person another is throwing a pillow down the stairs' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.5> to <10.9>." + } + ] + }, + { + "video_id": "XJLHC", + "question_id": "XJLHC", + "video_file": "Charades/videos/XJLHC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person close the lid to the laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.5> to <31.1>." + } + ] + }, + { + "video_id": "XJLHC", + "question_id": "XJLHC", + "video_file": "Charades/videos/XJLHC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.5> to <31.1>." + } + ] + }, + { + "video_id": "U2SJH", + "question_id": "U2SJH", + "video_file": "Charades/videos/U2SJH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person playing with their laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.3>." + } + ] + }, + { + "video_id": "U2SJH", + "question_id": "U2SJH", + "video_file": "Charades/videos/U2SJH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating a bag of chips' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.8> to <32.0>." + } + ] + }, + { + "video_id": "U2SJH", + "question_id": "U2SJH", + "video_file": "Charades/videos/U2SJH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts eating chips' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.8> to <32.0>." + } + ] + }, + { + "video_id": "7QH3I", + "question_id": "7QH3I", + "video_file": "Charades/videos/7QH3I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person talking on the phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.3> to <32.0>." + } + ] + }, + { + "video_id": "7QH3I", + "question_id": "7QH3I", + "video_file": "Charades/videos/7QH3I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a second person is sitting on the couch' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.6> to <19.1>." + } + ] + }, + { + "video_id": "V9XR6", + "question_id": "V9XR6", + "video_file": "Charades/videos/V9XR6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take a glass from the cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.7> to <39.0>." + } + ] + }, + { + "video_id": "V9XR6", + "question_id": "V9XR6", + "video_file": "Charades/videos/V9XR6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.7> to <37.7>." + } + ] + }, + { + "video_id": "ES10S", + "question_id": "ES10S", + "video_file": "Charades/videos/ES10S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <21.6>." + } + ] + }, + { + "video_id": "ES10S", + "question_id": "ES10S", + "video_file": "Charades/videos/ES10S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turn a light on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <22.3>." + } + ] + }, + { + "video_id": "ES10S", + "question_id": "ES10S", + "video_file": "Charades/videos/ES10S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turn on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <22.3>." + } + ] + }, + { + "video_id": "1NDVD", + "question_id": "1NDVD", + "video_file": "Charades/videos/1NDVD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is playing with a phone or camera' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.1> to <27.0>." + } + ] + }, + { + "video_id": "8ESHT", + "question_id": "8ESHT", + "video_file": "Charades/videos/8ESHT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a laptop'." + }, + { + "from": "gpt", + "value": "From <14.3> to <23.7>." + } + ] + }, + { + "video_id": "3SIOT", + "question_id": "3SIOT", + "video_file": "Charades/videos/3SIOT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person the other was throwing around a bag'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "UUX4G", + "question_id": "UUX4G", + "video_file": "Charades/videos/UUX4G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a seat in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.1> to <14.5>." + } + ] + }, + { + "video_id": "UUX4G", + "question_id": "UUX4G", + "video_file": "Charades/videos/UUX4G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.9> to <20.9>." + } + ] + }, + { + "video_id": "UUX4G", + "question_id": "UUX4G", + "video_file": "Charades/videos/UUX4G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats something from the desk'." + }, + { + "from": "gpt", + "value": "From <3.1> to <14.5>." + } + ] + }, + { + "video_id": "UUX4G", + "question_id": "UUX4G", + "video_file": "Charades/videos/UUX4G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turned on the lamp light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.9> to <20.9>." + } + ] + }, + { + "video_id": "ZCZ3X", + "question_id": "ZCZ3X", + "video_file": "Charades/videos/ZCZ3X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person watching television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.1> to <30.5>." + } + ] + }, + { + "video_id": "ZCZ3X", + "question_id": "ZCZ3X", + "video_file": "Charades/videos/ZCZ3X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person washes dishes with a towel' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.0> to <13.0>." + } + ] + }, + { + "video_id": "ZHDUR", + "question_id": "ZHDUR", + "video_file": "Charades/videos/ZHDUR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is dressing in front of the wardrobe' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.1>." + } + ] + }, + { + "video_id": "2R13Z", + "question_id": "2R13Z", + "video_file": "Charades/videos/2R13Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person does work on a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <11.7>." + } + ] + }, + { + "video_id": "DINAV", + "question_id": "DINAV", + "video_file": "Charades/videos/DINAV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is pouring a hot cup of coffee' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "DINAV", + "question_id": "DINAV", + "video_file": "Charades/videos/DINAV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is pouring a beverage into a cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "K48CQ", + "question_id": "K48CQ", + "video_file": "Charades/videos/K48CQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.1> to <36.8>." + } + ] + }, + { + "video_id": "L9YK1", + "question_id": "L9YK1", + "video_file": "Charades/videos/L9YK1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is running back'." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.1>." + } + ] + }, + { + "video_id": "L9YK1", + "question_id": "L9YK1", + "video_file": "Charades/videos/L9YK1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is running with a towel' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.1>." + } + ] + }, + { + "video_id": "LOP4H", + "question_id": "LOP4H", + "video_file": "Charades/videos/LOP4H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they stand up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.4> to <19.7>." + } + ] + }, + { + "video_id": "LOP4H", + "question_id": "LOP4H", + "video_file": "Charades/videos/LOP4H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is dressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.8> to <34.5>." + } + ] + }, + { + "video_id": "LOP4H", + "question_id": "LOP4H", + "video_file": "Charades/videos/LOP4H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they throw a pillow onto the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.7> to <13.8>." + } + ] + }, + { + "video_id": "IEZFM", + "question_id": "IEZFM", + "video_file": "Charades/videos/IEZFM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins working from a laptop'." + }, + { + "from": "gpt", + "value": "From <20.0> to <31.0>." + } + ] + }, + { + "video_id": "IEZFM", + "question_id": "IEZFM", + "video_file": "Charades/videos/IEZFM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sitting down at desk opening laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.7> to <31.0>." + } + ] + }, + { + "video_id": "IEZFM", + "question_id": "IEZFM", + "video_file": "Charades/videos/IEZFM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.0> to <16.4>." + } + ] + }, + { + "video_id": "IEZFM", + "question_id": "IEZFM", + "video_file": "Charades/videos/IEZFM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns the light on' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.4> to <8.3>." + } + ] + }, + { + "video_id": "IEZFM", + "question_id": "IEZFM", + "video_file": "Charades/videos/IEZFM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.0> to <16.4>." + } + ] + }, + { + "video_id": "IEZFM", + "question_id": "IEZFM", + "video_file": "Charades/videos/IEZFM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walking to wall turning lights on'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <8.3>." + } + ] + }, + { + "video_id": "Q07NE", + "question_id": "Q07NE", + "video_file": "Charades/videos/Q07NE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a closet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.5> to <22.7>." + } + ] + }, + { + "video_id": "Q07NE", + "question_id": "Q07NE", + "video_file": "Charades/videos/Q07NE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is working on their laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "Q07NE", + "question_id": "Q07NE", + "video_file": "Charades/videos/Q07NE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is doing work on a laptop,after a'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "490S2", + "question_id": "490S2", + "video_file": "Charades/videos/490S2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the water down to check their phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <25.4>." + } + ] + }, + { + "video_id": "490S2", + "question_id": "490S2", + "video_file": "Charades/videos/490S2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking water from a glass'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "UWRQV", + "question_id": "UWRQV", + "video_file": "Charades/videos/UWRQV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person working on a laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.8> to <15.0>." + } + ] + }, + { + "video_id": "UWRQV", + "question_id": "UWRQV", + "video_file": "Charades/videos/UWRQV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the cabinet door'." + }, + { + "from": "gpt", + "value": "From <26.4> to <31.0>." + } + ] + }, + { + "video_id": "UWRQV", + "question_id": "UWRQV", + "video_file": "Charades/videos/UWRQV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes some food off the shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.5> to <22.0>." + } + ] + }, + { + "video_id": "UWRQV", + "question_id": "UWRQV", + "video_file": "Charades/videos/UWRQV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.9> to <12.5>." + } + ] + }, + { + "video_id": "UWRQV", + "question_id": "UWRQV", + "video_file": "Charades/videos/UWRQV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.9> to <12.5>." + } + ] + }, + { + "video_id": "H26YY", + "question_id": "H26YY", + "video_file": "Charades/videos/H26YY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person enters garage holding a plate with food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.4> to <5.2>." + } + ] + }, + { + "video_id": "H26YY", + "question_id": "H26YY", + "video_file": "Charades/videos/H26YY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts the plate on the shelf'." + }, + { + "from": "gpt", + "value": "From <3.7> to <8.6>." + } + ] + }, + { + "video_id": "H26YY", + "question_id": "H26YY", + "video_file": "Charades/videos/H26YY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts the dishes down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.4> to <9.4>." + } + ] + }, + { + "video_id": "H26YY", + "question_id": "H26YY", + "video_file": "Charades/videos/H26YY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts a plate of food on the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.3> to <9.1>." + } + ] + }, + { + "video_id": "H26YY", + "question_id": "H26YY", + "video_file": "Charades/videos/H26YY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats a piece of food from one of them' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.2> to <23.0>." + } + ] + }, + { + "video_id": "H26YY", + "question_id": "H26YY", + "video_file": "Charades/videos/H26YY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.2> to <23.0>." + } + ] + }, + { + "video_id": "H26YY", + "question_id": "H26YY", + "video_file": "Charades/videos/H26YY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats the food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.2> to <23.0>." + } + ] + }, + { + "video_id": "JHU46", + "question_id": "JHU46", + "video_file": "Charades/videos/JHU46.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting the bag down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.0> to <15.9>." + } + ] + }, + { + "video_id": "88Z3M", + "question_id": "88Z3M", + "video_file": "Charades/videos/88Z3M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is walking through the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <5.4>." + } + ] + }, + { + "video_id": "ME7II", + "question_id": "ME7II", + "video_file": "Charades/videos/ME7II.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits on the sofa'." + }, + { + "from": "gpt", + "value": "From <5.4> to <10.5>." + } + ] + }, + { + "video_id": "ME7II", + "question_id": "ME7II", + "video_file": "Charades/videos/ME7II.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "ME7II", + "question_id": "ME7II", + "video_file": "Charades/videos/ME7II.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person is opening door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "EUK1Y", + "question_id": "EUK1Y", + "video_file": "Charades/videos/EUK1Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person is standing in bathroom reading book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "DRXNL", + "question_id": "DRXNL", + "video_file": "Charades/videos/DRXNL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person takes books from the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <17.8>." + } + ] + }, + { + "video_id": "DRXNL", + "question_id": "DRXNL", + "video_file": "Charades/videos/DRXNL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person is calmly putting things on a shelf'." + }, + { + "from": "gpt", + "value": "From <14.3> to <19.7>." + } + ] + }, + { + "video_id": "DRXNL", + "question_id": "DRXNL", + "video_file": "Charades/videos/DRXNL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts them on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <19.7>." + } + ] + }, + { + "video_id": "DRXNL", + "question_id": "DRXNL", + "video_file": "Charades/videos/DRXNL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person throws clothes from a table into a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <15.1>." + } + ] + }, + { + "video_id": "6RPUS", + "question_id": "6RPUS", + "video_file": "Charades/videos/6RPUS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'there is a person at the stove cooking something'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.8>." + } + ] + }, + { + "video_id": "6C2DJ", + "question_id": "6C2DJ", + "video_file": "Charades/videos/6C2DJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts them on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.8> to <12.6>." + } + ] + }, + { + "video_id": "6C2DJ", + "question_id": "6C2DJ", + "video_file": "Charades/videos/6C2DJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes dishes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.4> to <30.0>." + } + ] + }, + { + "video_id": "OH3R9", + "question_id": "OH3R9", + "video_file": "Charades/videos/OH3R9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the sandwich down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.2>." + } + ] + }, + { + "video_id": "2Q2LA", + "question_id": "2Q2LA", + "video_file": "Charades/videos/2Q2LA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person entering in the kitchen puts the dishes appropriatley'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.1>." + } + ] + }, + { + "video_id": "2BUDH", + "question_id": "2BUDH", + "video_file": "Charades/videos/2BUDH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take a drink of water from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.4> to <18.1>." + } + ] + }, + { + "video_id": "JC1DV", + "question_id": "JC1DV", + "video_file": "Charades/videos/JC1DV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting their shoes on the shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.0> to <16.4>." + } + ] + }, + { + "video_id": "PO74I", + "question_id": "PO74I", + "video_file": "Charades/videos/PO74I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs up the stairs'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "AQXBN", + "question_id": "AQXBN", + "video_file": "Charades/videos/AQXBN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting on a sofa'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <15.4>." + } + ] + }, + { + "video_id": "AQXBN", + "question_id": "AQXBN", + "video_file": "Charades/videos/AQXBN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person sits down on a couch with a laptop'." + }, + { + "from": "gpt", + "value": "From <2.2> to <15.4>." + } + ] + }, + { + "video_id": "R00LO", + "question_id": "R00LO", + "video_file": "Charades/videos/R00LO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begin to eat it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.2> to <31.3>." + } + ] + }, + { + "video_id": "R00LO", + "question_id": "R00LO", + "video_file": "Charades/videos/R00LO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eat the food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.2> to <31.3>." + } + ] + }, + { + "video_id": "R00LO", + "question_id": "R00LO", + "video_file": "Charades/videos/R00LO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks through a doorway wearing a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <14.2>." + } + ] + }, + { + "video_id": "R00LO", + "question_id": "R00LO", + "video_file": "Charades/videos/R00LO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person take out some food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.7> to <21.2>." + } + ] + }, + { + "video_id": "WPPAT", + "question_id": "WPPAT", + "video_file": "Charades/videos/WPPAT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.6> to <24.1>." + } + ] + }, + { + "video_id": "WPPAT", + "question_id": "WPPAT", + "video_file": "Charades/videos/WPPAT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person walks part the doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.2> to <16.1>." + } + ] + }, + { + "video_id": "NJU3G", + "question_id": "NJU3G", + "video_file": "Charades/videos/NJU3G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they start dressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <32.0>." + } + ] + }, + { + "video_id": "NJU3G", + "question_id": "NJU3G", + "video_file": "Charades/videos/NJU3G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.0> to <9.1>." + } + ] + }, + { + "video_id": "NJU3G", + "question_id": "NJU3G", + "video_file": "Charades/videos/NJU3G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is walking into a room eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <9.1>." + } + ] + }, + { + "video_id": "NJU3G", + "question_id": "NJU3G", + "video_file": "Charades/videos/NJU3G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begin dressing in the doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.7> to <32.0>." + } + ] + }, + { + "video_id": "7SLK1", + "question_id": "7SLK1", + "video_file": "Charades/videos/7SLK1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person slightly closes a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <21.9>." + } + ] + }, + { + "video_id": "7SLK1", + "question_id": "7SLK1", + "video_file": "Charades/videos/7SLK1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is holding a phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.6>." + } + ] + }, + { + "video_id": "02SK4", + "question_id": "02SK4", + "video_file": "Charades/videos/02SK4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks into a dining room holding food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.9>." + } + ] + }, + { + "video_id": "19SJ6", + "question_id": "19SJ6", + "video_file": "Charades/videos/19SJ6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'other person walks in sits on chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <10.8>." + } + ] + }, + { + "video_id": "19SJ6", + "question_id": "19SJ6", + "video_file": "Charades/videos/19SJ6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person is sitting in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <10.8>." + } + ] + }, + { + "video_id": "ZCH1J", + "question_id": "ZCH1J", + "video_file": "Charades/videos/ZCH1J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening a bag of chips'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "G183W", + "question_id": "G183W", + "video_file": "Charades/videos/G183W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts a blanket on it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.0>." + } + ] + }, + { + "video_id": "G183W", + "question_id": "G183W", + "video_file": "Charades/videos/G183W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person put a blanket on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.0>." + } + ] + }, + { + "video_id": "G183W", + "question_id": "G183W", + "video_file": "Charades/videos/G183W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinks from a cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.1> to <29.5>." + } + ] + }, + { + "video_id": "XIG01", + "question_id": "XIG01", + "video_file": "Charades/videos/XIG01.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person is holding a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "XIG01", + "question_id": "XIG01", + "video_file": "Charades/videos/XIG01.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats some of it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.4> to <10.6>." + } + ] + }, + { + "video_id": "XIG01", + "question_id": "XIG01", + "video_file": "Charades/videos/XIG01.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.4> to <10.6>." + } + ] + }, + { + "video_id": "MW2TQ", + "question_id": "MW2TQ", + "video_file": "Charades/videos/MW2TQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <16.5>." + } + ] + }, + { + "video_id": "F86M9", + "question_id": "F86M9", + "video_file": "Charades/videos/F86M9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is laughing at a television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.1> to <24.5>." + } + ] + }, + { + "video_id": "EGY7O", + "question_id": "EGY7O", + "video_file": "Charades/videos/EGY7O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <18.6>." + } + ] + }, + { + "video_id": "EGY7O", + "question_id": "EGY7O", + "video_file": "Charades/videos/EGY7O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they open the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <18.6>." + } + ] + }, + { + "video_id": "EGY7O", + "question_id": "EGY7O", + "video_file": "Charades/videos/EGY7O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person close the frig door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <10.3>." + } + ] + }, + { + "video_id": "EGY7O", + "question_id": "EGY7O", + "video_file": "Charades/videos/EGY7O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take out food to cook with' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <27.5> to <39.0>." + } + ] + }, + { + "video_id": "EGY7O", + "question_id": "EGY7O", + "video_file": "Charades/videos/EGY7O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they continue cooking' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <27.5> to <39.0>." + } + ] + }, + { + "video_id": "EGY7O", + "question_id": "EGY7O", + "video_file": "Charades/videos/EGY7O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is cooking at the stove'." + }, + { + "from": "gpt", + "value": "From <27.5> to <39.0>." + } + ] + }, + { + "video_id": "EGY7O", + "question_id": "EGY7O", + "video_file": "Charades/videos/EGY7O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person back to the stove to begin cooking the food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.5> to <39.0>." + } + ] + }, + { + "video_id": "HH0VG", + "question_id": "HH0VG", + "video_file": "Charades/videos/HH0VG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is looking at a picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.5> to <15.3>." + } + ] + }, + { + "video_id": "G4NOT", + "question_id": "G4NOT", + "video_file": "Charades/videos/G4NOT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.9> to <28.3>." + } + ] + }, + { + "video_id": "G4NOT", + "question_id": "G4NOT", + "video_file": "Charades/videos/G4NOT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing a doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.9> to <28.3>." + } + ] + }, + { + "video_id": "G4NOT", + "question_id": "G4NOT", + "video_file": "Charades/videos/G4NOT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks through doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.9> to <30.0>." + } + ] + }, + { + "video_id": "MERBU", + "question_id": "MERBU", + "video_file": "Charades/videos/MERBU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person in one hand they are holding a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.9> to <32.0>." + } + ] + }, + { + "video_id": "MERBU", + "question_id": "MERBU", + "video_file": "Charades/videos/MERBU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person is holding a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.7> to <32.0>." + } + ] + }, + { + "video_id": "MERBU", + "question_id": "MERBU", + "video_file": "Charades/videos/MERBU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'see another person eating a snack' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.8> to <32.0>." + } + ] + }, + { + "video_id": "LNWWJ", + "question_id": "LNWWJ", + "video_file": "Charades/videos/LNWWJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throwing a pillow onto a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.6> to <9.9>." + } + ] + }, + { + "video_id": "LNWWJ", + "question_id": "LNWWJ", + "video_file": "Charades/videos/LNWWJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws down a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.6> to <9.9>." + } + ] + }, + { + "video_id": "3EYLH", + "question_id": "3EYLH", + "video_file": "Charades/videos/3EYLH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin cooking something at a stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.8> to <52.8>." + } + ] + }, + { + "video_id": "4Z3PQ", + "question_id": "4Z3PQ", + "video_file": "Charades/videos/4Z3PQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <14.0>." + } + ] + }, + { + "video_id": "4Z3PQ", + "question_id": "4Z3PQ", + "video_file": "Charades/videos/4Z3PQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs into their bedroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <7.1>." + } + ] + }, + { + "video_id": "4Z3PQ", + "question_id": "4Z3PQ", + "video_file": "Charades/videos/4Z3PQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs into a room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <7.1>." + } + ] + }, + { + "video_id": "S2GDP", + "question_id": "S2GDP", + "video_file": "Charades/videos/S2GDP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person talking on a phone'." + }, + { + "from": "gpt", + "value": "From <24.3> to <34.0>." + } + ] + }, + { + "video_id": "AJ5YQ", + "question_id": "AJ5YQ", + "video_file": "Charades/videos/AJ5YQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes out a camera' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <9.1>." + } + ] + }, + { + "video_id": "L8V10", + "question_id": "L8V10", + "video_file": "Charades/videos/L8V10.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person smiling into a mirror in a bathroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.4> to <29.0>." + } + ] + }, + { + "video_id": "5GEJS", + "question_id": "5GEJS", + "video_file": "Charades/videos/5GEJS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts his shoes under a desk' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.8> to <29.8>." + } + ] + }, + { + "video_id": "MSN1B", + "question_id": "MSN1B", + "video_file": "Charades/videos/MSN1B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drink from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.7>." + } + ] + }, + { + "video_id": "MSN1B", + "question_id": "MSN1B", + "video_file": "Charades/videos/MSN1B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a laughing person awakens in their living room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.9>." + } + ] + }, + { + "video_id": "MSN1B", + "question_id": "MSN1B", + "video_file": "Charades/videos/MSN1B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a phone'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "D2CEL", + "question_id": "D2CEL", + "video_file": "Charades/videos/D2CEL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person is looking at a picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.1> to <33.0>." + } + ] + }, + { + "video_id": "D2CEL", + "question_id": "D2CEL", + "video_file": "Charades/videos/D2CEL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person gets dressed with their chosen shirt'." + }, + { + "from": "gpt", + "value": "From <21.8> to <32.8>." + } + ] + }, + { + "video_id": "5XYPL", + "question_id": "5XYPL", + "video_file": "Charades/videos/5XYPL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begin laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <12.7>." + } + ] + }, + { + "video_id": "QC9M1", + "question_id": "QC9M1", + "video_file": "Charades/videos/QC9M1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they are holding a phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "QC9M1", + "question_id": "QC9M1", + "video_file": "Charades/videos/QC9M1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding the phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "QC9M1", + "question_id": "QC9M1", + "video_file": "Charades/videos/QC9M1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person awakens by the desk in the entryway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.5> to <37.1>." + } + ] + }, + { + "video_id": "LFBCQ", + "question_id": "LFBCQ", + "video_file": "Charades/videos/LFBCQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person wearing a hoodie opens a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <6.5>." + } + ] + }, + { + "video_id": "LFBCQ", + "question_id": "LFBCQ", + "video_file": "Charades/videos/LFBCQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person in a basement is opening a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <6.5>." + } + ] + }, + { + "video_id": "LFBCQ", + "question_id": "LFBCQ", + "video_file": "Charades/videos/LFBCQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a cabinet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.6> to <6.5>." + } + ] + }, + { + "video_id": "LFBCQ", + "question_id": "LFBCQ", + "video_file": "Charades/videos/LFBCQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes a door to the outside that is open' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.5> to <12.7>." + } + ] + }, + { + "video_id": "LFBCQ", + "question_id": "LFBCQ", + "video_file": "Charades/videos/LFBCQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'closes the door.the person turns back to the cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.5> to <12.7>." + } + ] + }, + { + "video_id": "LFBCQ", + "question_id": "LFBCQ", + "video_file": "Charades/videos/LFBCQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closing a doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.5> to <12.7>." + } + ] + }, + { + "video_id": "JK4Q2", + "question_id": "JK4Q2", + "video_file": "Charades/videos/JK4Q2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins washing the mirror over their bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.6> to <15.0>." + } + ] + }, + { + "video_id": "JK4Q2", + "question_id": "JK4Q2", + "video_file": "Charades/videos/JK4Q2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins eating it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.3> to <32.0>." + } + ] + }, + { + "video_id": "JK4Q2", + "question_id": "JK4Q2", + "video_file": "Charades/videos/JK4Q2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes food from it'." + }, + { + "from": "gpt", + "value": "From <23.4> to <30.2>." + } + ] + }, + { + "video_id": "JK4Q2", + "question_id": "JK4Q2", + "video_file": "Charades/videos/JK4Q2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks to a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "JK4Q2", + "question_id": "JK4Q2", + "video_file": "Charades/videos/JK4Q2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person turns on the light in a room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "JK4Q2", + "question_id": "JK4Q2", + "video_file": "Charades/videos/JK4Q2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they wash a mirror over the sink'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <15.0>." + } + ] + }, + { + "video_id": "JK4Q2", + "question_id": "JK4Q2", + "video_file": "Charades/videos/JK4Q2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person turns the light on'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "JK4Q2", + "question_id": "JK4Q2", + "video_file": "Charades/videos/JK4Q2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.3> to <32.0>." + } + ] + }, + { + "video_id": "N2XCF", + "question_id": "N2XCF", + "video_file": "Charades/videos/N2XCF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting clothes on the table from the dryer' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.3> to <33.8>." + } + ] + }, + { + "video_id": "N2XCF", + "question_id": "N2XCF", + "video_file": "Charades/videos/N2XCF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put clothes in the washer' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <53.0> to <59.6>." + } + ] + }, + { + "video_id": "FOA8D", + "question_id": "FOA8D", + "video_file": "Charades/videos/FOA8D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a pair of shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "PABBK", + "question_id": "PABBK", + "video_file": "Charades/videos/PABBK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they take the medicine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.7> to <21.0>." + } + ] + }, + { + "video_id": "PABBK", + "question_id": "PABBK", + "video_file": "Charades/videos/PABBK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person start eating from a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.4> to <31.0>." + } + ] + }, + { + "video_id": "PABBK", + "question_id": "PABBK", + "video_file": "Charades/videos/PABBK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats something'." + }, + { + "from": "gpt", + "value": "From <17.4> to <31.0>." + } + ] + }, + { + "video_id": "PABBK", + "question_id": "PABBK", + "video_file": "Charades/videos/PABBK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begin to eat food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.4> to <31.0>." + } + ] + }, + { + "video_id": "7BFDL", + "question_id": "7BFDL", + "video_file": "Charades/videos/7BFDL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <15.8>." + } + ] + }, + { + "video_id": "7BFDL", + "question_id": "7BFDL", + "video_file": "Charades/videos/7BFDL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person holding some empty dishes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.3>." + } + ] + }, + { + "video_id": "7BFDL", + "question_id": "7BFDL", + "video_file": "Charades/videos/7BFDL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person start dressing by putting on a shirt' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <32.3>." + } + ] + }, + { + "video_id": "7BFDL", + "question_id": "7BFDL", + "video_file": "Charades/videos/7BFDL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person bow on top of dresser picked up clothes,hat' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.2> to <32.3>." + } + ] + }, + { + "video_id": "BMYMI", + "question_id": "BMYMI", + "video_file": "Charades/videos/BMYMI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "BMYMI", + "question_id": "BMYMI", + "video_file": "Charades/videos/BMYMI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opening a refrigerator door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "ZHZST", + "question_id": "ZHZST", + "video_file": "Charades/videos/ZHZST.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts the glass in the sink' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.8> to <12.2>." + } + ] + }, + { + "video_id": "HJJ32", + "question_id": "HJJ32", + "video_file": "Charades/videos/HJJ32.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person packing his bag thrown in the bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.9> to <21.0>." + } + ] + }, + { + "video_id": "RIE30", + "question_id": "RIE30", + "video_file": "Charades/videos/RIE30.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person grab a cup off the table take the drink' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <16.7>." + } + ] + }, + { + "video_id": "RIE30", + "question_id": "RIE30", + "video_file": "Charades/videos/RIE30.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <10.0>." + } + ] + }, + { + "video_id": "RIE30", + "question_id": "RIE30", + "video_file": "Charades/videos/RIE30.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking a beverage from a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.7> to <16.7>." + } + ] + }, + { + "video_id": "1Q4SL", + "question_id": "1Q4SL", + "video_file": "Charades/videos/1Q4SL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <27.2>." + } + ] + }, + { + "video_id": "1Q4SL", + "question_id": "1Q4SL", + "video_file": "Charades/videos/1Q4SL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a laughing person is drinking a soda in the bathroom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.1>." + } + ] + }, + { + "video_id": "1Q4SL", + "question_id": "1Q4SL", + "video_file": "Charades/videos/1Q4SL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.1> to <27.2>." + } + ] + }, + { + "video_id": "1Q4SL", + "question_id": "1Q4SL", + "video_file": "Charades/videos/1Q4SL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turning off a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.4> to <32.0>." + } + ] + }, + { + "video_id": "1Q4SL", + "question_id": "1Q4SL", + "video_file": "Charades/videos/1Q4SL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person laughs drinks' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.1>." + } + ] + }, + { + "video_id": "MX0CO", + "question_id": "MX0CO", + "video_file": "Charades/videos/MX0CO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is drinking a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "FYDYO", + "question_id": "FYDYO", + "video_file": "Charades/videos/FYDYO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begin sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <18.9>." + } + ] + }, + { + "video_id": "FYDYO", + "question_id": "FYDYO", + "video_file": "Charades/videos/FYDYO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they begin sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <18.9>." + } + ] + }, + { + "video_id": "N1EDG", + "question_id": "N1EDG", + "video_file": "Charades/videos/N1EDG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put the cup of coffee on the desk'." + }, + { + "from": "gpt", + "value": "From <25.9> to <31.0>." + } + ] + }, + { + "video_id": "SCHOY", + "question_id": "SCHOY", + "video_file": "Charades/videos/SCHOY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinks a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <14.0>." + } + ] + }, + { + "video_id": "SCHOY", + "question_id": "SCHOY", + "video_file": "Charades/videos/SCHOY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turned off the lights' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.9> to <26.6>." + } + ] + }, + { + "video_id": "SCHOY", + "question_id": "SCHOY", + "video_file": "Charades/videos/SCHOY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take a drink from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <14.0>." + } + ] + }, + { + "video_id": "SCHOY", + "question_id": "SCHOY", + "video_file": "Charades/videos/SCHOY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turning off a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.9> to <26.6>." + } + ] + }, + { + "video_id": "SCHOY", + "question_id": "SCHOY", + "video_file": "Charades/videos/SCHOY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they turn off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.9> to <26.6>." + } + ] + }, + { + "video_id": "SCHOY", + "question_id": "SCHOY", + "video_file": "Charades/videos/SCHOY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is undressing in the bathroom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "SCHOY", + "question_id": "SCHOY", + "video_file": "Charades/videos/SCHOY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person wash the glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.4> to <19.9>." + } + ] + }, + { + "video_id": "Z2A7Q", + "question_id": "Z2A7Q", + "video_file": "Charades/videos/Z2A7Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'this person is running into a room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.1>." + } + ] + }, + { + "video_id": "Z2A7Q", + "question_id": "Z2A7Q", + "video_file": "Charades/videos/Z2A7Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts throwing all the books on the ground' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <14.2>." + } + ] + }, + { + "video_id": "97HAH", + "question_id": "97HAH", + "video_file": "Charades/videos/97HAH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <10.5>." + } + ] + }, + { + "video_id": "97HAH", + "question_id": "97HAH", + "video_file": "Charades/videos/97HAH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person close the cabinet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.1> to <20.1>." + } + ] + }, + { + "video_id": "Z70KK", + "question_id": "Z70KK", + "video_file": "Charades/videos/Z70KK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person is washing table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.0> to <15.0>." + } + ] + }, + { + "video_id": "MR66M", + "question_id": "MR66M", + "video_file": "Charades/videos/MR66M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eat some food from it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.5> to <32.4>." + } + ] + }, + { + "video_id": "MR66M", + "question_id": "MR66M", + "video_file": "Charades/videos/MR66M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating food from a dish'." + }, + { + "from": "gpt", + "value": "From <22.5> to <32.4>." + } + ] + }, + { + "video_id": "MY5CQ", + "question_id": "MY5CQ", + "video_file": "Charades/videos/MY5CQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <19.6>." + } + ] + }, + { + "video_id": "MY5CQ", + "question_id": "MY5CQ", + "video_file": "Charades/videos/MY5CQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person closes their laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <17.4>." + } + ] + }, + { + "video_id": "MY5CQ", + "question_id": "MY5CQ", + "video_file": "Charades/videos/MY5CQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes laptop'." + }, + { + "from": "gpt", + "value": "From <11.5> to <17.4>." + } + ] + }, + { + "video_id": "MY5CQ", + "question_id": "MY5CQ", + "video_file": "Charades/videos/MY5CQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person working on his laptop'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "01ZWG", + "question_id": "01ZWG", + "video_file": "Charades/videos/01ZWG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opening a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <11.0>." + } + ] + }, + { + "video_id": "01ZWG", + "question_id": "01ZWG", + "video_file": "Charades/videos/01ZWG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <11.0>." + } + ] + }, + { + "video_id": "2P6ZP", + "question_id": "2P6ZP", + "video_file": "Charades/videos/2P6ZP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts eating a snack'." + }, + { + "from": "gpt", + "value": "From <19.6> to <31.3>." + } + ] + }, + { + "video_id": "2P6ZP", + "question_id": "2P6ZP", + "video_file": "Charades/videos/2P6ZP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a cookie from the table'." + }, + { + "from": "gpt", + "value": "From <19.6> to <31.3>." + } + ] + }, + { + "video_id": "VLMII", + "question_id": "VLMII", + "video_file": "Charades/videos/VLMII.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person closes the laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.0> to <7.5>." + } + ] + }, + { + "video_id": "VLMII", + "question_id": "VLMII", + "video_file": "Charades/videos/VLMII.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <16.1>." + } + ] + }, + { + "video_id": "VLMII", + "question_id": "VLMII", + "video_file": "Charades/videos/VLMII.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the clothes they were wearing in the washer' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.4> to <18.1>." + } + ] + }, + { + "video_id": "MRKTD", + "question_id": "MRKTD", + "video_file": "Charades/videos/MRKTD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person holding a towel sneezes into it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.9> to <17.9>." + } + ] + }, + { + "video_id": "57J92", + "question_id": "57J92", + "video_file": "Charades/videos/57J92.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'smiles at another person who is eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <29.8> to <35.4>." + } + ] + }, + { + "video_id": "57J92", + "question_id": "57J92", + "video_file": "Charades/videos/57J92.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person another eats a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.8> to <35.4>." + } + ] + }, + { + "video_id": "57J92", + "question_id": "57J92", + "video_file": "Charades/videos/57J92.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating something'." + }, + { + "from": "gpt", + "value": "From <29.8> to <35.4>." + } + ] + }, + { + "video_id": "1YC83", + "question_id": "1YC83", + "video_file": "Charades/videos/1YC83.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts down their sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.6>." + } + ] + }, + { + "video_id": "S673O", + "question_id": "S673O", + "video_file": "Charades/videos/S673O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats a bite' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.9> to <35.0>." + } + ] + }, + { + "video_id": "S673O", + "question_id": "S673O", + "video_file": "Charades/videos/S673O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person lies on a sofa in the living room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.2>." + } + ] + }, + { + "video_id": "S673O", + "question_id": "S673O", + "video_file": "Charades/videos/S673O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person pours a drink into their glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.4> to <35.0>." + } + ] + }, + { + "video_id": "S673O", + "question_id": "S673O", + "video_file": "Charades/videos/S673O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a bite of their sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.0> to <35.0>." + } + ] + }, + { + "video_id": "DFEG2", + "question_id": "DFEG2", + "video_file": "Charades/videos/DFEG2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is eating'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <21.0>." + } + ] + }, + { + "video_id": "DFEG2", + "question_id": "DFEG2", + "video_file": "Charades/videos/DFEG2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person undresses' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "DFEG2", + "question_id": "DFEG2", + "video_file": "Charades/videos/DFEG2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they begin to eat some food'." + }, + { + "from": "gpt", + "value": "From <14.3> to <21.0>." + } + ] + }, + { + "video_id": "DFEG2", + "question_id": "DFEG2", + "video_file": "Charades/videos/DFEG2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes off their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <12.1>." + } + ] + }, + { + "video_id": "DFEG2", + "question_id": "DFEG2", + "video_file": "Charades/videos/DFEG2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits on the couch' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <21.0>." + } + ] + }, + { + "video_id": "DFEG2", + "question_id": "DFEG2", + "video_file": "Charades/videos/DFEG2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "L9H3O", + "question_id": "L9H3O", + "video_file": "Charades/videos/L9H3O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pours a glass of cola'." + }, + { + "from": "gpt", + "value": "From <9.0> to <19.0>." + } + ] + }, + { + "video_id": "L9H3O", + "question_id": "L9H3O", + "video_file": "Charades/videos/L9H3O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pours a drink into a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.0> to <24.0>." + } + ] + }, + { + "video_id": "L9H3O", + "question_id": "L9H3O", + "video_file": "Charades/videos/L9H3O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'there is a person sitting in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.0>." + } + ] + }, + { + "video_id": "L9H3O", + "question_id": "L9H3O", + "video_file": "Charades/videos/L9H3O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the cup in a sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.8> to <28.8>." + } + ] + }, + { + "video_id": "L9H3O", + "question_id": "L9H3O", + "video_file": "Charades/videos/L9H3O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sitting back down again to eat it'." + }, + { + "from": "gpt", + "value": "From <29.0> to <34.0>." + } + ] + }, + { + "video_id": "L9H3O", + "question_id": "L9H3O", + "video_file": "Charades/videos/L9H3O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pours in a glass'." + }, + { + "from": "gpt", + "value": "From <9.0> to <19.0>." + } + ] + }, + { + "video_id": "L9H3O", + "question_id": "L9H3O", + "video_file": "Charades/videos/L9H3O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'upon finishing it person takes the glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.7> to <29.2>." + } + ] + }, + { + "video_id": "L9H3O", + "question_id": "L9H3O", + "video_file": "Charades/videos/L9H3O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.0> to <34.0>." + } + ] + }, + { + "video_id": "U4FR1", + "question_id": "U4FR1", + "video_file": "Charades/videos/U4FR1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opens the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "U4FR1", + "question_id": "U4FR1", + "video_file": "Charades/videos/U4FR1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "C69Y1", + "question_id": "C69Y1", + "video_file": "Charades/videos/C69Y1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <11.2>." + } + ] + }, + { + "video_id": "C69Y1", + "question_id": "C69Y1", + "video_file": "Charades/videos/C69Y1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes out a glass'." + }, + { + "from": "gpt", + "value": "From <7.0> to <12.3>." + } + ] + }, + { + "video_id": "C69Y1", + "question_id": "C69Y1", + "video_file": "Charades/videos/C69Y1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open a cabinet to take a glass out' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <12.3>." + } + ] + }, + { + "video_id": "C69Y1", + "question_id": "C69Y1", + "video_file": "Charades/videos/C69Y1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they pour some water into the glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.7> to <28.6>." + } + ] + }, + { + "video_id": "C69Y1", + "question_id": "C69Y1", + "video_file": "Charades/videos/C69Y1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin to pour soda into the glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.7> to <28.6>." + } + ] + }, + { + "video_id": "V4PKW", + "question_id": "V4PKW", + "video_file": "Charades/videos/V4PKW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking a glass of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.8>." + } + ] + }, + { + "video_id": "V4PKW", + "question_id": "V4PKW", + "video_file": "Charades/videos/V4PKW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <26.9>." + } + ] + }, + { + "video_id": "V4PKW", + "question_id": "V4PKW", + "video_file": "Charades/videos/V4PKW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they open the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <26.9>." + } + ] + }, + { + "video_id": "XRQDI", + "question_id": "XRQDI", + "video_file": "Charades/videos/XRQDI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the other person is laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.2> to <15.8>." + } + ] + }, + { + "video_id": "XRQDI", + "question_id": "XRQDI", + "video_file": "Charades/videos/XRQDI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the second person runs into the door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "XRQDI", + "question_id": "XRQDI", + "video_file": "Charades/videos/XRQDI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'two persons are running up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "XRQDI", + "question_id": "XRQDI", + "video_file": "Charades/videos/XRQDI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person two people are running through a hallway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "XRQDI", + "question_id": "XRQDI", + "video_file": "Charades/videos/XRQDI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person one runs into a doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "XRQDI", + "question_id": "XRQDI", + "video_file": "Charades/videos/XRQDI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'laughs at the second person' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.2> to <15.8>." + } + ] + }, + { + "video_id": "XRQDI", + "question_id": "XRQDI", + "video_file": "Charades/videos/XRQDI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person runs in a room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "G06QC", + "question_id": "G06QC", + "video_file": "Charades/videos/G06QC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts a broom back' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.6> to <18.7>." + } + ] + }, + { + "video_id": "G06QC", + "question_id": "G06QC", + "video_file": "Charades/videos/G06QC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is drinking a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "G06QC", + "question_id": "G06QC", + "video_file": "Charades/videos/G06QC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes vacuum out of closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <25.0>." + } + ] + }, + { + "video_id": "G06QC", + "question_id": "G06QC", + "video_file": "Charades/videos/G06QC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a drink from a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "G06QC", + "question_id": "G06QC", + "video_file": "Charades/videos/G06QC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the broom in the closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.6> to <18.7>." + } + ] + }, + { + "video_id": "G06QC", + "question_id": "G06QC", + "video_file": "Charades/videos/G06QC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes out the vacuum'." + }, + { + "from": "gpt", + "value": "From <16.8> to <25.0>." + } + ] + }, + { + "video_id": "Q8XEE", + "question_id": "Q8XEE", + "video_file": "Charades/videos/Q8XEE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <7.8>." + } + ] + }, + { + "video_id": "Q8XEE", + "question_id": "Q8XEE", + "video_file": "Charades/videos/Q8XEE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open up a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.9> to <7.8>." + } + ] + }, + { + "video_id": "0O81T", + "question_id": "0O81T", + "video_file": "Charades/videos/0O81T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person is sneezing'." + }, + { + "from": "gpt", + "value": "From <11.5> to <21.1>." + } + ] + }, + { + "video_id": "0O81T", + "question_id": "0O81T", + "video_file": "Charades/videos/0O81T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person someone started sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <21.1>." + } + ] + }, + { + "video_id": "W7SGQ", + "question_id": "W7SGQ", + "video_file": "Charades/videos/W7SGQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person prepares to take a picture'." + }, + { + "from": "gpt", + "value": "From <23.6> to <34.0>." + } + ] + }, + { + "video_id": "FO035", + "question_id": "FO035", + "video_file": "Charades/videos/FO035.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is awakening from a nap on their sofa' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.6> to <28.4>." + } + ] + }, + { + "video_id": "FO035", + "question_id": "FO035", + "video_file": "Charades/videos/FO035.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs over to their sofa' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.4> to <34.0>." + } + ] + }, + { + "video_id": "OKVGK", + "question_id": "OKVGK", + "video_file": "Charades/videos/OKVGK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closing the door'." + }, + { + "from": "gpt", + "value": "From <2.1> to <8.0>." + } + ] + }, + { + "video_id": "OKVGK", + "question_id": "OKVGK", + "video_file": "Charades/videos/OKVGK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'as the person begins undressing'." + }, + { + "from": "gpt", + "value": "From <8.6> to <18.2>." + } + ] + }, + { + "video_id": "BBAVF", + "question_id": "BBAVF", + "video_file": "Charades/videos/BBAVF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes off their shoes'." + }, + { + "from": "gpt", + "value": "From <6.0> to <16.4>." + } + ] + }, + { + "video_id": "BBAVF", + "question_id": "BBAVF", + "video_file": "Charades/videos/BBAVF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes off their shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.0> to <16.4>." + } + ] + }, + { + "video_id": "BBAVF", + "question_id": "BBAVF", + "video_file": "Charades/videos/BBAVF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person looks out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "IR7CA", + "question_id": "IR7CA", + "video_file": "Charades/videos/IR7CA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they begin sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <13.2>." + } + ] + }, + { + "video_id": "987P8", + "question_id": "987P8", + "video_file": "Charades/videos/987P8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person throws some clothes on the ground'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "987P8", + "question_id": "987P8", + "video_file": "Charades/videos/987P8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws that in the floor'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "987P8", + "question_id": "987P8", + "video_file": "Charades/videos/987P8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begins undressing by taking off a jacket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <11.8>." + } + ] + }, + { + "video_id": "UKCTK", + "question_id": "UKCTK", + "video_file": "Charades/videos/UKCTK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person close the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <27.0> to <32.0>." + } + ] + }, + { + "video_id": "S8TI7", + "question_id": "S8TI7", + "video_file": "Charades/videos/S8TI7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <15.3>." + } + ] + }, + { + "video_id": "S8TI7", + "question_id": "S8TI7", + "video_file": "Charades/videos/S8TI7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats more of the sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.7> to <15.3>." + } + ] + }, + { + "video_id": "S8TI7", + "question_id": "S8TI7", + "video_file": "Charades/videos/S8TI7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting in the bedroom holding a sandwich'." + }, + { + "from": "gpt", + "value": "From <0.2> to <15.8>." + } + ] + }, + { + "video_id": "S8TI7", + "question_id": "S8TI7", + "video_file": "Charades/videos/S8TI7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sitting on the bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.2> to <15.8>." + } + ] + }, + { + "video_id": "S8TI7", + "question_id": "S8TI7", + "video_file": "Charades/videos/S8TI7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits on a bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <15.8>." + } + ] + }, + { + "video_id": "S8TI7", + "question_id": "S8TI7", + "video_file": "Charades/videos/S8TI7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looks out the window'." + }, + { + "from": "gpt", + "value": "From <2.3> to <9.5>." + } + ] + }, + { + "video_id": "S8TI7", + "question_id": "S8TI7", + "video_file": "Charades/videos/S8TI7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person looks out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <9.5>." + } + ] + }, + { + "video_id": "5UUEP", + "question_id": "5UUEP", + "video_file": "Charades/videos/5UUEP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'that person pour some coffee into a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.4>." + } + ] + }, + { + "video_id": "CXO6P", + "question_id": "CXO6P", + "video_file": "Charades/videos/CXO6P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes a sandwich from the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <23.7>." + } + ] + }, + { + "video_id": "CXO6P", + "question_id": "CXO6P", + "video_file": "Charades/videos/CXO6P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person appears to be sneezing or sick'." + }, + { + "from": "gpt", + "value": "From <5.7> to <16.0>." + } + ] + }, + { + "video_id": "CLW8L", + "question_id": "CLW8L", + "video_file": "Charades/videos/CLW8L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes their pillow'." + }, + { + "from": "gpt", + "value": "From <7.1> to <13.1>." + } + ] + }, + { + "video_id": "CLW8L", + "question_id": "CLW8L", + "video_file": "Charades/videos/CLW8L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'awakening another person who is under a blanket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <23.0>." + } + ] + }, + { + "video_id": "JOUM7", + "question_id": "JOUM7", + "video_file": "Charades/videos/JOUM7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.2> to <30.1>." + } + ] + }, + { + "video_id": "2C2AI", + "question_id": "2C2AI", + "video_file": "Charades/videos/2C2AI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they start sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.5> to <13.5>." + } + ] + }, + { + "video_id": "05WY6", + "question_id": "05WY6", + "video_file": "Charades/videos/05WY6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person starts sneezing in a towel'." + }, + { + "from": "gpt", + "value": "From <15.9> to <28.4>." + } + ] + }, + { + "video_id": "RXJ04", + "question_id": "RXJ04", + "video_file": "Charades/videos/RXJ04.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the blanket in a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <30.6> to <43.5>." + } + ] + }, + { + "video_id": "2JT00", + "question_id": "2JT00", + "video_file": "Charades/videos/2JT00.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pick up a plate holding a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.8> to <22.0>." + } + ] + }, + { + "video_id": "2JT00", + "question_id": "2JT00", + "video_file": "Charades/videos/2JT00.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sits smiling in a pantry' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.8> to <12.3>." + } + ] + }, + { + "video_id": "2JT00", + "question_id": "2JT00", + "video_file": "Charades/videos/2JT00.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eat it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.3> to <22.0>." + } + ] + }, + { + "video_id": "2JT00", + "question_id": "2JT00", + "video_file": "Charades/videos/2JT00.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the broom down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.4> to <20.1>." + } + ] + }, + { + "video_id": "469E8", + "question_id": "469E8", + "video_file": "Charades/videos/469E8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.8> to <20.9>." + } + ] + }, + { + "video_id": "469E8", + "question_id": "469E8", + "video_file": "Charades/videos/469E8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "469E8", + "question_id": "469E8", + "video_file": "Charades/videos/469E8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person looks out a nearby window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.9> to <18.4>." + } + ] + }, + { + "video_id": "469E8", + "question_id": "469E8", + "video_file": "Charades/videos/469E8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sits on stairs drinking from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "469E8", + "question_id": "469E8", + "video_file": "Charades/videos/469E8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looks out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.9> to <18.4>." + } + ] + }, + { + "video_id": "469E8", + "question_id": "469E8", + "video_file": "Charades/videos/469E8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looks out a window on a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.9> to <18.4>." + } + ] + }, + { + "video_id": "469E8", + "question_id": "469E8", + "video_file": "Charades/videos/469E8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.8> to <20.9>." + } + ] + }, + { + "video_id": "3XUPT", + "question_id": "3XUPT", + "video_file": "Charades/videos/3XUPT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.4>." + } + ] + }, + { + "video_id": "3XUPT", + "question_id": "3XUPT", + "video_file": "Charades/videos/3XUPT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person with a broom opens the bathroom door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.4>." + } + ] + }, + { + "video_id": "7FCAW", + "question_id": "7FCAW", + "video_file": "Charades/videos/7FCAW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person laughs as they begin undressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.3> to <20.8>." + } + ] + }, + { + "video_id": "7FCAW", + "question_id": "7FCAW", + "video_file": "Charades/videos/7FCAW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person is holding a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.9> to <10.9>." + } + ] + }, + { + "video_id": "7FCAW", + "question_id": "7FCAW", + "video_file": "Charades/videos/7FCAW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins to undress'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <20.8>." + } + ] + }, + { + "video_id": "P76PG", + "question_id": "P76PG", + "video_file": "Charades/videos/P76PG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'this person is sitting down holding a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.8>." + } + ] + }, + { + "video_id": "P76PG", + "question_id": "P76PG", + "video_file": "Charades/videos/P76PG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.8>." + } + ] + }, + { + "video_id": "FJM2Y", + "question_id": "FJM2Y", + "video_file": "Charades/videos/FJM2Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.6> to <14.2>." + } + ] + }, + { + "video_id": "FJM2Y", + "question_id": "FJM2Y", + "video_file": "Charades/videos/FJM2Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person walks around holding a bag'." + }, + { + "from": "gpt", + "value": "From <10.5> to <16.7>." + } + ] + }, + { + "video_id": "FJM2Y", + "question_id": "FJM2Y", + "video_file": "Charades/videos/FJM2Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts bag in cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <16.9>." + } + ] + }, + { + "video_id": "FJM2Y", + "question_id": "FJM2Y", + "video_file": "Charades/videos/FJM2Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person holding a bag walks through a room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.5> to <16.7>." + } + ] + }, + { + "video_id": "FJM2Y", + "question_id": "FJM2Y", + "video_file": "Charades/videos/FJM2Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take a book from the cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.2> to <29.8>." + } + ] + }, + { + "video_id": "I5PU9", + "question_id": "I5PU9", + "video_file": "Charades/videos/I5PU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person holding a phone looks out a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.8> to <33.0>." + } + ] + }, + { + "video_id": "G3WBE", + "question_id": "G3WBE", + "video_file": "Charades/videos/G3WBE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking from a glass'." + }, + { + "from": "gpt", + "value": "From <33.3> to <42.6>." + } + ] + }, + { + "video_id": "G3WBE", + "question_id": "G3WBE", + "video_file": "Charades/videos/G3WBE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a drink from a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.3> to <42.6>." + } + ] + }, + { + "video_id": "G3WBE", + "question_id": "G3WBE", + "video_file": "Charades/videos/G3WBE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking from a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.3> to <42.6>." + } + ] + }, + { + "video_id": "X084M", + "question_id": "X084M", + "video_file": "Charades/videos/X084M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is fixing a light in the garage' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.3> to <21.8>." + } + ] + }, + { + "video_id": "X084M", + "question_id": "X084M", + "video_file": "Charades/videos/X084M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they turn on the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.3> to <21.8>." + } + ] + }, + { + "video_id": "X084M", + "question_id": "X084M", + "video_file": "Charades/videos/X084M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person walks up to the light'." + }, + { + "from": "gpt", + "value": "From <11.3> to <21.8>." + } + ] + }, + { + "video_id": "4MUXJ", + "question_id": "4MUXJ", + "video_file": "Charades/videos/4MUXJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.7> to <31.0>." + } + ] + }, + { + "video_id": "4MUXJ", + "question_id": "4MUXJ", + "video_file": "Charades/videos/4MUXJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the bag down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <23.6>." + } + ] + }, + { + "video_id": "4MUXJ", + "question_id": "4MUXJ", + "video_file": "Charades/videos/4MUXJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person begins putting items in the bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.1> to <23.6>." + } + ] + }, + { + "video_id": "4Q3KR", + "question_id": "4Q3KR", + "video_file": "Charades/videos/4Q3KR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person tidying the shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <23.0>." + } + ] + }, + { + "video_id": "7P5R2", + "question_id": "7P5R2", + "video_file": "Charades/videos/7P5R2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a door'." + }, + { + "from": "gpt", + "value": "From <1.8> to <16.1>." + } + ] + }, + { + "video_id": "7P5R2", + "question_id": "7P5R2", + "video_file": "Charades/videos/7P5R2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is opening a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <16.1>." + } + ] + }, + { + "video_id": "7P5R2", + "question_id": "7P5R2", + "video_file": "Charades/videos/7P5R2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a book out of the bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <24.4>." + } + ] + }, + { + "video_id": "0NJ26", + "question_id": "0NJ26", + "video_file": "Charades/videos/0NJ26.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is putting a blanket on the couch' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.4>." + } + ] + }, + { + "video_id": "VU4J8", + "question_id": "VU4J8", + "video_file": "Charades/videos/VU4J8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "6ALEL", + "question_id": "6ALEL", + "video_file": "Charades/videos/6ALEL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begin eating it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.9> to <27.0>." + } + ] + }, + { + "video_id": "6ALEL", + "question_id": "6ALEL", + "video_file": "Charades/videos/6ALEL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a bite of a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.3> to <24.3>." + } + ] + }, + { + "video_id": "6ALEL", + "question_id": "6ALEL", + "video_file": "Charades/videos/6ALEL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.3> to <24.3>." + } + ] + }, + { + "video_id": "6ALEL", + "question_id": "6ALEL", + "video_file": "Charades/videos/6ALEL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person finishing dressing themself'." + }, + { + "from": "gpt", + "value": "From <9.7> to <16.4>." + } + ] + }, + { + "video_id": "6ALEL", + "question_id": "6ALEL", + "video_file": "Charades/videos/6ALEL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person start eating it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.9> to <27.0>." + } + ] + }, + { + "video_id": "SMVKB", + "question_id": "SMVKB", + "video_file": "Charades/videos/SMVKB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a pillow laughing opens a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <18.6>." + } + ] + }, + { + "video_id": "SMVKB", + "question_id": "SMVKB", + "video_file": "Charades/videos/SMVKB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person starts smiling' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.9> to <22.8>." + } + ] + }, + { + "video_id": "SMVKB", + "question_id": "SMVKB", + "video_file": "Charades/videos/SMVKB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person is smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <22.8>." + } + ] + }, + { + "video_id": "SMVKB", + "question_id": "SMVKB", + "video_file": "Charades/videos/SMVKB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.5> to <23.6>." + } + ] + }, + { + "video_id": "SMVKB", + "question_id": "SMVKB", + "video_file": "Charades/videos/SMVKB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person standing close to the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.4> to <32.0>." + } + ] + }, + { + "video_id": "X9ULK", + "question_id": "X9ULK", + "video_file": "Charades/videos/X9ULK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pours a cup of coffee on to another cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <26.4>." + } + ] + }, + { + "video_id": "X9ULK", + "question_id": "X9ULK", + "video_file": "Charades/videos/X9ULK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person poured themselves a cup of coffee' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <26.4>." + } + ] + }, + { + "video_id": "X9ULK", + "question_id": "X9ULK", + "video_file": "Charades/videos/X9ULK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person after dressing in the bathroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <18.7>." + } + ] + }, + { + "video_id": "EOAR1", + "question_id": "EOAR1", + "video_file": "Charades/videos/EOAR1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is putting food in the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.6> to <16.6>." + } + ] + }, + { + "video_id": "0QES3", + "question_id": "0QES3", + "video_file": "Charades/videos/0QES3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting feet in shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <14.8>." + } + ] + }, + { + "video_id": "0QES3", + "question_id": "0QES3", + "video_file": "Charades/videos/0QES3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the door'." + }, + { + "from": "gpt", + "value": "From <13.5> to <20.1>." + } + ] + }, + { + "video_id": "0QES3", + "question_id": "0QES3", + "video_file": "Charades/videos/0QES3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts on shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.1> to <14.8>." + } + ] + }, + { + "video_id": "7HEWH", + "question_id": "7HEWH", + "video_file": "Charades/videos/7HEWH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person washes a dish in the sink'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <32.0>." + } + ] + }, + { + "video_id": "HCSPE", + "question_id": "HCSPE", + "video_file": "Charades/videos/HCSPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts down the phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.1> to <30.6>." + } + ] + }, + { + "video_id": "HCSPE", + "question_id": "HCSPE", + "video_file": "Charades/videos/HCSPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "HCSPE", + "question_id": "HCSPE", + "video_file": "Charades/videos/HCSPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts their phone on the night table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.1> to <30.6>." + } + ] + }, + { + "video_id": "HCSPE", + "question_id": "HCSPE", + "video_file": "Charades/videos/HCSPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <11.2>." + } + ] + }, + { + "video_id": "HCSPE", + "question_id": "HCSPE", + "video_file": "Charades/videos/HCSPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.4> to <22.0>." + } + ] + }, + { + "video_id": "HCSPE", + "question_id": "HCSPE", + "video_file": "Charades/videos/HCSPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person someone walks into a room holding a phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.9> to <13.9>." + } + ] + }, + { + "video_id": "HCSPE", + "question_id": "HCSPE", + "video_file": "Charades/videos/HCSPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they close the door'." + }, + { + "from": "gpt", + "value": "From <15.4> to <22.0>." + } + ] + }, + { + "video_id": "HCSPE", + "question_id": "HCSPE", + "video_file": "Charades/videos/HCSPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks back over to the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <11.2>." + } + ] + }, + { + "video_id": "HCSPE", + "question_id": "HCSPE", + "video_file": "Charades/videos/HCSPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'closes a door .the person goes back to the stand'." + }, + { + "from": "gpt", + "value": "From <15.4> to <22.0>." + } + ] + }, + { + "video_id": "AI4NP", + "question_id": "AI4NP", + "video_file": "Charades/videos/AI4NP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take out some clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <11.4>." + } + ] + }, + { + "video_id": "69GFN", + "question_id": "69GFN", + "video_file": "Charades/videos/69GFN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looking at a picture on the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <12.4>." + } + ] + }, + { + "video_id": "69GFN", + "question_id": "69GFN", + "video_file": "Charades/videos/69GFN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person sits on a sofa' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <30.3>." + } + ] + }, + { + "video_id": "69GFN", + "question_id": "69GFN", + "video_file": "Charades/videos/69GFN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person gets undressed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.1> to <19.5>." + } + ] + }, + { + "video_id": "69GFN", + "question_id": "69GFN", + "video_file": "Charades/videos/69GFN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person after undressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <19.5>." + } + ] + }, + { + "video_id": "654PD", + "question_id": "654PD", + "video_file": "Charades/videos/654PD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening a book on the stairs' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.7> to <17.5>." + } + ] + }, + { + "video_id": "654PD", + "question_id": "654PD", + "video_file": "Charades/videos/654PD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person he puts the paper down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.6> to <13.8>." + } + ] + }, + { + "video_id": "V4C0P", + "question_id": "V4C0P", + "video_file": "Charades/videos/V4C0P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throw a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.0> to <18.5>." + } + ] + }, + { + "video_id": "J2DBI", + "question_id": "J2DBI", + "video_file": "Charades/videos/J2DBI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person start smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <27.0>." + } + ] + }, + { + "video_id": "J2DBI", + "question_id": "J2DBI", + "video_file": "Charades/videos/J2DBI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person finally smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <27.0>." + } + ] + }, + { + "video_id": "CWBHH", + "question_id": "CWBHH", + "video_file": "Charades/videos/CWBHH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is reading a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "30JQT", + "question_id": "30JQT", + "video_file": "Charades/videos/30JQT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they pour themselves a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.7> to <37.0>." + } + ] + }, + { + "video_id": "30JQT", + "question_id": "30JQT", + "video_file": "Charades/videos/30JQT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pours a glass of water on the table'." + }, + { + "from": "gpt", + "value": "From <27.7> to <37.0>." + } + ] + }, + { + "video_id": "ANAR5", + "question_id": "ANAR5", + "video_file": "Charades/videos/ANAR5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating food from a bag'." + }, + { + "from": "gpt", + "value": "From <3.1> to <11.6>." + } + ] + }, + { + "video_id": "B4MNI", + "question_id": "B4MNI", + "video_file": "Charades/videos/B4MNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is pouring a cup of coffee' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <11.9>." + } + ] + }, + { + "video_id": "B4MNI", + "question_id": "B4MNI", + "video_file": "Charades/videos/B4MNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they snuggle with a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <33.7> to <43.6>." + } + ] + }, + { + "video_id": "B4MNI", + "question_id": "B4MNI", + "video_file": "Charades/videos/B4MNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person pours coffee from a cup to another cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <11.9>." + } + ] + }, + { + "video_id": "B4MNI", + "question_id": "B4MNI", + "video_file": "Charades/videos/B4MNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pours water from one cup into another' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <11.9>." + } + ] + }, + { + "video_id": "B4MNI", + "question_id": "B4MNI", + "video_file": "Charades/videos/B4MNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person snuggling with a blanket in a living room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.7> to <44.0>." + } + ] + }, + { + "video_id": "B4MNI", + "question_id": "B4MNI", + "video_file": "Charades/videos/B4MNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person sits on the couch' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <8.6>." + } + ] + }, + { + "video_id": "B4MNI", + "question_id": "B4MNI", + "video_file": "Charades/videos/B4MNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person sits on a couch'." + }, + { + "from": "gpt", + "value": "From <2.5> to <8.6>." + } + ] + }, + { + "video_id": "B4MNI", + "question_id": "B4MNI", + "video_file": "Charades/videos/B4MNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person snuggles with their blanket' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <33.7> to <44.0>." + } + ] + }, + { + "video_id": "NV4FC", + "question_id": "NV4FC", + "video_file": "Charades/videos/NV4FC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes off the shoes'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.2>." + } + ] + }, + { + "video_id": "NV4FC", + "question_id": "NV4FC", + "video_file": "Charades/videos/NV4FC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting on their bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.1>." + } + ] + }, + { + "video_id": "NV4FC", + "question_id": "NV4FC", + "video_file": "Charades/videos/NV4FC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they take off their shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.2>." + } + ] + }, + { + "video_id": "NV4FC", + "question_id": "NV4FC", + "video_file": "Charades/videos/NV4FC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a book from the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.6> to <25.8>." + } + ] + }, + { + "video_id": "NV4FC", + "question_id": "NV4FC", + "video_file": "Charades/videos/NV4FC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they open the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.3> to <30.3>." + } + ] + }, + { + "video_id": "Q8QGT", + "question_id": "Q8QGT", + "video_file": "Charades/videos/Q8QGT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats the sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.5> to <13.9>." + } + ] + }, + { + "video_id": "809QT", + "question_id": "809QT", + "video_file": "Charades/videos/809QT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a book from it'." + }, + { + "from": "gpt", + "value": "From <3.0> to <8.0>." + } + ] + }, + { + "video_id": "803H6", + "question_id": "803H6", + "video_file": "Charades/videos/803H6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'opens the window another person comes in' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.3> to <12.3>." + } + ] + }, + { + "video_id": "803H6", + "question_id": "803H6", + "video_file": "Charades/videos/803H6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a cabinet door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.3> to <12.3>." + } + ] + }, + { + "video_id": "803H6", + "question_id": "803H6", + "video_file": "Charades/videos/803H6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws a pillow around complaining about the pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <15.9>." + } + ] + }, + { + "video_id": "803H6", + "question_id": "803H6", + "video_file": "Charades/videos/803H6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is opening a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.3> to <12.3>." + } + ] + }, + { + "video_id": "UG6NT", + "question_id": "UG6NT", + "video_file": "Charades/videos/UG6NT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person #2 throws a pillow across the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.1> to <8.3>." + } + ] + }, + { + "video_id": "UG6NT", + "question_id": "UG6NT", + "video_file": "Charades/videos/UG6NT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is throwing a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <8.3>." + } + ] + }, + { + "video_id": "TXJK7", + "question_id": "TXJK7", + "video_file": "Charades/videos/TXJK7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is fixing some string lights' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.2> to <11.9>." + } + ] + }, + { + "video_id": "TXJK7", + "question_id": "TXJK7", + "video_file": "Charades/videos/TXJK7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is fixing lights'." + }, + { + "from": "gpt", + "value": "From <0.2> to <11.9>." + } + ] + }, + { + "video_id": "TXJK7", + "question_id": "TXJK7", + "video_file": "Charades/videos/TXJK7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is fixing a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <11.9>." + } + ] + }, + { + "video_id": "MMV3D", + "question_id": "MMV3D", + "video_file": "Charades/videos/MMV3D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a blanket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "QNXGG", + "question_id": "QNXGG", + "video_file": "Charades/videos/QNXGG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is undressing near a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.8> to <11.4>." + } + ] + }, + { + "video_id": "QNXGG", + "question_id": "QNXGG", + "video_file": "Charades/videos/QNXGG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they begin pouring medicine into a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <22.0>." + } + ] + }, + { + "video_id": "QNXGG", + "question_id": "QNXGG", + "video_file": "Charades/videos/QNXGG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pour it into a glass of water'." + }, + { + "from": "gpt", + "value": "From <12.3> to <22.0>." + } + ] + }, + { + "video_id": "NJCDX", + "question_id": "NJCDX", + "video_file": "Charades/videos/NJCDX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they close the cabinet door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.4> to <32.0>." + } + ] + }, + { + "video_id": "NJCDX", + "question_id": "NJCDX", + "video_file": "Charades/videos/NJCDX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take food'." + }, + { + "from": "gpt", + "value": "From <12.9> to <25.4>." + } + ] + }, + { + "video_id": "NJCDX", + "question_id": "NJCDX", + "video_file": "Charades/videos/NJCDX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open the cabinet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <13.2>." + } + ] + }, + { + "video_id": "4ATDB", + "question_id": "4ATDB", + "video_file": "Charades/videos/4ATDB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns the light off' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.3> to <21.8>." + } + ] + }, + { + "video_id": "4ATDB", + "question_id": "4ATDB", + "video_file": "Charades/videos/4ATDB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns off a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.3> to <21.8>." + } + ] + }, + { + "video_id": "V4HBP", + "question_id": "V4HBP", + "video_file": "Charades/videos/V4HBP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person finally throwing a box onto the bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.2> to <30.3>." + } + ] + }, + { + "video_id": "V4HBP", + "question_id": "V4HBP", + "video_file": "Charades/videos/V4HBP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws another blanket off the bed,' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <19.3>." + } + ] + }, + { + "video_id": "V4HBP", + "question_id": "V4HBP", + "video_file": "Charades/videos/V4HBP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits on a bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.6> to <32.0>." + } + ] + }, + { + "video_id": "V4HBP", + "question_id": "V4HBP", + "video_file": "Charades/videos/V4HBP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting on a bed folding a towel' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.6> to <32.0>." + } + ] + }, + { + "video_id": "V4HBP", + "question_id": "V4HBP", + "video_file": "Charades/videos/V4HBP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they throw a blanket on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.2> to <19.3>." + } + ] + }, + { + "video_id": "O630X", + "question_id": "O630X", + "video_file": "Charades/videos/O630X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person suddenly awakens'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.1> to <34.9>." + } + ] + }, + { + "video_id": "Q5GVZ", + "question_id": "Q5GVZ", + "video_file": "Charades/videos/Q5GVZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a cup of coffee' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <15.0>." + } + ] + }, + { + "video_id": "Q5GVZ", + "question_id": "Q5GVZ", + "video_file": "Charades/videos/Q5GVZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the towel on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <16.4>." + } + ] + }, + { + "video_id": "Q5GVZ", + "question_id": "Q5GVZ", + "video_file": "Charades/videos/Q5GVZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts the towel onto a pile of clothes'." + }, + { + "from": "gpt", + "value": "From <11.0> to <16.4>." + } + ] + }, + { + "video_id": "UCDL4", + "question_id": "UCDL4", + "video_file": "Charades/videos/UCDL4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.1> to <6.2>." + } + ] + }, + { + "video_id": "UCDL4", + "question_id": "UCDL4", + "video_file": "Charades/videos/UCDL4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a picture off the wall' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <14.0>." + } + ] + }, + { + "video_id": "UCDL4", + "question_id": "UCDL4", + "video_file": "Charades/videos/UCDL4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes a picture off the wall' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <14.0>." + } + ] + }, + { + "video_id": "4UT8K", + "question_id": "4UT8K", + "video_file": "Charades/videos/4UT8K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take out their camera' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <24.5>." + } + ] + }, + { + "video_id": "0QHR4", + "question_id": "0QHR4", + "video_file": "Charades/videos/0QHR4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is opening the door to their closer' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "0QHR4", + "question_id": "0QHR4", + "video_file": "Charades/videos/0QHR4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opened the closet door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.1>." + } + ] + }, + { + "video_id": "0QHR4", + "question_id": "0QHR4", + "video_file": "Charades/videos/0QHR4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they eat some food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.3> to <8.9>." + } + ] + }, + { + "video_id": "VML1Z", + "question_id": "VML1Z", + "video_file": "Charades/videos/VML1Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "6AVDE", + "question_id": "6AVDE", + "video_file": "Charades/videos/6AVDE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating some food from a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "6AVDE", + "question_id": "6AVDE", + "video_file": "Charades/videos/6AVDE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person was eating some food'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "1K0GV", + "question_id": "1K0GV", + "video_file": "Charades/videos/1K0GV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a glass of water'." + }, + { + "from": "gpt", + "value": "From <10.8> to <15.5>." + } + ] + }, + { + "video_id": "1K0GV", + "question_id": "1K0GV", + "video_file": "Charades/videos/1K0GV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throw it on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.1> to <14.8>." + } + ] + }, + { + "video_id": "1K0GV", + "question_id": "1K0GV", + "video_file": "Charades/videos/1K0GV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is throwing trash in a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.1> to <14.8>." + } + ] + }, + { + "video_id": "IYD9U", + "question_id": "IYD9U", + "video_file": "Charades/videos/IYD9U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is holding their camera' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "IYD9U", + "question_id": "IYD9U", + "video_file": "Charades/videos/IYD9U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person is standing takes a couple pictures with his phone'." + }, + { + "from": "gpt", + "value": "From <3.9> to <10.0>." + } + ] + }, + { + "video_id": "IYD9U", + "question_id": "IYD9U", + "video_file": "Charades/videos/IYD9U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person runs to the blanket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.1> to <27.9>." + } + ] + }, + { + "video_id": "I5U4Q", + "question_id": "I5U4Q", + "video_file": "Charades/videos/I5U4Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person puts a camera on top of the fridge' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.1> to <20.8>." + } + ] + }, + { + "video_id": "I5U4Q", + "question_id": "I5U4Q", + "video_file": "Charades/videos/I5U4Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person puts a camera on a refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <20.8>." + } + ] + }, + { + "video_id": "Q1K00", + "question_id": "Q1K00", + "video_file": "Charades/videos/Q1K00.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they open a window'." + }, + { + "from": "gpt", + "value": "From <8.8> to <14.8>." + } + ] + }, + { + "video_id": "Q1K00", + "question_id": "Q1K00", + "video_file": "Charades/videos/Q1K00.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes off shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.6> to <8.3>." + } + ] + }, + { + "video_id": "Q1K00", + "question_id": "Q1K00", + "video_file": "Charades/videos/Q1K00.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person moves curtain looks out of the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.6> to <15.4>." + } + ] + }, + { + "video_id": "Q1K00", + "question_id": "Q1K00", + "video_file": "Charades/videos/Q1K00.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes off their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.6> to <8.3>." + } + ] + }, + { + "video_id": "Q1K00", + "question_id": "Q1K00", + "video_file": "Charades/videos/Q1K00.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "HIO3X", + "question_id": "HIO3X", + "video_file": "Charades/videos/HIO3X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person proceeds to wash their hands'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <34.6> to <48.2>." + } + ] + }, + { + "video_id": "HIO3X", + "question_id": "HIO3X", + "video_file": "Charades/videos/HIO3X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person wash their hands' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <34.6> to <48.2>." + } + ] + }, + { + "video_id": "4S3UZ", + "question_id": "4S3UZ", + "video_file": "Charades/videos/4S3UZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person watching tv smiling'." + }, + { + "from": "gpt", + "value": "From <14.3> to <23.2>." + } + ] + }, + { + "video_id": "4S3UZ", + "question_id": "4S3UZ", + "video_file": "Charades/videos/4S3UZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person is sitting down smiling at the television' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <23.2>." + } + ] + }, + { + "video_id": "E40Y8", + "question_id": "E40Y8", + "video_file": "Charades/videos/E40Y8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open a box of cereal' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.3> to <17.0>." + } + ] + }, + { + "video_id": "E40Y8", + "question_id": "E40Y8", + "video_file": "Charades/videos/E40Y8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a box of cookies' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.3> to <17.0>." + } + ] + }, + { + "video_id": "JN2E8", + "question_id": "JN2E8", + "video_file": "Charades/videos/JN2E8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person throws the blanket down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.8> to <19.2>." + } + ] + }, + { + "video_id": "JN2E8", + "question_id": "JN2E8", + "video_file": "Charades/videos/JN2E8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throw the blanket off of themself'." + }, + { + "from": "gpt", + "value": "From <11.8> to <19.2>." + } + ] + }, + { + "video_id": "2Q3ZJ", + "question_id": "2Q3ZJ", + "video_file": "Charades/videos/2Q3ZJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person goes into the closet talking on the phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.0> to <28.9>." + } + ] + }, + { + "video_id": "2Q3ZJ", + "question_id": "2Q3ZJ", + "video_file": "Charades/videos/2Q3ZJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person talking on the phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <28.9>." + } + ] + }, + { + "video_id": "K0IGF", + "question_id": "K0IGF", + "video_file": "Charades/videos/K0IGF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sneezing'." + }, + { + "from": "gpt", + "value": "From <2.5> to <11.4>." + } + ] + }, + { + "video_id": "OZ6SB", + "question_id": "OZ6SB", + "video_file": "Charades/videos/OZ6SB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'at this point the person had a sneezing fit'." + }, + { + "from": "gpt", + "value": "From <7.5> to <16.3>." + } + ] + }, + { + "video_id": "OZ6SB", + "question_id": "OZ6SB", + "video_file": "Charades/videos/OZ6SB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they begin sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <16.3>." + } + ] + }, + { + "video_id": "Z320B", + "question_id": "Z320B", + "video_file": "Charades/videos/Z320B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes a blanket' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.6> to <22.5>." + } + ] + }, + { + "video_id": "Z320B", + "question_id": "Z320B", + "video_file": "Charades/videos/Z320B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts broom in corner by door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.5> to <20.5>." + } + ] + }, + { + "video_id": "Z320B", + "question_id": "Z320B", + "video_file": "Charades/videos/Z320B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding broom'." + }, + { + "from": "gpt", + "value": "From <0.0> to <19.9>." + } + ] + }, + { + "video_id": "5AW54", + "question_id": "5AW54", + "video_file": "Charades/videos/5AW54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is holding a bag of clothes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.6>." + } + ] + }, + { + "video_id": "5AW54", + "question_id": "5AW54", + "video_file": "Charades/videos/5AW54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the bag in front of a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <10.1>." + } + ] + }, + { + "video_id": "7VZ5M", + "question_id": "7VZ5M", + "video_file": "Charades/videos/7VZ5M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sitting in a chair near a laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <8.3>." + } + ] + }, + { + "video_id": "DECCY", + "question_id": "DECCY", + "video_file": "Charades/videos/DECCY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person reading the book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.8> to <34.0>." + } + ] + }, + { + "video_id": "RSFBP", + "question_id": "RSFBP", + "video_file": "Charades/videos/RSFBP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is walking down the stair sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <12.2>." + } + ] + }, + { + "video_id": "RSFBP", + "question_id": "RSFBP", + "video_file": "Charades/videos/RSFBP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person looks out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.9> to <31.0>." + } + ] + }, + { + "video_id": "QSQK2", + "question_id": "QSQK2", + "video_file": "Charades/videos/QSQK2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.5> to <29.0>." + } + ] + }, + { + "video_id": "QSQK2", + "question_id": "QSQK2", + "video_file": "Charades/videos/QSQK2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person walks through the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.0> to <29.9>." + } + ] + }, + { + "video_id": "QSQK2", + "question_id": "QSQK2", + "video_file": "Charades/videos/QSQK2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs'." + }, + { + "from": "gpt", + "value": "From <7.0> to <12.7>." + } + ] + }, + { + "video_id": "QSQK2", + "question_id": "QSQK2", + "video_file": "Charades/videos/QSQK2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.2> to <26.8>." + } + ] + }, + { + "video_id": "3JSX7", + "question_id": "3JSX7", + "video_file": "Charades/videos/3JSX7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person works on a burnt-out light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.2> to <15.6>." + } + ] + }, + { + "video_id": "3JSX7", + "question_id": "3JSX7", + "video_file": "Charades/videos/3JSX7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns on a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.2> to <15.6>." + } + ] + }, + { + "video_id": "8U3SS", + "question_id": "8U3SS", + "video_file": "Charades/videos/8U3SS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they begin to laugh at each other' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <19.7>." + } + ] + }, + { + "video_id": "92QGQ", + "question_id": "92QGQ", + "video_file": "Charades/videos/92QGQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.6> to <31.0>." + } + ] + }, + { + "video_id": "92QGQ", + "question_id": "92QGQ", + "video_file": "Charades/videos/92QGQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a blanket' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "92QGQ", + "question_id": "92QGQ", + "video_file": "Charades/videos/92QGQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person goes to sit in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.6> to <31.0>." + } + ] + }, + { + "video_id": "92QGQ", + "question_id": "92QGQ", + "video_file": "Charades/videos/92QGQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a blanket out'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "92QGQ", + "question_id": "92QGQ", + "video_file": "Charades/videos/92QGQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sit in a chair'." + }, + { + "from": "gpt", + "value": "From <22.6> to <31.0>." + } + ] + }, + { + "video_id": "92QGQ", + "question_id": "92QGQ", + "video_file": "Charades/videos/92QGQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person puts clothes in a dryer' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.4> to <12.9>." + } + ] + }, + { + "video_id": "92QGQ", + "question_id": "92QGQ", + "video_file": "Charades/videos/92QGQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person takes a blanket out of the dryer' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "92QGQ", + "question_id": "92QGQ", + "video_file": "Charades/videos/92QGQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person snuggle into the blanket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.1> to <31.0>." + } + ] + }, + { + "video_id": "92QGQ", + "question_id": "92QGQ", + "video_file": "Charades/videos/92QGQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts in some clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <12.9>." + } + ] + }, + { + "video_id": "3TUUI", + "question_id": "3TUUI", + "video_file": "Charades/videos/3TUUI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begins putting their shoes on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.1> to <34.2>." + } + ] + }, + { + "video_id": "A6NH2", + "question_id": "A6NH2", + "video_file": "Charades/videos/A6NH2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person in the kitchen opens a small closet door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <20.0>." + } + ] + }, + { + "video_id": "R2ZYS", + "question_id": "R2ZYS", + "video_file": "Charades/videos/R2ZYS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking cup of coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.1> to <18.8>." + } + ] + }, + { + "video_id": "R2ZYS", + "question_id": "R2ZYS", + "video_file": "Charades/videos/R2ZYS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a drink out of a glass'." + }, + { + "from": "gpt", + "value": "From <5.1> to <18.8>." + } + ] + }, + { + "video_id": "1OIPY", + "question_id": "1OIPY", + "video_file": "Charades/videos/1OIPY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person sits in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "1OIPY", + "question_id": "1OIPY", + "video_file": "Charades/videos/1OIPY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.8> to <30.9>." + } + ] + }, + { + "video_id": "1OIPY", + "question_id": "1OIPY", + "video_file": "Charades/videos/1OIPY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person someone is sitting on a chair in the kitchen' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "LRLNC", + "question_id": "LRLNC", + "video_file": "Charades/videos/LRLNC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the phone down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.7> to <31.0>." + } + ] + }, + { + "video_id": "LRLNC", + "question_id": "LRLNC", + "video_file": "Charades/videos/LRLNC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts the phone down on the bed'." + }, + { + "from": "gpt", + "value": "From <27.7> to <31.0>." + } + ] + }, + { + "video_id": "1E7VU", + "question_id": "1E7VU", + "video_file": "Charades/videos/1E7VU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is putting dishes away' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.1> to <18.5>." + } + ] + }, + { + "video_id": "1E7VU", + "question_id": "1E7VU", + "video_file": "Charades/videos/1E7VU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes some dishes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.7> to <16.5>." + } + ] + }, + { + "video_id": "1E7VU", + "question_id": "1E7VU", + "video_file": "Charades/videos/1E7VU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they start sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <32.0>." + } + ] + }, + { + "video_id": "1E7VU", + "question_id": "1E7VU", + "video_file": "Charades/videos/1E7VU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.5> to <32.0>." + } + ] + }, + { + "video_id": "1E7VU", + "question_id": "1E7VU", + "video_file": "Charades/videos/1E7VU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting away dishes'." + }, + { + "from": "gpt", + "value": "From <3.1> to <18.5>." + } + ] + }, + { + "video_id": "1MD1H", + "question_id": "1MD1H", + "video_file": "Charades/videos/1MD1H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person awakens on the sofa in a living room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <13.5>." + } + ] + }, + { + "video_id": "1MD1H", + "question_id": "1MD1H", + "video_file": "Charades/videos/1MD1H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they sit back down on the couch' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.4> to <33.0>." + } + ] + }, + { + "video_id": "1MD1H", + "question_id": "1MD1H", + "video_file": "Charades/videos/1MD1H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pillow they wake up take off their shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <19.2>." + } + ] + }, + { + "video_id": "FNLLG", + "question_id": "FNLLG", + "video_file": "Charades/videos/FNLLG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person looks at a picture on the wall' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.3> to <13.0>." + } + ] + }, + { + "video_id": "ZYDGF", + "question_id": "ZYDGF", + "video_file": "Charades/videos/ZYDGF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a bag out'." + }, + { + "from": "gpt", + "value": "From <13.2> to <20.4>." + } + ] + }, + { + "video_id": "AFAXI", + "question_id": "AFAXI", + "video_file": "Charades/videos/AFAXI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person who is lying down awakens'." + }, + { + "from": "gpt", + "value": "From <9.2> to <19.3>." + } + ] + }, + { + "video_id": "AFAXI", + "question_id": "AFAXI", + "video_file": "Charades/videos/AFAXI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sitting on the bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.7> to <28.0>." + } + ] + }, + { + "video_id": "AFAXI", + "question_id": "AFAXI", + "video_file": "Charades/videos/AFAXI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is lying on a bed they sit up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.7> to <28.0>." + } + ] + }, + { + "video_id": "APQSV", + "question_id": "APQSV", + "video_file": "Charades/videos/APQSV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens another door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.6> to <9.7>." + } + ] + }, + { + "video_id": "APQSV", + "question_id": "APQSV", + "video_file": "Charades/videos/APQSV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <13.8>." + } + ] + }, + { + "video_id": "APQSV", + "question_id": "APQSV", + "video_file": "Charades/videos/APQSV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the towel on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.5> to <24.6>." + } + ] + }, + { + "video_id": "APQSV", + "question_id": "APQSV", + "video_file": "Charades/videos/APQSV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a smiling person opens the door to their basement' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.6> to <9.7>." + } + ] + }, + { + "video_id": "DA7E9", + "question_id": "DA7E9", + "video_file": "Charades/videos/DA7E9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is cooking food on a stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.2> to <13.2>." + } + ] + }, + { + "video_id": "DA7E9", + "question_id": "DA7E9", + "video_file": "Charades/videos/DA7E9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person reads the book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.9> to <25.8>." + } + ] + }, + { + "video_id": "DA7E9", + "question_id": "DA7E9", + "video_file": "Charades/videos/DA7E9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person continues cooking' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.2> to <13.2>." + } + ] + }, + { + "video_id": "DA7E9", + "question_id": "DA7E9", + "video_file": "Charades/videos/DA7E9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person a cookbook'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.2> to <13.2>." + } + ] + }, + { + "video_id": "DA7E9", + "question_id": "DA7E9", + "video_file": "Charades/videos/DA7E9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is cooking at the stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.2> to <13.2>." + } + ] + }, + { + "video_id": "ATWCT", + "question_id": "ATWCT", + "video_file": "Charades/videos/ATWCT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <30.0>." + } + ] + }, + { + "video_id": "YXJBA", + "question_id": "YXJBA", + "video_file": "Charades/videos/YXJBA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person reading a book that is put on a table'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.8>." + } + ] + }, + { + "video_id": "O18DS", + "question_id": "O18DS", + "video_file": "Charades/videos/O18DS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.2> to <13.1>." + } + ] + }, + { + "video_id": "HP94W", + "question_id": "HP94W", + "video_file": "Charades/videos/HP94W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person is putting phone on shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <11.2>." + } + ] + }, + { + "video_id": "YUOQW", + "question_id": "YUOQW", + "video_file": "Charades/videos/YUOQW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is holding a phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <12.4>." + } + ] + }, + { + "video_id": "HEZBG", + "question_id": "HEZBG", + "video_file": "Charades/videos/HEZBG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins cooking at the stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <35.0>." + } + ] + }, + { + "video_id": "MNBEI", + "question_id": "MNBEI", + "video_file": "Charades/videos/MNBEI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.3> to <21.4>." + } + ] + }, + { + "video_id": "MNBEI", + "question_id": "MNBEI", + "video_file": "Charades/videos/MNBEI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "MNBEI", + "question_id": "MNBEI", + "video_file": "Charades/videos/MNBEI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'this person opens door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "MNBEI", + "question_id": "MNBEI", + "video_file": "Charades/videos/MNBEI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits in a yellow chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <10.2>." + } + ] + }, + { + "video_id": "MNBEI", + "question_id": "MNBEI", + "video_file": "Charades/videos/MNBEI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walk towards the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "MNBEI", + "question_id": "MNBEI", + "video_file": "Charades/videos/MNBEI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person proceeds to sit in the chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.4> to <10.2>." + } + ] + }, + { + "video_id": "MNBEI", + "question_id": "MNBEI", + "video_file": "Charades/videos/MNBEI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting on their shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.5> to <18.4>." + } + ] + }, + { + "video_id": "MNBEI", + "question_id": "MNBEI", + "video_file": "Charades/videos/MNBEI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts on his shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.5> to <18.4>." + } + ] + }, + { + "video_id": "MNBEI", + "question_id": "MNBEI", + "video_file": "Charades/videos/MNBEI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "Z0Q5D", + "question_id": "Z0Q5D", + "video_file": "Charades/videos/Z0Q5D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts the phone down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.2> to <18.3>." + } + ] + }, + { + "video_id": "APJQ8", + "question_id": "APJQ8", + "video_file": "Charades/videos/APJQ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts the blanket into the washing machine'." + }, + { + "from": "gpt", + "value": "From <4.2> to <11.1>." + } + ] + }, + { + "video_id": "APJQ8", + "question_id": "APJQ8", + "video_file": "Charades/videos/APJQ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts a blanket into a washer' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <11.1>." + } + ] + }, + { + "video_id": "R5K2K", + "question_id": "R5K2K", + "video_file": "Charades/videos/R5K2K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person is smiling at the clothes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <7.4>." + } + ] + }, + { + "video_id": "6BFKO", + "question_id": "6BFKO", + "video_file": "Charades/videos/6BFKO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person the laptop lights up'." + }, + { + "from": "gpt", + "value": "From <23.8> to <29.0>." + } + ] + }, + { + "video_id": "25BZ1", + "question_id": "25BZ1", + "video_file": "Charades/videos/25BZ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs out with their homework' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.3> to <25.0>." + } + ] + }, + { + "video_id": "NXC5X", + "question_id": "NXC5X", + "video_file": "Charades/videos/NXC5X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks through the doorway holding a towel' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "NXC5X", + "question_id": "NXC5X", + "video_file": "Charades/videos/NXC5X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts eating an apple' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <21.4>." + } + ] + }, + { + "video_id": "NXC5X", + "question_id": "NXC5X", + "video_file": "Charades/videos/NXC5X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begin eating food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <21.4>." + } + ] + }, + { + "video_id": "NXC5X", + "question_id": "NXC5X", + "video_file": "Charades/videos/NXC5X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating something' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <21.4>." + } + ] + }, + { + "video_id": "F3RM9", + "question_id": "F3RM9", + "video_file": "Charades/videos/F3RM9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person is smiling at a picture' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.2> to <21.9>." + } + ] + }, + { + "video_id": "F3RM9", + "question_id": "F3RM9", + "video_file": "Charades/videos/F3RM9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looking at a picture'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "VTICD", + "question_id": "VTICD", + "video_file": "Charades/videos/VTICD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person cooks on the stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "VTICD", + "question_id": "VTICD", + "video_file": "Charades/videos/VTICD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is cooking at a stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "HGLHS", + "question_id": "HGLHS", + "video_file": "Charades/videos/HGLHS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks past a doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <6.7>." + } + ] + }, + { + "video_id": "HGLHS", + "question_id": "HGLHS", + "video_file": "Charades/videos/HGLHS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person begins laughing'." + }, + { + "from": "gpt", + "value": "From <7.0> to <17.5>." + } + ] + }, + { + "video_id": "J40UT", + "question_id": "J40UT", + "video_file": "Charades/videos/J40UT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person in a bathroom is holding a glass'." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.1>." + } + ] + }, + { + "video_id": "J40UT", + "question_id": "J40UT", + "video_file": "Charades/videos/J40UT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is drinking a glass of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "MF8AX", + "question_id": "MF8AX", + "video_file": "Charades/videos/MF8AX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is running into a kitchen' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "MF8AX", + "question_id": "MF8AX", + "video_file": "Charades/videos/MF8AX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs into a room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "6PN8O", + "question_id": "6PN8O", + "video_file": "Charades/videos/6PN8O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turn off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.0> to <30.9>." + } + ] + }, + { + "video_id": "6PN8O", + "question_id": "6PN8O", + "video_file": "Charades/videos/6PN8O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is in the bathroom holding a towel' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "NQ1V6", + "question_id": "NQ1V6", + "video_file": "Charades/videos/NQ1V6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts undressing to change clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.6> to <27.2>." + } + ] + }, + { + "video_id": "NQ1V6", + "question_id": "NQ1V6", + "video_file": "Charades/videos/NQ1V6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts on different clothes'." + }, + { + "from": "gpt", + "value": "From <12.3> to <16.9>." + } + ] + }, + { + "video_id": "8YZ8X", + "question_id": "8YZ8X", + "video_file": "Charades/videos/8YZ8X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns off the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.5> to <29.9>." + } + ] + }, + { + "video_id": "9ZQ0R", + "question_id": "9ZQ0R", + "video_file": "Charades/videos/9ZQ0R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take a drink of their glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.0> to <13.8>." + } + ] + }, + { + "video_id": "X6D8E", + "question_id": "X6D8E", + "video_file": "Charades/videos/X6D8E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pours themselves a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.4> to <30.5>." + } + ] + }, + { + "video_id": "909RL", + "question_id": "909RL", + "video_file": "Charades/videos/909RL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking a glass of water'." + }, + { + "from": "gpt", + "value": "From <13.2> to <19.1>." + } + ] + }, + { + "video_id": "909RL", + "question_id": "909RL", + "video_file": "Charades/videos/909RL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put the glass of water onto a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <20.4>." + } + ] + }, + { + "video_id": "909RL", + "question_id": "909RL", + "video_file": "Charades/videos/909RL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looks out window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.5> to <30.0>." + } + ] + }, + { + "video_id": "909RL", + "question_id": "909RL", + "video_file": "Charades/videos/909RL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinks from glass of water places glass onto table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.2> to <19.1>." + } + ] + }, + { + "video_id": "UPXBJ", + "question_id": "UPXBJ", + "video_file": "Charades/videos/UPXBJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <14.9>." + } + ] + }, + { + "video_id": "UPXBJ", + "question_id": "UPXBJ", + "video_file": "Charades/videos/UPXBJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person opens a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.0> to <14.9>." + } + ] + }, + { + "video_id": "9GMHF", + "question_id": "9GMHF", + "video_file": "Charades/videos/9GMHF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens another door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.8> to <18.7>." + } + ] + }, + { + "video_id": "9GMHF", + "question_id": "9GMHF", + "video_file": "Charades/videos/9GMHF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person opens the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <18.7>." + } + ] + }, + { + "video_id": "9GMHF", + "question_id": "9GMHF", + "video_file": "Charades/videos/9GMHF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes out a broom'." + }, + { + "from": "gpt", + "value": "From <14.3> to <19.9>." + } + ] + }, + { + "video_id": "Z7O9G", + "question_id": "Z7O9G", + "video_file": "Charades/videos/Z7O9G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a camera' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <18.1>." + } + ] + }, + { + "video_id": "Z7O9G", + "question_id": "Z7O9G", + "video_file": "Charades/videos/Z7O9G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a drink from a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.1> to <30.1>." + } + ] + }, + { + "video_id": "OJ4IE", + "question_id": "OJ4IE", + "video_file": "Charades/videos/OJ4IE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the 2nd person opens the door'." + }, + { + "from": "gpt", + "value": "From <16.9> to <25.4>." + } + ] + }, + { + "video_id": "OJ4IE", + "question_id": "OJ4IE", + "video_file": "Charades/videos/OJ4IE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding clothes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.6> to <26.9>." + } + ] + }, + { + "video_id": "MSW4E", + "question_id": "MSW4E", + "video_file": "Charades/videos/MSW4E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person is also eating some food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.8> to <19.3>." + } + ] + }, + { + "video_id": "XYTEU", + "question_id": "XYTEU", + "video_file": "Charades/videos/XYTEU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes a pile of clothes from a desk' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.1> to <14.1>." + } + ] + }, + { + "video_id": "CUII4", + "question_id": "CUII4", + "video_file": "Charades/videos/CUII4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "CUII4", + "question_id": "CUII4", + "video_file": "Charades/videos/CUII4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person undress the same sweater' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.8> to <34.8>." + } + ] + }, + { + "video_id": "CUII4", + "question_id": "CUII4", + "video_file": "Charades/videos/CUII4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throw it on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.3> to <13.7>." + } + ] + }, + { + "video_id": "CUII4", + "question_id": "CUII4", + "video_file": "Charades/videos/CUII4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes out some clothes from it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <13.1>." + } + ] + }, + { + "video_id": "CUII4", + "question_id": "CUII4", + "video_file": "Charades/videos/CUII4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes some clothes out of it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.4> to <13.1>." + } + ] + }, + { + "video_id": "DYVUY", + "question_id": "DYVUY", + "video_file": "Charades/videos/DYVUY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.8> to <31.0>." + } + ] + }, + { + "video_id": "DYVUY", + "question_id": "DYVUY", + "video_file": "Charades/videos/DYVUY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns the light off behind them' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.6> to <28.0>." + } + ] + }, + { + "video_id": "DYVUY", + "question_id": "DYVUY", + "video_file": "Charades/videos/DYVUY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.6> to <28.0>." + } + ] + }, + { + "video_id": "DYVUY", + "question_id": "DYVUY", + "video_file": "Charades/videos/DYVUY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person shutting off the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.6> to <28.0>." + } + ] + }, + { + "video_id": "BLPAB", + "question_id": "BLPAB", + "video_file": "Charades/videos/BLPAB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pours a glass of juice' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.8> to <31.0>." + } + ] + }, + { + "video_id": "BLPAB", + "question_id": "BLPAB", + "video_file": "Charades/videos/BLPAB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is pouring something into a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.8> to <31.0>." + } + ] + }, + { + "video_id": "6LGTT", + "question_id": "6LGTT", + "video_file": "Charades/videos/6LGTT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'i person is walking into the pantry smiling'." + }, + { + "from": "gpt", + "value": "From <9.0> to <16.9>." + } + ] + }, + { + "video_id": "6LGTT", + "question_id": "6LGTT", + "video_file": "Charades/videos/6LGTT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.0> to <10.9>." + } + ] + }, + { + "video_id": "6LGTT", + "question_id": "6LGTT", + "video_file": "Charades/videos/6LGTT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person close the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.7> to <22.2>." + } + ] + }, + { + "video_id": "KFOX4", + "question_id": "KFOX4", + "video_file": "Charades/videos/KFOX4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is holding a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.8> to <17.8>." + } + ] + }, + { + "video_id": "KFOX4", + "question_id": "KFOX4", + "video_file": "Charades/videos/KFOX4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person the start throwing a blanket around' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.8> to <19.0>." + } + ] + }, + { + "video_id": "KFOX4", + "question_id": "KFOX4", + "video_file": "Charades/videos/KFOX4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person throws the blanket on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.8> to <19.0>." + } + ] + }, + { + "video_id": "AHLZY", + "question_id": "AHLZY", + "video_file": "Charades/videos/AHLZY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person start sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.4> to <21.2>." + } + ] + }, + { + "video_id": "3JJ7C", + "question_id": "3JJ7C", + "video_file": "Charades/videos/3JJ7C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person tidy's up a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "ZG1EM", + "question_id": "ZG1EM", + "video_file": "Charades/videos/ZG1EM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'next the person sits on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.8> to <26.7>." + } + ] + }, + { + "video_id": "ZG1EM", + "question_id": "ZG1EM", + "video_file": "Charades/videos/ZG1EM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits down on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.8> to <26.7>." + } + ] + }, + { + "video_id": "ZG1EM", + "question_id": "ZG1EM", + "video_file": "Charades/videos/ZG1EM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinks from a glass'." + }, + { + "from": "gpt", + "value": "From <23.4> to <32.0>." + } + ] + }, + { + "video_id": "ZG1EM", + "question_id": "ZG1EM", + "video_file": "Charades/videos/ZG1EM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person undresses by taking the jacket back off' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.0>." + } + ] + }, + { + "video_id": "ZG1EM", + "question_id": "ZG1EM", + "video_file": "Charades/videos/ZG1EM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person dressing by putting a jacket on' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.0>." + } + ] + }, + { + "video_id": "UMT5R", + "question_id": "UMT5R", + "video_file": "Charades/videos/UMT5R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put it back on the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.5> to <31.0>." + } + ] + }, + { + "video_id": "6BUU6", + "question_id": "6BUU6", + "video_file": "Charades/videos/6BUU6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is closing the hallway door grasping the doorknob' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.2>." + } + ] + }, + { + "video_id": "6BUU6", + "question_id": "6BUU6", + "video_file": "Charades/videos/6BUU6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a laptop opens' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <29.2> to <32.0>." + } + ] + }, + { + "video_id": "6BUU6", + "question_id": "6BUU6", + "video_file": "Charades/videos/6BUU6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.2>." + } + ] + }, + { + "video_id": "29UVO", + "question_id": "29UVO", + "video_file": "Charades/videos/29UVO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting on the couch snuggling a blanket'." + }, + { + "from": "gpt", + "value": "From <0.8> to <12.3>." + } + ] + }, + { + "video_id": "29UVO", + "question_id": "29UVO", + "video_file": "Charades/videos/29UVO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person is snuggling with a blanket on the sofa'." + }, + { + "from": "gpt", + "value": "From <0.8> to <12.3>." + } + ] + }, + { + "video_id": "H69BE", + "question_id": "H69BE", + "video_file": "Charades/videos/H69BE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person closes the laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <27.8>." + } + ] + }, + { + "video_id": "S3TR3", + "question_id": "S3TR3", + "video_file": "Charades/videos/S3TR3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.5> to <33.2>." + } + ] + }, + { + "video_id": "S3TR3", + "question_id": "S3TR3", + "video_file": "Charades/videos/S3TR3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take their shoes off'." + }, + { + "from": "gpt", + "value": "From <21.5> to <33.2>." + } + ] + }, + { + "video_id": "E7CJ0", + "question_id": "E7CJ0", + "video_file": "Charades/videos/E7CJ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person is sneezing into a blanket' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.9> to <24.7>." + } + ] + }, + { + "video_id": "T38JM", + "question_id": "T38JM", + "video_file": "Charades/videos/T38JM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begin laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.3> to <23.2>." + } + ] + }, + { + "video_id": "T38JM", + "question_id": "T38JM", + "video_file": "Charades/videos/T38JM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the broom down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.2> to <15.3>." + } + ] + }, + { + "video_id": "ON2VH", + "question_id": "ON2VH", + "video_file": "Charades/videos/ON2VH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.1> to <23.7>." + } + ] + }, + { + "video_id": "03M0K", + "question_id": "03M0K", + "video_file": "Charades/videos/03M0K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <13.9>." + } + ] + }, + { + "video_id": "03M0K", + "question_id": "03M0K", + "video_file": "Charades/videos/03M0K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is dressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.2> to <9.9>." + } + ] + }, + { + "video_id": "03M0K", + "question_id": "03M0K", + "video_file": "Charades/videos/03M0K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens their laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.5> to <13.9>." + } + ] + }, + { + "video_id": "BNSQN", + "question_id": "BNSQN", + "video_file": "Charades/videos/BNSQN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks with it through a doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <28.0>." + } + ] + }, + { + "video_id": "BNSQN", + "question_id": "BNSQN", + "video_file": "Charades/videos/BNSQN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is walking through the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.3> to <28.0>." + } + ] + }, + { + "video_id": "V3DJ5", + "question_id": "V3DJ5", + "video_file": "Charades/videos/V3DJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <27.0>." + } + ] + }, + { + "video_id": "V3DJ5", + "question_id": "V3DJ5", + "video_file": "Charades/videos/V3DJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a piece of food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.2> to <24.2>." + } + ] + }, + { + "video_id": "0UQU6", + "question_id": "0UQU6", + "video_file": "Charades/videos/0UQU6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person awakens in bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <17.0>." + } + ] + }, + { + "video_id": "0UQU6", + "question_id": "0UQU6", + "video_file": "Charades/videos/0UQU6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walks through a doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <18.0>." + } + ] + }, + { + "video_id": "493WS", + "question_id": "493WS", + "video_file": "Charades/videos/493WS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the refrigerator door'." + }, + { + "from": "gpt", + "value": "From <7.5> to <13.4>." + } + ] + }, + { + "video_id": "4BZI6", + "question_id": "4BZI6", + "video_file": "Charades/videos/4BZI6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a box'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "AR4S6", + "question_id": "AR4S6", + "video_file": "Charades/videos/AR4S6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the basement door behind them' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.6> to <28.3>." + } + ] + }, + { + "video_id": "TXNDA", + "question_id": "TXNDA", + "video_file": "Charades/videos/TXNDA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person laughing at whats outside the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <12.6>." + } + ] + }, + { + "video_id": "T1FTD", + "question_id": "T1FTD", + "video_file": "Charades/videos/T1FTD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person putting it on a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.4> to <33.4>." + } + ] + }, + { + "video_id": "T1FTD", + "question_id": "T1FTD", + "video_file": "Charades/videos/T1FTD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person closes the dryer door'." + }, + { + "from": "gpt", + "value": "From <29.6> to <35.0>." + } + ] + }, + { + "video_id": "T1FTD", + "question_id": "T1FTD", + "video_file": "Charades/videos/T1FTD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the pillow onto a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.5> to <33.4>." + } + ] + }, + { + "video_id": "DG5TX", + "question_id": "DG5TX", + "video_file": "Charades/videos/DG5TX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits down in a chair'." + }, + { + "from": "gpt", + "value": "From <17.9> to <23.4>." + } + ] + }, + { + "video_id": "DG5TX", + "question_id": "DG5TX", + "video_file": "Charades/videos/DG5TX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins working on some paperwork' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.6> to <31.0>." + } + ] + }, + { + "video_id": "E8C24", + "question_id": "E8C24", + "video_file": "Charades/videos/E8C24.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs into the kitchen holding a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <8.9>." + } + ] + }, + { + "video_id": "EJIKG", + "question_id": "EJIKG", + "video_file": "Charades/videos/EJIKG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a book from a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.8> to <16.9>." + } + ] + }, + { + "video_id": "8X49S", + "question_id": "8X49S", + "video_file": "Charades/videos/8X49S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.4> to <8.4>." + } + ] + }, + { + "video_id": "8X49S", + "question_id": "8X49S", + "video_file": "Charades/videos/8X49S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <13.7>." + } + ] + }, + { + "video_id": "8X49S", + "question_id": "8X49S", + "video_file": "Charades/videos/8X49S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they eat some food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <8.4>." + } + ] + }, + { + "video_id": "U3T5A", + "question_id": "U3T5A", + "video_file": "Charades/videos/U3T5A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opened the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.3> to <9.5>." + } + ] + }, + { + "video_id": "UGWOG", + "question_id": "UGWOG", + "video_file": "Charades/videos/UGWOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a book off of the desk'." + }, + { + "from": "gpt", + "value": "From <52.0> to <55.0>." + } + ] + }, + { + "video_id": "UGWOG", + "question_id": "UGWOG", + "video_file": "Charades/videos/UGWOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person turns on a light'." + }, + { + "from": "gpt", + "value": "From <10.7> to <20.1>." + } + ] + }, + { + "video_id": "UGWOG", + "question_id": "UGWOG", + "video_file": "Charades/videos/UGWOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person turns on the lights' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.7> to <20.1>." + } + ] + }, + { + "video_id": "VT64S", + "question_id": "VT64S", + "video_file": "Charades/videos/VT64S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes a box of food from the table'." + }, + { + "from": "gpt", + "value": "From <3.5> to <10.0>." + } + ] + }, + { + "video_id": "DSP2U", + "question_id": "DSP2U", + "video_file": "Charades/videos/DSP2U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.8> to <22.0>." + } + ] + }, + { + "video_id": "DSP2U", + "question_id": "DSP2U", + "video_file": "Charades/videos/DSP2U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits down in the chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.8> to <22.0>." + } + ] + }, + { + "video_id": "LRJS4", + "question_id": "LRJS4", + "video_file": "Charades/videos/LRJS4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.2>." + } + ] + }, + { + "video_id": "QEFJF", + "question_id": "QEFJF", + "video_file": "Charades/videos/QEFJF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating from a plate' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.4>." + } + ] + }, + { + "video_id": "QEFJF", + "question_id": "QEFJF", + "video_file": "Charades/videos/QEFJF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take their dishes'." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.8>." + } + ] + }, + { + "video_id": "QEFJF", + "question_id": "QEFJF", + "video_file": "Charades/videos/QEFJF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating in the dining room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.4>." + } + ] + }, + { + "video_id": "QEFJF", + "question_id": "QEFJF", + "video_file": "Charades/videos/QEFJF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person is eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.4>." + } + ] + }, + { + "video_id": "SAT9P", + "question_id": "SAT9P", + "video_file": "Charades/videos/SAT9P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person reads a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.4> to <29.0>." + } + ] + }, + { + "video_id": "SAT9P", + "question_id": "SAT9P", + "video_file": "Charades/videos/SAT9P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a man looks out a bathroom window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.4>." + } + ] + }, + { + "video_id": "SAT9P", + "question_id": "SAT9P", + "video_file": "Charades/videos/SAT9P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.4> to <29.0>." + } + ] + }, + { + "video_id": "SAT9P", + "question_id": "SAT9P", + "video_file": "Charades/videos/SAT9P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take out a blanket'." + }, + { + "from": "gpt", + "value": "From <11.3> to <22.0>." + } + ] + }, + { + "video_id": "LSY08", + "question_id": "LSY08", + "video_file": "Charades/videos/LSY08.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts down the stack of dishes on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <25.4>." + } + ] + }, + { + "video_id": "TQXMV", + "question_id": "TQXMV", + "video_file": "Charades/videos/TQXMV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.6> to <14.7>." + } + ] + }, + { + "video_id": "M5W51", + "question_id": "M5W51", + "video_file": "Charades/videos/M5W51.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes food that is next to them' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.0> to <31.0>." + } + ] + }, + { + "video_id": "M5W51", + "question_id": "M5W51", + "video_file": "Charades/videos/M5W51.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sleeping at a table awakens' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.3> to <26.9>." + } + ] + }, + { + "video_id": "WRW7O", + "question_id": "WRW7O", + "video_file": "Charades/videos/WRW7O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts some food onto a dish'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <21.4>." + } + ] + }, + { + "video_id": "WRW7O", + "question_id": "WRW7O", + "video_file": "Charades/videos/WRW7O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs out' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <26.0>." + } + ] + }, + { + "video_id": "W9XJX", + "question_id": "W9XJX", + "video_file": "Charades/videos/W9XJX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person puts down a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.3> to <9.8>." + } + ] + }, + { + "video_id": "OOE32", + "question_id": "OOE32", + "video_file": "Charades/videos/OOE32.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person is sneezing'." + }, + { + "from": "gpt", + "value": "From <12.8> to <18.8>." + } + ] + }, + { + "video_id": "VOTA7", + "question_id": "VOTA7", + "video_file": "Charades/videos/VOTA7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes the camera' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <14.3>." + } + ] + }, + { + "video_id": "VOTA7", + "question_id": "VOTA7", + "video_file": "Charades/videos/VOTA7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person runs in'." + }, + { + "from": "gpt", + "value": "From <9.8> to <17.0>." + } + ] + }, + { + "video_id": "VOTA7", + "question_id": "VOTA7", + "video_file": "Charades/videos/VOTA7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person runs out'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <17.0>." + } + ] + }, + { + "video_id": "7XJLH", + "question_id": "7XJLH", + "video_file": "Charades/videos/7XJLH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a cup'." + }, + { + "from": "gpt", + "value": "From <16.5> to <22.3>." + } + ] + }, + { + "video_id": "7XJLH", + "question_id": "7XJLH", + "video_file": "Charades/videos/7XJLH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes a drink from a glass of water'." + }, + { + "from": "gpt", + "value": "From <16.9> to <24.1>." + } + ] + }, + { + "video_id": "53FPM", + "question_id": "53FPM", + "video_file": "Charades/videos/53FPM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes some medicine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.0> to <38.4>." + } + ] + }, + { + "video_id": "3CTZL", + "question_id": "3CTZL", + "video_file": "Charades/videos/3CTZL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting on some clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.5> to <15.7>." + } + ] + }, + { + "video_id": "FYP5L", + "question_id": "FYP5L", + "video_file": "Charades/videos/FYP5L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take some shoes out from behind a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.4> to <32.0>." + } + ] + }, + { + "video_id": "FYP5L", + "question_id": "FYP5L", + "video_file": "Charades/videos/FYP5L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes shoes down from a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.4> to <32.0>." + } + ] + }, + { + "video_id": "FYP5L", + "question_id": "FYP5L", + "video_file": "Charades/videos/FYP5L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'there is a person running down a hall' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.2>." + } + ] + }, + { + "video_id": "FYP5L", + "question_id": "FYP5L", + "video_file": "Charades/videos/FYP5L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs in laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.2>." + } + ] + }, + { + "video_id": "FYP5L", + "question_id": "FYP5L", + "video_file": "Charades/videos/FYP5L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs down a hall'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.2>." + } + ] + }, + { + "video_id": "ZY2B1", + "question_id": "ZY2B1", + "video_file": "Charades/videos/ZY2B1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person the throws the clothes at a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.6> to <24.7>." + } + ] + }, + { + "video_id": "857LD", + "question_id": "857LD", + "video_file": "Charades/videos/857LD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <17.3>." + } + ] + }, + { + "video_id": "857LD", + "question_id": "857LD", + "video_file": "Charades/videos/857LD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they start closing the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.8> to <17.3>." + } + ] + }, + { + "video_id": "QNVMD", + "question_id": "QNVMD", + "video_file": "Charades/videos/QNVMD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts a pillow into a grocery bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.8> to <26.1>." + } + ] + }, + { + "video_id": "DK59J", + "question_id": "DK59J", + "video_file": "Charades/videos/DK59J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.3> to <25.4>." + } + ] + }, + { + "video_id": "DK59J", + "question_id": "DK59J", + "video_file": "Charades/videos/DK59J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens the book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <25.4>." + } + ] + }, + { + "video_id": "G5LEA", + "question_id": "G5LEA", + "video_file": "Charades/videos/G5LEA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opened the cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.5> to <13.0>." + } + ] + }, + { + "video_id": "G5LEA", + "question_id": "G5LEA", + "video_file": "Charades/videos/G5LEA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person closes the cabinet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.1> to <22.3>." + } + ] + }, + { + "video_id": "C1A2T", + "question_id": "C1A2T", + "video_file": "Charades/videos/C1A2T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens the bedroom door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "5V8WH", + "question_id": "5V8WH", + "video_file": "Charades/videos/5V8WH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks through a doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.7> to <6.6>." + } + ] + }, + { + "video_id": "5IDF4", + "question_id": "5IDF4", + "video_file": "Charades/videos/5IDF4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person throws the book on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.7> to <16.0>." + } + ] + }, + { + "video_id": "5IDF4", + "question_id": "5IDF4", + "video_file": "Charades/videos/5IDF4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs down the hallway holding a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.8> to <14.4>." + } + ] + }, + { + "video_id": "5IDF4", + "question_id": "5IDF4", + "video_file": "Charades/videos/5IDF4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a man runs into a room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.8> to <14.4>." + } + ] + }, + { + "video_id": "5IDF4", + "question_id": "5IDF4", + "video_file": "Charades/videos/5IDF4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws a book down onto the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.7> to <16.0>." + } + ] + }, + { + "video_id": "5O31N", + "question_id": "5O31N", + "video_file": "Charades/videos/5O31N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throw their clothes on the chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.1> to <38.8>." + } + ] + }, + { + "video_id": "5O31N", + "question_id": "5O31N", + "video_file": "Charades/videos/5O31N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person closes a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <9.1>." + } + ] + }, + { + "video_id": "5O31N", + "question_id": "5O31N", + "video_file": "Charades/videos/5O31N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person get undressed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.8> to <35.9>." + } + ] + }, + { + "video_id": "5O31N", + "question_id": "5O31N", + "video_file": "Charades/videos/5O31N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts undressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.8> to <35.9>." + } + ] + }, + { + "video_id": "GISE8", + "question_id": "GISE8", + "video_file": "Charades/videos/GISE8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closing the door to a medicine cabinet'." + }, + { + "from": "gpt", + "value": "From <16.5> to <21.9>." + } + ] + }, + { + "video_id": "GISE8", + "question_id": "GISE8", + "video_file": "Charades/videos/GISE8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts a towel in a closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.8> to <19.5>." + } + ] + }, + { + "video_id": "GISE8", + "question_id": "GISE8", + "video_file": "Charades/videos/GISE8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person began dressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.8> to <28.5>." + } + ] + }, + { + "video_id": "GISE8", + "question_id": "GISE8", + "video_file": "Charades/videos/GISE8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts the towel away on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <16.6>." + } + ] + }, + { + "video_id": "1221S", + "question_id": "1221S", + "video_file": "Charades/videos/1221S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting the broom in the corner' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.1> to <23.5>." + } + ] + }, + { + "video_id": "BVCCX", + "question_id": "BVCCX", + "video_file": "Charades/videos/BVCCX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is working on a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "B3SF6", + "question_id": "B3SF6", + "video_file": "Charades/videos/B3SF6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is watching television' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.4> to <16.3>." + } + ] + }, + { + "video_id": "B3SF6", + "question_id": "B3SF6", + "video_file": "Charades/videos/B3SF6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open cabinet doors' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.9> to <27.4>." + } + ] + }, + { + "video_id": "B3SF6", + "question_id": "B3SF6", + "video_file": "Charades/videos/B3SF6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person watches television' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <16.3>." + } + ] + }, + { + "video_id": "CP3Q6", + "question_id": "CP3Q6", + "video_file": "Charades/videos/CP3Q6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they begin cooking on the stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.4> to <29.0>." + } + ] + }, + { + "video_id": "2CR09", + "question_id": "2CR09", + "video_file": "Charades/videos/2CR09.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts bag on car to pull out sweater' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.6> to <8.9>." + } + ] + }, + { + "video_id": "2CR09", + "question_id": "2CR09", + "video_file": "Charades/videos/2CR09.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begin dressing holding a bag'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.2>." + } + ] + }, + { + "video_id": "FABO7", + "question_id": "FABO7", + "video_file": "Charades/videos/FABO7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts smiling at picture on wall' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.9> to <32.0>." + } + ] + }, + { + "video_id": "PM9HG", + "question_id": "PM9HG", + "video_file": "Charades/videos/PM9HG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins eating candy from the candy dish'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.5> to <35.0>." + } + ] + }, + { + "video_id": "PM9HG", + "question_id": "PM9HG", + "video_file": "Charades/videos/PM9HG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eat it'." + }, + { + "from": "gpt", + "value": "From <30.5> to <35.0>." + } + ] + }, + { + "video_id": "PM9HG", + "question_id": "PM9HG", + "video_file": "Charades/videos/PM9HG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "PM9HG", + "question_id": "PM9HG", + "video_file": "Charades/videos/PM9HG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walks through a doorway into a room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.4> to <6.6>." + } + ] + }, + { + "video_id": "O5FR0", + "question_id": "O5FR0", + "video_file": "Charades/videos/O5FR0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is undressing by the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "X9WTR", + "question_id": "X9WTR", + "video_file": "Charades/videos/X9WTR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person sits in a chair'." + }, + { + "from": "gpt", + "value": "From <0.0> to <25.2>." + } + ] + }, + { + "video_id": "X9WTR", + "question_id": "X9WTR", + "video_file": "Charades/videos/X9WTR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person sits down in a chair'." + }, + { + "from": "gpt", + "value": "From <0.0> to <25.2>." + } + ] + }, + { + "video_id": "X9WTR", + "question_id": "X9WTR", + "video_file": "Charades/videos/X9WTR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person watching television' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.8> to <22.3>." + } + ] + }, + { + "video_id": "FRXS5", + "question_id": "FRXS5", + "video_file": "Charades/videos/FRXS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person runs into the room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.9>." + } + ] + }, + { + "video_id": "FRXS5", + "question_id": "FRXS5", + "video_file": "Charades/videos/FRXS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs in'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.9>." + } + ] + }, + { + "video_id": "FRXS5", + "question_id": "FRXS5", + "video_file": "Charades/videos/FRXS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'this person runs into the living room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.9>." + } + ] + }, + { + "video_id": "TUVM8", + "question_id": "TUVM8", + "video_file": "Charades/videos/TUVM8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person as they watch the screen they begin laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.2>." + } + ] + }, + { + "video_id": "K5DIB", + "question_id": "K5DIB", + "video_file": "Charades/videos/K5DIB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns off a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <18.4>." + } + ] + }, + { + "video_id": "K5DIB", + "question_id": "K5DIB", + "video_file": "Charades/videos/K5DIB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <18.4>." + } + ] + }, + { + "video_id": "K5DIB", + "question_id": "K5DIB", + "video_file": "Charades/videos/K5DIB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looking at a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.3>." + } + ] + }, + { + "video_id": "AQ0CP", + "question_id": "AQ0CP", + "video_file": "Charades/videos/AQ0CP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closing a door in a laundry room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.9> to <26.6>." + } + ] + }, + { + "video_id": "AQ0CP", + "question_id": "AQ0CP", + "video_file": "Charades/videos/AQ0CP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.9> to <26.6>." + } + ] + }, + { + "video_id": "KQT19", + "question_id": "KQT19", + "video_file": "Charades/videos/KQT19.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person stands over them holding a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.7> to <15.1>." + } + ] + }, + { + "video_id": "KOF86", + "question_id": "KOF86", + "video_file": "Charades/videos/KOF86.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.3> to <33.0>." + } + ] + }, + { + "video_id": "KOF86", + "question_id": "KOF86", + "video_file": "Charades/videos/KOF86.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begin eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.5> to <31.6>." + } + ] + }, + { + "video_id": "KOF86", + "question_id": "KOF86", + "video_file": "Charades/videos/KOF86.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a sandwich in a bathroom'." + }, + { + "from": "gpt", + "value": "From <24.5> to <31.6>." + } + ] + }, + { + "video_id": "2CFCW", + "question_id": "2CFCW", + "video_file": "Charades/videos/2CFCW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is holding a broom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.8> to <12.5>." + } + ] + }, + { + "video_id": "M7DIB", + "question_id": "M7DIB", + "video_file": "Charades/videos/M7DIB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person looks out a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "M7DIB", + "question_id": "M7DIB", + "video_file": "Charades/videos/M7DIB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they eat a bite of food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <10.0>." + } + ] + }, + { + "video_id": "M7DIB", + "question_id": "M7DIB", + "video_file": "Charades/videos/M7DIB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put a coffee cup in the sink' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.7> to <19.0>." + } + ] + }, + { + "video_id": "M7DIB", + "question_id": "M7DIB", + "video_file": "Charades/videos/M7DIB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats from it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.6> to <10.0>." + } + ] + }, + { + "video_id": "YFZRG", + "question_id": "YFZRG", + "video_file": "Charades/videos/YFZRG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person turns on a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "YFZRG", + "question_id": "YFZRG", + "video_file": "Charades/videos/YFZRG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person turns on the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "6ERRL", + "question_id": "6ERRL", + "video_file": "Charades/videos/6ERRL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <18.0>." + } + ] + }, + { + "video_id": "6ERRL", + "question_id": "6ERRL", + "video_file": "Charades/videos/6ERRL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits with an open book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.8> to <13.4>." + } + ] + }, + { + "video_id": "6ERRL", + "question_id": "6ERRL", + "video_file": "Charades/videos/6ERRL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits on the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.6> to <15.8>." + } + ] + }, + { + "video_id": "5ZXR3", + "question_id": "5ZXR3", + "video_file": "Charades/videos/5ZXR3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person proceeds to open the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.8> to <27.6>." + } + ] + }, + { + "video_id": "C3HIN", + "question_id": "C3HIN", + "video_file": "Charades/videos/C3HIN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they take a sip from a glass of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <19.9>." + } + ] + }, + { + "video_id": "C3HIN", + "question_id": "C3HIN", + "video_file": "Charades/videos/C3HIN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person smiling at a picture' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "C3HIN", + "question_id": "C3HIN", + "video_file": "Charades/videos/C3HIN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they are smiling because they are happy' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "UQPO0", + "question_id": "UQPO0", + "video_file": "Charades/videos/UQPO0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits down in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <39.0> to <49.3>." + } + ] + }, + { + "video_id": "UQPO0", + "question_id": "UQPO0", + "video_file": "Charades/videos/UQPO0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they sit in a nearby chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <39.0> to <49.3>." + } + ] + }, + { + "video_id": "U9LKA", + "question_id": "U9LKA", + "video_file": "Charades/videos/U9LKA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person get up open the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <21.4>." + } + ] + }, + { + "video_id": "G838Q", + "question_id": "G838Q", + "video_file": "Charades/videos/G838Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins to tidy the closet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <18.4>." + } + ] + }, + { + "video_id": "G838Q", + "question_id": "G838Q", + "video_file": "Charades/videos/G838Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes a vacuum out of the closet'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.0>." + } + ] + }, + { + "video_id": "MIV2M", + "question_id": "MIV2M", + "video_file": "Charades/videos/MIV2M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits down on a sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.0> to <23.4>." + } + ] + }, + { + "video_id": "MIV2M", + "question_id": "MIV2M", + "video_file": "Charades/videos/MIV2M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is holding a bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.8>." + } + ] + }, + { + "video_id": "MIV2M", + "question_id": "MIV2M", + "video_file": "Charades/videos/MIV2M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "MIV2M", + "question_id": "MIV2M", + "video_file": "Charades/videos/MIV2M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pours a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.7> to <18.9>." + } + ] + }, + { + "video_id": "MIV2M", + "question_id": "MIV2M", + "video_file": "Charades/videos/MIV2M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they sit down on the couch holding the glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <23.4>." + } + ] + }, + { + "video_id": "MIV2M", + "question_id": "MIV2M", + "video_file": "Charades/videos/MIV2M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sit on the sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <23.4>." + } + ] + }, + { + "video_id": "MIV2M", + "question_id": "MIV2M", + "video_file": "Charades/videos/MIV2M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walks into their living room holding a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.8>." + } + ] + }, + { + "video_id": "R13US", + "question_id": "R13US", + "video_file": "Charades/videos/R13US.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <13.0>." + } + ] + }, + { + "video_id": "R13US", + "question_id": "R13US", + "video_file": "Charades/videos/R13US.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person smiles in the mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.0> to <13.0>." + } + ] + }, + { + "video_id": "0JA9E", + "question_id": "0JA9E", + "video_file": "Charades/videos/0JA9E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.2> to <8.6>." + } + ] + }, + { + "video_id": "0JA9E", + "question_id": "0JA9E", + "video_file": "Charades/videos/0JA9E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throw it on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <12.2>." + } + ] + }, + { + "video_id": "L9J5I", + "question_id": "L9J5I", + "video_file": "Charades/videos/L9J5I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is in a bedroom eating by a window'." + }, + { + "from": "gpt", + "value": "From <0.3> to <11.2>." + } + ] + }, + { + "video_id": "L9J5I", + "question_id": "L9J5I", + "video_file": "Charades/videos/L9J5I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they are eating a snack'." + }, + { + "from": "gpt", + "value": "From <0.3> to <11.2>." + } + ] + }, + { + "video_id": "L9J5I", + "question_id": "L9J5I", + "video_file": "Charades/videos/L9J5I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting on the bed eating some food'." + }, + { + "from": "gpt", + "value": "From <0.3> to <11.2>." + } + ] + }, + { + "video_id": "EY6P4", + "question_id": "EY6P4", + "video_file": "Charades/videos/EY6P4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a door to their study' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.9>." + } + ] + }, + { + "video_id": "EY6P4", + "question_id": "EY6P4", + "video_file": "Charades/videos/EY6P4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person reading a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.8> to <11.0>." + } + ] + }, + { + "video_id": "EY6P4", + "question_id": "EY6P4", + "video_file": "Charades/videos/EY6P4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.9>." + } + ] + }, + { + "video_id": "EY6P4", + "question_id": "EY6P4", + "video_file": "Charades/videos/EY6P4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person reads a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <11.0>." + } + ] + }, + { + "video_id": "TS880", + "question_id": "TS880", + "video_file": "Charades/videos/TS880.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks through a doorway with a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "TS880", + "question_id": "TS880", + "video_file": "Charades/videos/TS880.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is holding a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <21.1>." + } + ] + }, + { + "video_id": "TS880", + "question_id": "TS880", + "video_file": "Charades/videos/TS880.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person dressing in the doorway of the bedroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.3> to <38.5>." + } + ] + }, + { + "video_id": "III9I", + "question_id": "III9I", + "video_file": "Charades/videos/III9I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closing the refrigerator door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.6> to <7.4>." + } + ] + }, + { + "video_id": "III9I", + "question_id": "III9I", + "video_file": "Charades/videos/III9I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person stands in front of an open refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.2>." + } + ] + }, + { + "video_id": "III9I", + "question_id": "III9I", + "video_file": "Charades/videos/III9I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they close the refrigerator door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.6> to <7.4>." + } + ] + }, + { + "video_id": "DD57W", + "question_id": "DD57W", + "video_file": "Charades/videos/DD57W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person in the entryway is working on their laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <22.6>." + } + ] + }, + { + "video_id": "03TL7", + "question_id": "03TL7", + "video_file": "Charades/videos/03TL7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a man eating a cracker' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.9>." + } + ] + }, + { + "video_id": "03TL7", + "question_id": "03TL7", + "video_file": "Charades/videos/03TL7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is sitting down eating food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.9>." + } + ] + }, + { + "video_id": "B5VY1", + "question_id": "B5VY1", + "video_file": "Charades/videos/B5VY1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person walks through a doorway holding a glass'." + }, + { + "from": "gpt", + "value": "From <0.2> to <5.7>." + } + ] + }, + { + "video_id": "B5VY1", + "question_id": "B5VY1", + "video_file": "Charades/videos/B5VY1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the refrigerator door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.7> to <16.0>." + } + ] + }, + { + "video_id": "B5VY1", + "question_id": "B5VY1", + "video_file": "Charades/videos/B5VY1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the glass in a sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.7> to <11.1>." + } + ] + }, + { + "video_id": "ORD96", + "question_id": "ORD96", + "video_file": "Charades/videos/ORD96.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is holding a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.3>." + } + ] + }, + { + "video_id": "IO3CI", + "question_id": "IO3CI", + "video_file": "Charades/videos/IO3CI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is smiling' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.7> to <11.0>." + } + ] + }, + { + "video_id": "IO3CI", + "question_id": "IO3CI", + "video_file": "Charades/videos/IO3CI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sits on a chair eating food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.8> to <18.0>." + } + ] + }, + { + "video_id": "IO3CI", + "question_id": "IO3CI", + "video_file": "Charades/videos/IO3CI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'eating with another person' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.8> to <18.0>." + } + ] + }, + { + "video_id": "IO3CI", + "question_id": "IO3CI", + "video_file": "Charades/videos/IO3CI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person smiling at each other as they eat' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <11.0>." + } + ] + }, + { + "video_id": "IO3CI", + "question_id": "IO3CI", + "video_file": "Charades/videos/IO3CI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person sits on a couch also eating food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.8> to <18.0>." + } + ] + }, + { + "video_id": "JCUT2", + "question_id": "JCUT2", + "video_file": "Charades/videos/JCUT2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting them on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "JCUT2", + "question_id": "JCUT2", + "video_file": "Charades/videos/JCUT2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is tidying up a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "JCUT2", + "question_id": "JCUT2", + "video_file": "Charades/videos/JCUT2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is tidying up a table by taking plates'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "JCUT2", + "question_id": "JCUT2", + "video_file": "Charades/videos/JCUT2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is putting dishes on the shelf'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "3DLKC", + "question_id": "3DLKC", + "video_file": "Charades/videos/3DLKC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person opens the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "3DLKC", + "question_id": "3DLKC", + "video_file": "Charades/videos/3DLKC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a second person opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "LWB5E", + "question_id": "LWB5E", + "video_file": "Charades/videos/LWB5E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person putting homework on the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "C9HLB", + "question_id": "C9HLB", + "video_file": "Charades/videos/C9HLB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'this person closes the laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.4> to <35.0>." + } + ] + }, + { + "video_id": "SPJIE", + "question_id": "SPJIE", + "video_file": "Charades/videos/SPJIE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking coffee from a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "IIJ5I", + "question_id": "IIJ5I", + "video_file": "Charades/videos/IIJ5I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they begin cooking on the stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.5> to <29.0>." + } + ] + }, + { + "video_id": "LHPQS", + "question_id": "LHPQS", + "video_file": "Charades/videos/LHPQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take a picture of something' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.7> to <16.1>." + } + ] + }, + { + "video_id": "MTUR3", + "question_id": "MTUR3", + "video_file": "Charades/videos/MTUR3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person two people walk through a doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.9> to <13.1>." + } + ] + }, + { + "video_id": "MTUR3", + "question_id": "MTUR3", + "video_file": "Charades/videos/MTUR3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person smiling at each other'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.8> to <50.0>." + } + ] + }, + { + "video_id": "MTUR3", + "question_id": "MTUR3", + "video_file": "Charades/videos/MTUR3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they both lie down together in a bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.8> to <30.8>." + } + ] + }, + { + "video_id": "MTUR3", + "question_id": "MTUR3", + "video_file": "Charades/videos/MTUR3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person the people snuggle a blanket' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.8> to <36.5>." + } + ] + }, + { + "video_id": "MTUR3", + "question_id": "MTUR3", + "video_file": "Charades/videos/MTUR3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person snuggle with a blanket'." + }, + { + "from": "gpt", + "value": "From <26.8> to <36.5>." + } + ] + }, + { + "video_id": "ZKGKR", + "question_id": "ZKGKR", + "video_file": "Charades/videos/ZKGKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <15.9>." + } + ] + }, + { + "video_id": "ZKGKR", + "question_id": "ZKGKR", + "video_file": "Charades/videos/ZKGKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'aa person is undressing'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "ZKGKR", + "question_id": "ZKGKR", + "video_file": "Charades/videos/ZKGKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <15.9>." + } + ] + }, + { + "video_id": "ZKGKR", + "question_id": "ZKGKR", + "video_file": "Charades/videos/ZKGKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they close the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <15.9>." + } + ] + }, + { + "video_id": "ZKGKR", + "question_id": "ZKGKR", + "video_file": "Charades/videos/ZKGKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is undressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "OVICR", + "question_id": "OVICR", + "video_file": "Charades/videos/OVICR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <21.3>." + } + ] + }, + { + "video_id": "T1KK1", + "question_id": "T1KK1", + "video_file": "Charades/videos/T1KK1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating from a plate' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.3> to <17.3>." + } + ] + }, + { + "video_id": "T1KK1", + "question_id": "T1KK1", + "video_file": "Charades/videos/T1KK1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating some food out of a glass container'." + }, + { + "from": "gpt", + "value": "From <5.3> to <17.3>." + } + ] + }, + { + "video_id": "T1KK1", + "question_id": "T1KK1", + "video_file": "Charades/videos/T1KK1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is walking around drinking a cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.9>." + } + ] + }, + { + "video_id": "XHG13", + "question_id": "XHG13", + "video_file": "Charades/videos/XHG13.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pours the drink into another cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.2>." + } + ] + }, + { + "video_id": "ENCXV", + "question_id": "ENCXV", + "video_file": "Charades/videos/ENCXV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes a medicine bottle from the desk'." + }, + { + "from": "gpt", + "value": "From <15.3> to <20.9>." + } + ] + }, + { + "video_id": "GXVX8", + "question_id": "GXVX8", + "video_file": "Charades/videos/GXVX8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours water into a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <12.8>." + } + ] + }, + { + "video_id": "25YVM", + "question_id": "25YVM", + "video_file": "Charades/videos/25YVM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'puts the person dirty dishes in it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.8>." + } + ] + }, + { + "video_id": "25YVM", + "question_id": "25YVM", + "video_file": "Charades/videos/25YVM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person washes the dishes in the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.0> to <26.6>." + } + ] + }, + { + "video_id": "IS7PL", + "question_id": "IS7PL", + "video_file": "Charades/videos/IS7PL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks into the bathroom eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "IS7PL", + "question_id": "IS7PL", + "video_file": "Charades/videos/IS7PL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person someone walks through the doorway eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "IS7PL", + "question_id": "IS7PL", + "video_file": "Charades/videos/IS7PL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "8BG1T", + "question_id": "8BG1T", + "video_file": "Charades/videos/8BG1T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person laughs really hard at the television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.5> to <29.8>." + } + ] + }, + { + "video_id": "8BG1T", + "question_id": "8BG1T", + "video_file": "Charades/videos/8BG1T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person laugh at the television' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.5> to <29.8>." + } + ] + }, + { + "video_id": "1JYPW", + "question_id": "1JYPW", + "video_file": "Charades/videos/1JYPW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting some boxes away'." + }, + { + "from": "gpt", + "value": "From <14.1> to <21.2>." + } + ] + }, + { + "video_id": "FPXL7", + "question_id": "FPXL7", + "video_file": "Charades/videos/FPXL7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts a book on a shelf'." + }, + { + "from": "gpt", + "value": "From <4.0> to <8.5>." + } + ] + }, + { + "video_id": "6HR54", + "question_id": "6HR54", + "video_file": "Charades/videos/6HR54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person is at a chair at the table smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <8.4>." + } + ] + }, + { + "video_id": "91SQ4", + "question_id": "91SQ4", + "video_file": "Charades/videos/91SQ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person is opening box full of clothes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "91SQ4", + "question_id": "91SQ4", + "video_file": "Charades/videos/91SQ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person runs out of the room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.7> to <20.3>." + } + ] + }, + { + "video_id": "91SQ4", + "question_id": "91SQ4", + "video_file": "Charades/videos/91SQ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts running to doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.7> to <20.3>." + } + ] + }, + { + "video_id": "M8NEU", + "question_id": "M8NEU", + "video_file": "Charades/videos/M8NEU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is running into a room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "M8NEU", + "question_id": "M8NEU", + "video_file": "Charades/videos/M8NEU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person leave the room closing the door behind them' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.5> to <29.0>." + } + ] + }, + { + "video_id": "M8NEU", + "question_id": "M8NEU", + "video_file": "Charades/videos/M8NEU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs into their bedroom'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "M8NEU", + "question_id": "M8NEU", + "video_file": "Charades/videos/M8NEU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open the door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "OIRJ0", + "question_id": "OIRJ0", + "video_file": "Charades/videos/OIRJ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person laughs more'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <26.9>." + } + ] + }, + { + "video_id": "OIRJ0", + "question_id": "OIRJ0", + "video_file": "Charades/videos/OIRJ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person laughs' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.6> to <26.9>." + } + ] + }, + { + "video_id": "OIRJ0", + "question_id": "OIRJ0", + "video_file": "Charades/videos/OIRJ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits down laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.6> to <26.9>." + } + ] + }, + { + "video_id": "OIRJ0", + "question_id": "OIRJ0", + "video_file": "Charades/videos/OIRJ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person sits down at the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <22.0>." + } + ] + }, + { + "video_id": "OIRJ0", + "question_id": "OIRJ0", + "video_file": "Charades/videos/OIRJ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looks out window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.1>." + } + ] + }, + { + "video_id": "OIRJ0", + "question_id": "OIRJ0", + "video_file": "Charades/videos/OIRJ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is watching something out of the window'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.1>." + } + ] + }, + { + "video_id": "J9BXT", + "question_id": "J9BXT", + "video_file": "Charades/videos/J9BXT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person started sneezing because of all the dust' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <33.9> to <39.0>." + } + ] + }, + { + "video_id": "6GY1M", + "question_id": "6GY1M", + "video_file": "Charades/videos/6GY1M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person smiling walks to a desk' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.0> to <9.2>." + } + ] + }, + { + "video_id": "UB1QP", + "question_id": "UB1QP", + "video_file": "Charades/videos/UB1QP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is smiling' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.0> to <31.1>." + } + ] + }, + { + "video_id": "UB1QP", + "question_id": "UB1QP", + "video_file": "Charades/videos/UB1QP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a other person is smiling at a towel' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.0> to <31.1>." + } + ] + }, + { + "video_id": "UB1QP", + "question_id": "UB1QP", + "video_file": "Charades/videos/UB1QP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.5>." + } + ] + }, + { + "video_id": "759UE", + "question_id": "759UE", + "video_file": "Charades/videos/759UE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens their bathroom door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.6> to <10.4>." + } + ] + }, + { + "video_id": "759UE", + "question_id": "759UE", + "video_file": "Charades/videos/759UE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is opening a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.6> to <10.4>." + } + ] + }, + { + "video_id": "759UE", + "question_id": "759UE", + "video_file": "Charades/videos/759UE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a box from the sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <27.1>." + } + ] + }, + { + "video_id": "R6XU9", + "question_id": "R6XU9", + "video_file": "Charades/videos/R6XU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts on shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "MURXK", + "question_id": "MURXK", + "video_file": "Charades/videos/MURXK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting it on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <12.2>." + } + ] + }, + { + "video_id": "MURXK", + "question_id": "MURXK", + "video_file": "Charades/videos/MURXK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is holding a camera' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "MURXK", + "question_id": "MURXK", + "video_file": "Charades/videos/MURXK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts a camera onto the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <12.2>." + } + ] + }, + { + "video_id": "Z6ZYF", + "question_id": "Z6ZYF", + "video_file": "Charades/videos/Z6ZYF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person was done eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <14.5> to <23.2>." + } + ] + }, + { + "video_id": "Z6ZYF", + "question_id": "Z6ZYF", + "video_file": "Charades/videos/Z6ZYF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes a dish from the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.0> to <23.0>." + } + ] + }, + { + "video_id": "Z6ZYF", + "question_id": "Z6ZYF", + "video_file": "Charades/videos/Z6ZYF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating something'." + }, + { + "from": "gpt", + "value": "From <14.5> to <23.2>." + } + ] + }, + { + "video_id": "Z6ZYF", + "question_id": "Z6ZYF", + "video_file": "Charades/videos/Z6ZYF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs across the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "Z6ZYF", + "question_id": "Z6ZYF", + "video_file": "Charades/videos/Z6ZYF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person doing homework in the recreation room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <23.2>." + } + ] + }, + { + "video_id": "GIC0G", + "question_id": "GIC0G", + "video_file": "Charades/videos/GIC0G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person takes a picture of them with a camera'." + }, + { + "from": "gpt", + "value": "From <14.7> to <28.8>." + } + ] + }, + { + "video_id": "D98TV", + "question_id": "D98TV", + "video_file": "Charades/videos/D98TV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts a bag on the bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.4> to <7.5>." + } + ] + }, + { + "video_id": "D98TV", + "question_id": "D98TV", + "video_file": "Charades/videos/D98TV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts his bag onto his bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <7.5>." + } + ] + }, + { + "video_id": "6703W", + "question_id": "6703W", + "video_file": "Charades/videos/6703W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking from a cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.0> to <14.0>." + } + ] + }, + { + "video_id": "6703W", + "question_id": "6703W", + "video_file": "Charades/videos/6703W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinks out of a glass off a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.0> to <14.0>." + } + ] + }, + { + "video_id": "GQYR9", + "question_id": "GQYR9", + "video_file": "Charades/videos/GQYR9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a drink from a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.0> to <35.0>." + } + ] + }, + { + "video_id": "GQYR9", + "question_id": "GQYR9", + "video_file": "Charades/videos/GQYR9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinks from a glass'." + }, + { + "from": "gpt", + "value": "From <25.0> to <35.0>." + } + ] + }, + { + "video_id": "CLU7Z", + "question_id": "CLU7Z", + "video_file": "Charades/videos/CLU7Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a picture out the window'." + }, + { + "from": "gpt", + "value": "From <10.7> to <21.9>." + } + ] + }, + { + "video_id": "CLU7Z", + "question_id": "CLU7Z", + "video_file": "Charades/videos/CLU7Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is opening a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <12.3>." + } + ] + }, + { + "video_id": "CLU7Z", + "question_id": "CLU7Z", + "video_file": "Charades/videos/CLU7Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.5> to <12.3>." + } + ] + }, + { + "video_id": "W3CPB", + "question_id": "W3CPB", + "video_file": "Charades/videos/W3CPB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throw them on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <11.6>." + } + ] + }, + { + "video_id": "W3CPB", + "question_id": "W3CPB", + "video_file": "Charades/videos/W3CPB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.6> to <12.0>." + } + ] + }, + { + "video_id": "6TV29", + "question_id": "6TV29", + "video_file": "Charades/videos/6TV29.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is putting clothes into a washer' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.5> to <21.4>." + } + ] + }, + { + "video_id": "6TV29", + "question_id": "6TV29", + "video_file": "Charades/videos/6TV29.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts them on the clothes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.5> to <21.4>." + } + ] + }, + { + "video_id": "6TV29", + "question_id": "6TV29", + "video_file": "Charades/videos/6TV29.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person opens the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.0> to <39.9>." + } + ] + }, + { + "video_id": "6TV29", + "question_id": "6TV29", + "video_file": "Charades/videos/6TV29.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts clothes'." + }, + { + "from": "gpt", + "value": "From <13.5> to <21.4>." + } + ] + }, + { + "video_id": "HXCGX", + "question_id": "HXCGX", + "video_file": "Charades/videos/HXCGX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sitting on the bed eating pick up a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.7>." + } + ] + }, + { + "video_id": "HXCGX", + "question_id": "HXCGX", + "video_file": "Charades/videos/HXCGX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating on a bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.7>." + } + ] + }, + { + "video_id": "HXCGX", + "question_id": "HXCGX", + "video_file": "Charades/videos/HXCGX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting on the bed eating food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.7>." + } + ] + }, + { + "video_id": "HXCGX", + "question_id": "HXCGX", + "video_file": "Charades/videos/HXCGX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the box'." + }, + { + "from": "gpt", + "value": "From <25.0> to <33.0>." + } + ] + }, + { + "video_id": "TOY9V", + "question_id": "TOY9V", + "video_file": "Charades/videos/TOY9V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "TOY9V", + "question_id": "TOY9V", + "video_file": "Charades/videos/TOY9V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person closes the cabinet door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.7> to <13.6>." + } + ] + }, + { + "video_id": "6FSAH", + "question_id": "6FSAH", + "video_file": "Charades/videos/6FSAH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person takes food out of the wrapper' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <15.2>." + } + ] + }, + { + "video_id": "TF0KM", + "question_id": "TF0KM", + "video_file": "Charades/videos/TF0KM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is undressing near the window'." + }, + { + "from": "gpt", + "value": "From <13.5> to <26.3>." + } + ] + }, + { + "video_id": "9EP7R", + "question_id": "9EP7R", + "video_file": "Charades/videos/9EP7R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a picture'." + }, + { + "from": "gpt", + "value": "From <21.7> to <32.0>." + } + ] + }, + { + "video_id": "9EP7R", + "question_id": "9EP7R", + "video_file": "Charades/videos/9EP7R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'takes a picture of person a'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.7> to <32.0>." + } + ] + }, + { + "video_id": "7HLTQ", + "question_id": "7HLTQ", + "video_file": "Charades/videos/7HLTQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person begins pouring a glass of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.9> to <11.2>." + } + ] + }, + { + "video_id": "UNXLF", + "question_id": "UNXLF", + "video_file": "Charades/videos/UNXLF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a towel from the stove'." + }, + { + "from": "gpt", + "value": "From <17.7> to <22.4>." + } + ] + }, + { + "video_id": "K5CRM", + "question_id": "K5CRM", + "video_file": "Charades/videos/K5CRM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the pillow on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.2> to <11.3>." + } + ] + }, + { + "video_id": "WT59C", + "question_id": "WT59C", + "video_file": "Charades/videos/WT59C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws it onto the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <23.1>." + } + ] + }, + { + "video_id": "WT59C", + "question_id": "WT59C", + "video_file": "Charades/videos/WT59C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throwing a broom on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.6> to <23.1>." + } + ] + }, + { + "video_id": "8D464", + "question_id": "8D464", + "video_file": "Charades/videos/8D464.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes a picture of a laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.5> to <17.3>." + } + ] + }, + { + "video_id": "8D464", + "question_id": "8D464", + "video_file": "Charades/videos/8D464.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they turn off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.8> to <20.1>." + } + ] + }, + { + "video_id": "8D464", + "question_id": "8D464", + "video_file": "Charades/videos/8D464.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is standing holding a phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "8D464", + "question_id": "8D464", + "video_file": "Charades/videos/8D464.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns off the light'." + }, + { + "from": "gpt", + "value": "From <14.8> to <20.1>." + } + ] + }, + { + "video_id": "AZFKK", + "question_id": "AZFKK", + "video_file": "Charades/videos/AZFKK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is throwing a box towards a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.4>." + } + ] + }, + { + "video_id": "AZFKK", + "question_id": "AZFKK", + "video_file": "Charades/videos/AZFKK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is in a room opening a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.7> to <8.8>." + } + ] + }, + { + "video_id": "AZFKK", + "question_id": "AZFKK", + "video_file": "Charades/videos/AZFKK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is holding the box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "8YNSB", + "question_id": "8YNSB", + "video_file": "Charades/videos/8YNSB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is laying on the ground reading a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "8YNSB", + "question_id": "8YNSB", + "video_file": "Charades/videos/8YNSB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person reading a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "2NV6L", + "question_id": "2NV6L", + "video_file": "Charades/videos/2NV6L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person puts a pair of shoes into the closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <19.0>." + } + ] + }, + { + "video_id": "2NV6L", + "question_id": "2NV6L", + "video_file": "Charades/videos/2NV6L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a pillow from the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.8> to <27.8>." + } + ] + }, + { + "video_id": "P0ILN", + "question_id": "P0ILN", + "video_file": "Charades/videos/P0ILN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks in a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "P0ILN", + "question_id": "P0ILN", + "video_file": "Charades/videos/P0ILN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.1>." + } + ] + }, + { + "video_id": "P0ILN", + "question_id": "P0ILN", + "video_file": "Charades/videos/P0ILN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person playing on their phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <32.0>." + } + ] + }, + { + "video_id": "P0ILN", + "question_id": "P0ILN", + "video_file": "Charades/videos/P0ILN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person play with the phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <32.0>." + } + ] + }, + { + "video_id": "P0ILN", + "question_id": "P0ILN", + "video_file": "Charades/videos/P0ILN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns on a light'." + }, + { + "from": "gpt", + "value": "From <6.2> to <12.8>." + } + ] + }, + { + "video_id": "P0ILN", + "question_id": "P0ILN", + "video_file": "Charades/videos/P0ILN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes off their shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <20.7>." + } + ] + }, + { + "video_id": "P0ILN", + "question_id": "P0ILN", + "video_file": "Charades/videos/P0ILN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.9> to <20.7>." + } + ] + }, + { + "video_id": "P0ILN", + "question_id": "P0ILN", + "video_file": "Charades/videos/P0ILN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person turns on the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.2> to <12.8>." + } + ] + }, + { + "video_id": "9VBJ5", + "question_id": "9VBJ5", + "video_file": "Charades/videos/9VBJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is eating some food'." + }, + { + "from": "gpt", + "value": "From <0.0> to <18.4>." + } + ] + }, + { + "video_id": "9VBJ5", + "question_id": "9VBJ5", + "video_file": "Charades/videos/9VBJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person is eating food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <18.4>." + } + ] + }, + { + "video_id": "9VBJ5", + "question_id": "9VBJ5", + "video_file": "Charades/videos/9VBJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begin washing the glass out in the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.4> to <39.5>." + } + ] + }, + { + "video_id": "9VBJ5", + "question_id": "9VBJ5", + "video_file": "Charades/videos/9VBJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drink a cup of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <28.8>." + } + ] + }, + { + "video_id": "9VBJ5", + "question_id": "9VBJ5", + "video_file": "Charades/videos/9VBJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person wash the glass in the sink' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.4> to <39.5>." + } + ] + }, + { + "video_id": "9CDB6", + "question_id": "9CDB6", + "video_file": "Charades/videos/9CDB6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person runs up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.8> to <28.8>." + } + ] + }, + { + "video_id": "U7E1T", + "question_id": "U7E1T", + "video_file": "Charades/videos/U7E1T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is putting a glass down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <8.6>." + } + ] + }, + { + "video_id": "5I330", + "question_id": "5I330", + "video_file": "Charades/videos/5I330.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a box off the shelf'." + }, + { + "from": "gpt", + "value": "From <11.0> to <18.0>." + } + ] + }, + { + "video_id": "7BX06", + "question_id": "7BX06", + "video_file": "Charades/videos/7BX06.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they awaken'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.5>." + } + ] + }, + { + "video_id": "Z6OEM", + "question_id": "Z6OEM", + "video_file": "Charades/videos/Z6OEM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is opening a closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "3OAGK", + "question_id": "3OAGK", + "video_file": "Charades/videos/3OAGK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the bedroom door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <30.0>." + } + ] + }, + { + "video_id": "3OAGK", + "question_id": "3OAGK", + "video_file": "Charades/videos/3OAGK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <30.0>." + } + ] + }, + { + "video_id": "F0N38", + "question_id": "F0N38", + "video_file": "Charades/videos/F0N38.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes out a cell phone from their pocket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <32.0>." + } + ] + }, + { + "video_id": "F0N38", + "question_id": "F0N38", + "video_file": "Charades/videos/F0N38.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person talks on the phone'." + }, + { + "from": "gpt", + "value": "From <23.0> to <32.0>." + } + ] + }, + { + "video_id": "TLCLG", + "question_id": "TLCLG", + "video_file": "Charades/videos/TLCLG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs down a hallway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.6> to <32.9>." + } + ] + }, + { + "video_id": "TLCLG", + "question_id": "TLCLG", + "video_file": "Charades/videos/TLCLG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person talking on the phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.8> to <22.5>." + } + ] + }, + { + "video_id": "TLCLG", + "question_id": "TLCLG", + "video_file": "Charades/videos/TLCLG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person go for a run' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.6> to <32.9>." + } + ] + }, + { + "video_id": "TLCLG", + "question_id": "TLCLG", + "video_file": "Charades/videos/TLCLG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is talking on the phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.8> to <22.5>." + } + ] + }, + { + "video_id": "TLCLG", + "question_id": "TLCLG", + "video_file": "Charades/videos/TLCLG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person quickly runs out'." + }, + { + "from": "gpt", + "value": "From <23.6> to <32.9>." + } + ] + }, + { + "video_id": "YLNPX", + "question_id": "YLNPX", + "video_file": "Charades/videos/YLNPX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person talks on a phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.8> to <39.0>." + } + ] + }, + { + "video_id": "YLNPX", + "question_id": "YLNPX", + "video_file": "Charades/videos/YLNPX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes out of a cell phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.0> to <39.0>." + } + ] + }, + { + "video_id": "ZOLVU", + "question_id": "ZOLVU", + "video_file": "Charades/videos/ZOLVU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <27.9>." + } + ] + }, + { + "video_id": "2ODLD", + "question_id": "2ODLD", + "video_file": "Charades/videos/2ODLD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting in the walk in closet undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.3>." + } + ] + }, + { + "video_id": "J3ZZH", + "question_id": "J3ZZH", + "video_file": "Charades/videos/J3ZZH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person lying on the floor awakens'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.4>." + } + ] + }, + { + "video_id": "J3ZZH", + "question_id": "J3ZZH", + "video_file": "Charades/videos/J3ZZH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person closes the cabinet'." + }, + { + "from": "gpt", + "value": "From <27.1> to <32.5>." + } + ] + }, + { + "video_id": "J3ZZH", + "question_id": "J3ZZH", + "video_file": "Charades/videos/J3ZZH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens in a pantry'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.4>." + } + ] + }, + { + "video_id": "J3ZZH", + "question_id": "J3ZZH", + "video_file": "Charades/videos/J3ZZH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put a box of food on the shelf'." + }, + { + "from": "gpt", + "value": "From <23.3> to <31.5>." + } + ] + }, + { + "video_id": "J3ZZH", + "question_id": "J3ZZH", + "video_file": "Charades/videos/J3ZZH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts it on a shelf in a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.3> to <31.5>." + } + ] + }, + { + "video_id": "C1LJS", + "question_id": "C1LJS", + "video_file": "Charades/videos/C1LJS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the bedroom door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.8> to <7.7>." + } + ] + }, + { + "video_id": "P73UC", + "question_id": "P73UC", + "video_file": "Charades/videos/P73UC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes a phone from a pocket' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.9> to <30.0>." + } + ] + }, + { + "video_id": "P73UC", + "question_id": "P73UC", + "video_file": "Charades/videos/P73UC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person awakens in bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <9.5>." + } + ] + }, + { + "video_id": "P73UC", + "question_id": "P73UC", + "video_file": "Charades/videos/P73UC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a picture with it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <31.0>." + } + ] + }, + { + "video_id": "A30ZM", + "question_id": "A30ZM", + "video_file": "Charades/videos/A30ZM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is opening a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "UJXBC", + "question_id": "UJXBC", + "video_file": "Charades/videos/UJXBC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.7> to <8.1>." + } + ] + }, + { + "video_id": "Y1BWP", + "question_id": "Y1BWP", + "video_file": "Charades/videos/Y1BWP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a door with a box of tissues' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.6> to <10.1>." + } + ] + }, + { + "video_id": "Y1BWP", + "question_id": "Y1BWP", + "video_file": "Charades/videos/Y1BWP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.6> to <10.1>." + } + ] + }, + { + "video_id": "Y1BWP", + "question_id": "Y1BWP", + "video_file": "Charades/videos/Y1BWP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they put the kleenex back in the box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.4> to <26.2>." + } + ] + }, + { + "video_id": "7L7P4", + "question_id": "7L7P4", + "video_file": "Charades/videos/7L7P4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person put a camera on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.4> to <8.7>." + } + ] + }, + { + "video_id": "0R0DJ", + "question_id": "0R0DJ", + "video_file": "Charades/videos/0R0DJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person laughs as they look at the laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <17.0>." + } + ] + }, + { + "video_id": "0R0DJ", + "question_id": "0R0DJ", + "video_file": "Charades/videos/0R0DJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person sitting on the stairs opens a laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.4> to <9.5>." + } + ] + }, + { + "video_id": "0R0DJ", + "question_id": "0R0DJ", + "video_file": "Charades/videos/0R0DJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person sits on the stairs holding a laptop'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "9O889", + "question_id": "9O889", + "video_file": "Charades/videos/9O889.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person is working on a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.1> to <32.0>." + } + ] + }, + { + "video_id": "2H56B", + "question_id": "2H56B", + "video_file": "Charades/videos/2H56B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the bag down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.9> to <6.0>." + } + ] + }, + { + "video_id": "KTOJX", + "question_id": "KTOJX", + "video_file": "Charades/videos/KTOJX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person was undressing his clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "2YXJO", + "question_id": "2YXJO", + "video_file": "Charades/videos/2YXJO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'when another person runs in with a broom'." + }, + { + "from": "gpt", + "value": "From <7.6> to <13.9>." + } + ] + }, + { + "video_id": "T6BS9", + "question_id": "T6BS9", + "video_file": "Charades/videos/T6BS9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <10.9>." + } + ] + }, + { + "video_id": "T6BS9", + "question_id": "T6BS9", + "video_file": "Charades/videos/T6BS9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws the shoes into a closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.7> to <11.2>." + } + ] + }, + { + "video_id": "T6BS9", + "question_id": "T6BS9", + "video_file": "Charades/videos/T6BS9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throw the shoes in' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.7> to <11.2>." + } + ] + }, + { + "video_id": "T6BS9", + "question_id": "T6BS9", + "video_file": "Charades/videos/T6BS9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws the shoes inside' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.7> to <11.2>." + } + ] + }, + { + "video_id": "T6BS9", + "question_id": "T6BS9", + "video_file": "Charades/videos/T6BS9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they open a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.8> to <9.9>." + } + ] + }, + { + "video_id": "HI81L", + "question_id": "HI81L", + "video_file": "Charades/videos/HI81L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person is sitting on floor wrapped in blanket' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "HI81L", + "question_id": "HI81L", + "video_file": "Charades/videos/HI81L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting on the floor with a blanket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "HI81L", + "question_id": "HI81L", + "video_file": "Charades/videos/HI81L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person opens a small box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <9.5>." + } + ] + }, + { + "video_id": "HI81L", + "question_id": "HI81L", + "video_file": "Charades/videos/HI81L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts laughing'." + }, + { + "from": "gpt", + "value": "From <9.6> to <18.0>." + } + ] + }, + { + "video_id": "BART9", + "question_id": "BART9", + "video_file": "Charades/videos/BART9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person puts groceries onto a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.9> to <22.2>." + } + ] + }, + { + "video_id": "K7FNC", + "question_id": "K7FNC", + "video_file": "Charades/videos/K7FNC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a door'." + }, + { + "from": "gpt", + "value": "From <0.7> to <9.8>." + } + ] + }, + { + "video_id": "K7FNC", + "question_id": "K7FNC", + "video_file": "Charades/videos/K7FNC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens their laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.7> to <23.2>." + } + ] + }, + { + "video_id": "ADK5J", + "question_id": "ADK5J", + "video_file": "Charades/videos/ADK5J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the camera onto the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <20.8>." + } + ] + }, + { + "video_id": "ADK5J", + "question_id": "ADK5J", + "video_file": "Charades/videos/ADK5J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put their camera on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.2> to <20.8>." + } + ] + }, + { + "video_id": "5776I", + "question_id": "5776I", + "video_file": "Charades/videos/5776I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws clothes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.0> to <17.3>." + } + ] + }, + { + "video_id": "5776I", + "question_id": "5776I", + "video_file": "Charades/videos/5776I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person sits on a sofa using a computer' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "5776I", + "question_id": "5776I", + "video_file": "Charades/videos/5776I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person throws clothes on the couch' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.0> to <17.3>." + } + ] + }, + { + "video_id": "5776I", + "question_id": "5776I", + "video_file": "Charades/videos/5776I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person throws their clothes onto a bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.0> to <17.3>." + } + ] + }, + { + "video_id": "HM7J7", + "question_id": "HM7J7", + "video_file": "Charades/videos/HM7J7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turning on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "HM7J7", + "question_id": "HM7J7", + "video_file": "Charades/videos/HM7J7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person turns on the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "U44OC", + "question_id": "U44OC", + "video_file": "Charades/videos/U44OC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person runs through the room holding their cell phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.3> to <32.0>." + } + ] + }, + { + "video_id": "FY9UT", + "question_id": "FY9UT", + "video_file": "Charades/videos/FY9UT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is snuggling a blanket on the bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <17.0>." + } + ] + }, + { + "video_id": "IF7KE", + "question_id": "IF7KE", + "video_file": "Charades/videos/IF7KE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats food from a bowl' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.0> to <18.8>." + } + ] + }, + { + "video_id": "IF7KE", + "question_id": "IF7KE", + "video_file": "Charades/videos/IF7KE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person stands in the same place eating some food'." + }, + { + "from": "gpt", + "value": "From <11.0> to <18.8>." + } + ] + }, + { + "video_id": "TW2F4", + "question_id": "TW2F4", + "video_file": "Charades/videos/TW2F4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sit on the sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.5> to <32.0>." + } + ] + }, + { + "video_id": "TW2F4", + "question_id": "TW2F4", + "video_file": "Charades/videos/TW2F4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person begins undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <16.3>." + } + ] + }, + { + "video_id": "TW2F4", + "question_id": "TW2F4", + "video_file": "Charades/videos/TW2F4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sitting on the sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.5> to <32.0>." + } + ] + }, + { + "video_id": "TW2F4", + "question_id": "TW2F4", + "video_file": "Charades/videos/TW2F4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person undresses out of some clothes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <16.3>." + } + ] + }, + { + "video_id": "QQG8X", + "question_id": "QQG8X", + "video_file": "Charades/videos/QQG8X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person awaken get up' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "I9TIU", + "question_id": "I9TIU", + "video_file": "Charades/videos/I9TIU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person goes to open a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.2> to <28.0>." + } + ] + }, + { + "video_id": "I9TIU", + "question_id": "I9TIU", + "video_file": "Charades/videos/I9TIU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <24.3>." + } + ] + }, + { + "video_id": "X95MU", + "question_id": "X95MU", + "video_file": "Charades/videos/X95MU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take a picture' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.1> to <37.0>." + } + ] + }, + { + "video_id": "X95MU", + "question_id": "X95MU", + "video_file": "Charades/videos/X95MU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts on their shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.5> to <8.2>." + } + ] + }, + { + "video_id": "X95MU", + "question_id": "X95MU", + "video_file": "Charades/videos/X95MU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting on shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.5> to <8.2>." + } + ] + }, + { + "video_id": "X95MU", + "question_id": "X95MU", + "video_file": "Charades/videos/X95MU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take a picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.1> to <37.0>." + } + ] + }, + { + "video_id": "YLRVT", + "question_id": "YLRVT", + "video_file": "Charades/videos/YLRVT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person the throw some clothes on a bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.9> to <16.1>." + } + ] + }, + { + "video_id": "JFSJF", + "question_id": "JFSJF", + "video_file": "Charades/videos/JFSJF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is reading a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "JFSJF", + "question_id": "JFSJF", + "video_file": "Charades/videos/JFSJF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sits in a chair reading a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "JFSJF", + "question_id": "JFSJF", + "video_file": "Charades/videos/JFSJF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts undressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.9> to <27.0>." + } + ] + }, + { + "video_id": "EPP0O", + "question_id": "EPP0O", + "video_file": "Charades/videos/EPP0O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person laughs on camera' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "YUGJJ", + "question_id": "YUGJJ", + "video_file": "Charades/videos/YUGJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <21.3>." + } + ] + }, + { + "video_id": "I713Z", + "question_id": "I713Z", + "video_file": "Charades/videos/I713Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person open the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.4> to <30.8>." + } + ] + }, + { + "video_id": "UN6O9", + "question_id": "UN6O9", + "video_file": "Charades/videos/UN6O9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is throwing clothes on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <9.3>." + } + ] + }, + { + "video_id": "UN6O9", + "question_id": "UN6O9", + "video_file": "Charades/videos/UN6O9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is throwing clothes onto the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.4> to <9.3>." + } + ] + }, + { + "video_id": "W3LUB", + "question_id": "W3LUB", + "video_file": "Charades/videos/W3LUB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the phone into a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.4> to <28.1>." + } + ] + }, + { + "video_id": "33OVO", + "question_id": "33OVO", + "video_file": "Charades/videos/33OVO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "33OVO", + "question_id": "33OVO", + "video_file": "Charades/videos/33OVO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person works on a laptop on a desk' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.7> to <31.0>." + } + ] + }, + { + "video_id": "33OVO", + "question_id": "33OVO", + "video_file": "Charades/videos/33OVO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they close the door behind them'." + }, + { + "from": "gpt", + "value": "From <3.2> to <9.1>." + } + ] + }, + { + "video_id": "33OVO", + "question_id": "33OVO", + "video_file": "Charades/videos/33OVO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person start playing around with the laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <31.0>." + } + ] + }, + { + "video_id": "VEKTO", + "question_id": "VEKTO", + "video_file": "Charades/videos/VEKTO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts a pair of shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "VEKTO", + "question_id": "VEKTO", + "video_file": "Charades/videos/VEKTO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put the shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "VEKTO", + "question_id": "VEKTO", + "video_file": "Charades/videos/VEKTO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a towel'." + }, + { + "from": "gpt", + "value": "From <4.0> to <10.0>." + } + ] + }, + { + "video_id": "RYQI0", + "question_id": "RYQI0", + "video_file": "Charades/videos/RYQI0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they are holding a pillow'." + }, + { + "from": "gpt", + "value": "From <9.3> to <24.5>." + } + ] + }, + { + "video_id": "RYQI0", + "question_id": "RYQI0", + "video_file": "Charades/videos/RYQI0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they throw a pillow on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.5> to <23.9>." + } + ] + }, + { + "video_id": "NQKZB", + "question_id": "NQKZB", + "video_file": "Charades/videos/NQKZB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is tidying up some clothes on their bed'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.5>." + } + ] + }, + { + "video_id": "NQKZB", + "question_id": "NQKZB", + "video_file": "Charades/videos/NQKZB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting clothes in the hamper' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <37.4> to <48.0>." + } + ] + }, + { + "video_id": "ZS7X6", + "question_id": "ZS7X6", + "video_file": "Charades/videos/ZS7X6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a picture out of the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <12.7>." + } + ] + }, + { + "video_id": "ZS7X6", + "question_id": "ZS7X6", + "video_file": "Charades/videos/ZS7X6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they take a picture out of the box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <12.7>." + } + ] + }, + { + "video_id": "FNNTW", + "question_id": "FNNTW", + "video_file": "Charades/videos/FNNTW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person periodically drinking from a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.0> to <32.0>." + } + ] + }, + { + "video_id": "FNNTW", + "question_id": "FNNTW", + "video_file": "Charades/videos/FNNTW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person drinks from a cup of coffee' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.0> to <32.0>." + } + ] + }, + { + "video_id": "G1NFB", + "question_id": "G1NFB", + "video_file": "Charades/videos/G1NFB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person close the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.3> to <28.3>." + } + ] + }, + { + "video_id": "G1NFB", + "question_id": "G1NFB", + "video_file": "Charades/videos/G1NFB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sneezing as the person dusts shelves' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "G1NFB", + "question_id": "G1NFB", + "video_file": "Charades/videos/G1NFB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person begins washing the mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <26.0>." + } + ] + }, + { + "video_id": "Y4LT8", + "question_id": "Y4LT8", + "video_file": "Charades/videos/Y4LT8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person is washing dishes'." + }, + { + "from": "gpt", + "value": "From <17.3> to <31.0>." + } + ] + }, + { + "video_id": "WYZCW", + "question_id": "WYZCW", + "video_file": "Charades/videos/WYZCW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts on a pair of shoes'." + }, + { + "from": "gpt", + "value": "From <12.7> to <24.5>." + } + ] + }, + { + "video_id": "WYZCW", + "question_id": "WYZCW", + "video_file": "Charades/videos/WYZCW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they stand up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "WYZCW", + "question_id": "WYZCW", + "video_file": "Charades/videos/WYZCW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.1> to <16.3>." + } + ] + }, + { + "video_id": "WYZCW", + "question_id": "WYZCW", + "video_file": "Charades/videos/WYZCW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts on their shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.7> to <24.5>." + } + ] + }, + { + "video_id": "WYZCW", + "question_id": "WYZCW", + "video_file": "Charades/videos/WYZCW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.1> to <16.3>." + } + ] + }, + { + "video_id": "55DDW", + "question_id": "55DDW", + "video_file": "Charades/videos/55DDW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they sat their smiling in comfort' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.9> to <40.0>." + } + ] + }, + { + "video_id": "6ZJAL", + "question_id": "6ZJAL", + "video_file": "Charades/videos/6ZJAL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits on the sofa'." + }, + { + "from": "gpt", + "value": "From <11.2> to <22.4>." + } + ] + }, + { + "video_id": "6ZJAL", + "question_id": "6ZJAL", + "video_file": "Charades/videos/6ZJAL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws it on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <16.5>." + } + ] + }, + { + "video_id": "6ZJAL", + "question_id": "6ZJAL", + "video_file": "Charades/videos/6ZJAL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins to eat it'." + }, + { + "from": "gpt", + "value": "From <25.3> to <34.0>." + } + ] + }, + { + "video_id": "6ZJAL", + "question_id": "6ZJAL", + "video_file": "Charades/videos/6ZJAL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person start to eat some food'." + }, + { + "from": "gpt", + "value": "From <25.3> to <34.0>." + } + ] + }, + { + "video_id": "5CZRC", + "question_id": "5CZRC", + "video_file": "Charades/videos/5CZRC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.1> to <17.2>." + } + ] + }, + { + "video_id": "5CZRC", + "question_id": "5CZRC", + "video_file": "Charades/videos/5CZRC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is eating a piece of fruit' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "7SAFU", + "question_id": "7SAFU", + "video_file": "Charades/videos/7SAFU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is awakening to a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <7.1>." + } + ] + }, + { + "video_id": "D5886", + "question_id": "D5886", + "video_file": "Charades/videos/D5886.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person turns to watch the television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.2> to <28.0>." + } + ] + }, + { + "video_id": "D5886", + "question_id": "D5886", + "video_file": "Charades/videos/D5886.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "FQWTE", + "question_id": "FQWTE", + "video_file": "Charades/videos/FQWTE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person start undressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.1> to <20.5>." + } + ] + }, + { + "video_id": "S2A89", + "question_id": "S2A89", + "video_file": "Charades/videos/S2A89.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <10.8>." + } + ] + }, + { + "video_id": "NL26D", + "question_id": "NL26D", + "video_file": "Charades/videos/NL26D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person puts an item on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.2> to <10.2>." + } + ] + }, + { + "video_id": "NL26D", + "question_id": "NL26D", + "video_file": "Charades/videos/NL26D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person walks through a doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "RTULL", + "question_id": "RTULL", + "video_file": "Charades/videos/RTULL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person washes their hands' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.7> to <33.0>." + } + ] + }, + { + "video_id": "RTULL", + "question_id": "RTULL", + "video_file": "Charades/videos/RTULL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person washes hands under the tap'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.7> to <33.0>." + } + ] + }, + { + "video_id": "RTULL", + "question_id": "RTULL", + "video_file": "Charades/videos/RTULL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person walks in to wash their hands' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.7> to <33.0>." + } + ] + }, + { + "video_id": "G8YWY", + "question_id": "G8YWY", + "video_file": "Charades/videos/G8YWY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person is running to put food on stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <8.8>." + } + ] + }, + { + "video_id": "G82VI", + "question_id": "G82VI", + "video_file": "Charades/videos/G82VI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes out a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <17.0>." + } + ] + }, + { + "video_id": "G82VI", + "question_id": "G82VI", + "video_file": "Charades/videos/G82VI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <13.4>." + } + ] + }, + { + "video_id": "G82VI", + "question_id": "G82VI", + "video_file": "Charades/videos/G82VI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "G82VI", + "question_id": "G82VI", + "video_file": "Charades/videos/G82VI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take out a laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <17.0>." + } + ] + }, + { + "video_id": "G82VI", + "question_id": "G82VI", + "video_file": "Charades/videos/G82VI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <13.4>." + } + ] + }, + { + "video_id": "G82VI", + "question_id": "G82VI", + "video_file": "Charades/videos/G82VI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person holding a box enters the garage' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <17.2>." + } + ] + }, + { + "video_id": "4MFE0", + "question_id": "4MFE0", + "video_file": "Charades/videos/4MFE0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they look at a book on the table'." + }, + { + "from": "gpt", + "value": "From <11.7> to <20.9>." + } + ] + }, + { + "video_id": "4MFE0", + "question_id": "4MFE0", + "video_file": "Charades/videos/4MFE0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is eating food'." + }, + { + "from": "gpt", + "value": "From <0.4> to <6.1>." + } + ] + }, + { + "video_id": "4MFE0", + "question_id": "4MFE0", + "video_file": "Charades/videos/4MFE0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating a bag of chips' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.4> to <6.1>." + } + ] + }, + { + "video_id": "4MFE0", + "question_id": "4MFE0", + "video_file": "Charades/videos/4MFE0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they take the book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.8> to <27.2>." + } + ] + }, + { + "video_id": "UIERL", + "question_id": "UIERL", + "video_file": "Charades/videos/UIERL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person playing a game on a laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <17.2>." + } + ] + }, + { + "video_id": "UIERL", + "question_id": "UIERL", + "video_file": "Charades/videos/UIERL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person start eating' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.6> to <24.9>." + } + ] + }, + { + "video_id": "UIERL", + "question_id": "UIERL", + "video_file": "Charades/videos/UIERL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person stops to eat' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.6> to <24.9>." + } + ] + }, + { + "video_id": "UIERL", + "question_id": "UIERL", + "video_file": "Charades/videos/UIERL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person playing on a laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.4> to <17.2>." + } + ] + }, + { + "video_id": "UIERL", + "question_id": "UIERL", + "video_file": "Charades/videos/UIERL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating food in a hallway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.6> to <24.9>." + } + ] + }, + { + "video_id": "UIERL", + "question_id": "UIERL", + "video_file": "Charades/videos/UIERL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is running place' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.9> to <10.2>." + } + ] + }, + { + "video_id": "UIERL", + "question_id": "UIERL", + "video_file": "Charades/videos/UIERL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a man running in place'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <10.2>." + } + ] + }, + { + "video_id": "DV6U8", + "question_id": "DV6U8", + "video_file": "Charades/videos/DV6U8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <19.4>." + } + ] + }, + { + "video_id": "DV6U8", + "question_id": "DV6U8", + "video_file": "Charades/videos/DV6U8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person is washing their hands' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.1>." + } + ] + }, + { + "video_id": "DV6U8", + "question_id": "DV6U8", + "video_file": "Charades/videos/DV6U8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person runs out of the room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.4> to <32.0>." + } + ] + }, + { + "video_id": "DV6U8", + "question_id": "DV6U8", + "video_file": "Charades/videos/DV6U8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes out a camera' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.7> to <21.1>." + } + ] + }, + { + "video_id": "DV6U8", + "question_id": "DV6U8", + "video_file": "Charades/videos/DV6U8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is washing their hands in the bathroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.1>." + } + ] + }, + { + "video_id": "DV6U8", + "question_id": "DV6U8", + "video_file": "Charades/videos/DV6U8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin running out the room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.4> to <32.0>." + } + ] + }, + { + "video_id": "UPUPM", + "question_id": "UPUPM", + "video_file": "Charades/videos/UPUPM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they start throwing clothes off the bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.6> to <27.9>." + } + ] + }, + { + "video_id": "UPUPM", + "question_id": "UPUPM", + "video_file": "Charades/videos/UPUPM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins throwing the clothes into the corner'." + }, + { + "from": "gpt", + "value": "From <16.6> to <27.9>." + } + ] + }, + { + "video_id": "CR1OK", + "question_id": "CR1OK", + "video_file": "Charades/videos/CR1OK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.1> to <12.7>." + } + ] + }, + { + "video_id": "CR1OK", + "question_id": "CR1OK", + "video_file": "Charades/videos/CR1OK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits in a chair at a desk'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.1> to <12.7>." + } + ] + }, + { + "video_id": "CR1OK", + "question_id": "CR1OK", + "video_file": "Charades/videos/CR1OK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they go sit in the chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.1> to <12.7>." + } + ] + }, + { + "video_id": "PVAJ8", + "question_id": "PVAJ8", + "video_file": "Charades/videos/PVAJ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person grab the closet doorknob'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.2> to <37.9>." + } + ] + }, + { + "video_id": "KLO0K", + "question_id": "KLO0K", + "video_file": "Charades/videos/KLO0K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is holding a doorknob' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.0> to <17.4>." + } + ] + }, + { + "video_id": "KLO0K", + "question_id": "KLO0K", + "video_file": "Charades/videos/KLO0K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person lays down on a pillow on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.9> to <33.0>." + } + ] + }, + { + "video_id": "KLO0K", + "question_id": "KLO0K", + "video_file": "Charades/videos/KLO0K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person lays down on the floor'." + }, + { + "from": "gpt", + "value": "From <17.9> to <33.0>." + } + ] + }, + { + "video_id": "UPYPG", + "question_id": "UPYPG", + "video_file": "Charades/videos/UPYPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person runs through the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <21.8>." + } + ] + }, + { + "video_id": "UPYPG", + "question_id": "UPYPG", + "video_file": "Charades/videos/UPYPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person suddenly comes running through'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <21.8>." + } + ] + }, + { + "video_id": "RDHNQ", + "question_id": "RDHNQ", + "video_file": "Charades/videos/RDHNQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person turns on a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <5.0>." + } + ] + }, + { + "video_id": "RDHNQ", + "question_id": "RDHNQ", + "video_file": "Charades/videos/RDHNQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person turns on the light in the garage' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <5.0>." + } + ] + }, + { + "video_id": "RDHNQ", + "question_id": "RDHNQ", + "video_file": "Charades/videos/RDHNQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they take their phone out of their pocket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.3> to <38.0>." + } + ] + }, + { + "video_id": "RDHNQ", + "question_id": "RDHNQ", + "video_file": "Charades/videos/RDHNQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <14.2>." + } + ] + }, + { + "video_id": "OEIRN", + "question_id": "OEIRN", + "video_file": "Charades/videos/OEIRN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.3> to <18.0>." + } + ] + }, + { + "video_id": "OEIRN", + "question_id": "OEIRN", + "video_file": "Charades/videos/OEIRN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person take some medicine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.3> to <36.7>." + } + ] + }, + { + "video_id": "OEIRN", + "question_id": "OEIRN", + "video_file": "Charades/videos/OEIRN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take the medicine'." + }, + { + "from": "gpt", + "value": "From <19.3> to <36.7>." + } + ] + }, + { + "video_id": "OEIRN", + "question_id": "OEIRN", + "video_file": "Charades/videos/OEIRN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes some with a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.6> to <30.0>." + } + ] + }, + { + "video_id": "OEIRN", + "question_id": "OEIRN", + "video_file": "Charades/videos/OEIRN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person sits on a sofa eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <18.0>." + } + ] + }, + { + "video_id": "OEIRN", + "question_id": "OEIRN", + "video_file": "Charades/videos/OEIRN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes medicine off a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.3> to <36.7>." + } + ] + }, + { + "video_id": "V8GK6", + "question_id": "V8GK6", + "video_file": "Charades/videos/V8GK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns off a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.8> to <33.0>." + } + ] + }, + { + "video_id": "V8GK6", + "question_id": "V8GK6", + "video_file": "Charades/videos/V8GK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.8> to <33.0>." + } + ] + }, + { + "video_id": "MU6UQ", + "question_id": "MU6UQ", + "video_file": "Charades/videos/MU6UQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is awakening in bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "5D351", + "question_id": "5D351", + "video_file": "Charades/videos/5D351.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throw their shoes on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.3> to <29.4>." + } + ] + }, + { + "video_id": "5D351", + "question_id": "5D351", + "video_file": "Charades/videos/5D351.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throw the shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.3> to <29.4>." + } + ] + }, + { + "video_id": "13IOT", + "question_id": "13IOT", + "video_file": "Charades/videos/13IOT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person was laughing at something that was on television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "13IOT", + "question_id": "13IOT", + "video_file": "Charades/videos/13IOT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person is watching television' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.8>." + } + ] + }, + { + "video_id": "LH6LE", + "question_id": "LH6LE", + "video_file": "Charades/videos/LH6LE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.5> to <8.5>." + } + ] + }, + { + "video_id": "LH6LE", + "question_id": "LH6LE", + "video_file": "Charades/videos/LH6LE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.5> to <8.5>." + } + ] + }, + { + "video_id": "LH6LE", + "question_id": "LH6LE", + "video_file": "Charades/videos/LH6LE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.6> to <26.9>." + } + ] + }, + { + "video_id": "LH6LE", + "question_id": "LH6LE", + "video_file": "Charades/videos/LH6LE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts to put away dishes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.2> to <22.0>." + } + ] + }, + { + "video_id": "LH6LE", + "question_id": "LH6LE", + "video_file": "Charades/videos/LH6LE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person seen holding some dishes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.4> to <15.3>." + } + ] + }, + { + "video_id": "5XA4R", + "question_id": "5XA4R", + "video_file": "Charades/videos/5XA4R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits on the sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.4> to <8.8>." + } + ] + }, + { + "video_id": "5XA4R", + "question_id": "5XA4R", + "video_file": "Charades/videos/5XA4R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person to go sit on a couch'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <8.8>." + } + ] + }, + { + "video_id": "VICUM", + "question_id": "VICUM", + "video_file": "Charades/videos/VICUM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person close the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.5> to <5.9>." + } + ] + }, + { + "video_id": "VICUM", + "question_id": "VICUM", + "video_file": "Charades/videos/VICUM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door multiple times' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <5.9>." + } + ] + }, + { + "video_id": "VICUM", + "question_id": "VICUM", + "video_file": "Charades/videos/VICUM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closing the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.5> to <5.9>." + } + ] + }, + { + "video_id": "KU4UP", + "question_id": "KU4UP", + "video_file": "Charades/videos/KU4UP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes a cup from a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <8.0>." + } + ] + }, + { + "video_id": "O7FDB", + "question_id": "O7FDB", + "video_file": "Charades/videos/O7FDB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they put 3 bottles into the fridge' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.1> to <11.0>." + } + ] + }, + { + "video_id": "O7FDB", + "question_id": "O7FDB", + "video_file": "Charades/videos/O7FDB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person enters the room talking on their cell phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.9> to <16.5>." + } + ] + }, + { + "video_id": "O7FDB", + "question_id": "O7FDB", + "video_file": "Charades/videos/O7FDB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person talking on the phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.9> to <16.5>." + } + ] + }, + { + "video_id": "O7FDB", + "question_id": "O7FDB", + "video_file": "Charades/videos/O7FDB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person started putting beer on the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.3> to <24.0>." + } + ] + }, + { + "video_id": "3B81O", + "question_id": "3B81O", + "video_file": "Charades/videos/3B81O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person opens a window'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "3B81O", + "question_id": "3B81O", + "video_file": "Charades/videos/3B81O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts away food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.7> to <26.8>." + } + ] + }, + { + "video_id": "3B81O", + "question_id": "3B81O", + "video_file": "Charades/videos/3B81O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <28.6> to <34.0>." + } + ] + }, + { + "video_id": "3B81O", + "question_id": "3B81O", + "video_file": "Charades/videos/3B81O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person closes the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <28.6> to <34.0>." + } + ] + }, + { + "video_id": "3B81O", + "question_id": "3B81O", + "video_file": "Charades/videos/3B81O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "43FG9", + "question_id": "43FG9", + "video_file": "Charades/videos/43FG9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting clothes on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.2> to <8.7>." + } + ] + }, + { + "video_id": "EVU9O", + "question_id": "EVU9O", + "video_file": "Charades/videos/EVU9O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person is sneezing by the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.0> to <25.3>." + } + ] + }, + { + "video_id": "YXSI7", + "question_id": "YXSI7", + "video_file": "Charades/videos/YXSI7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the front door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.9> to <6.1>." + } + ] + }, + { + "video_id": "YXSI7", + "question_id": "YXSI7", + "video_file": "Charades/videos/YXSI7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person walks through a doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.6> to <7.6>." + } + ] + }, + { + "video_id": "YXSI7", + "question_id": "YXSI7", + "video_file": "Charades/videos/YXSI7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.1> to <20.7>." + } + ] + }, + { + "video_id": "YXSI7", + "question_id": "YXSI7", + "video_file": "Charades/videos/YXSI7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person proceeded to have a sneezing fit'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.1> to <20.7>." + } + ] + }, + { + "video_id": "VDUO1", + "question_id": "VDUO1", + "video_file": "Charades/videos/VDUO1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person washing dishes in a pantry'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <37.0>." + } + ] + }, + { + "video_id": "5V5EL", + "question_id": "5V5EL", + "video_file": "Charades/videos/5V5EL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person takes a picture with a camera' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "G4853", + "question_id": "G4853", + "video_file": "Charades/videos/G4853.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sneezing into a phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.2> to <13.3>." + } + ] + }, + { + "video_id": "G4853", + "question_id": "G4853", + "video_file": "Charades/videos/G4853.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening a cabinet in a garage'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.4> to <31.0>." + } + ] + }, + { + "video_id": "97086", + "question_id": "97086", + "video_file": "Charades/videos/97086.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.4> to <27.3>." + } + ] + }, + { + "video_id": "0IJ8T", + "question_id": "0IJ8T", + "video_file": "Charades/videos/0IJ8T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person uses a tool to fix a doorknob' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.1> to <11.7>." + } + ] + }, + { + "video_id": "0IJ8T", + "question_id": "0IJ8T", + "video_file": "Charades/videos/0IJ8T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person is fixing the doorknob'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <11.7>." + } + ] + }, + { + "video_id": "HVBES", + "question_id": "HVBES", + "video_file": "Charades/videos/HVBES.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person immediately started running' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <16.9>." + } + ] + }, + { + "video_id": "X9R37", + "question_id": "X9R37", + "video_file": "Charades/videos/X9R37.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closing the door they reopen the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.3> to <5.7>." + } + ] + }, + { + "video_id": "X9R37", + "question_id": "X9R37", + "video_file": "Charades/videos/X9R37.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person to remember to turn off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.3> to <9.4>." + } + ] + }, + { + "video_id": "X9R37", + "question_id": "X9R37", + "video_file": "Charades/videos/X9R37.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closing the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <5.7>." + } + ] + }, + { + "video_id": "X9R37", + "question_id": "X9R37", + "video_file": "Charades/videos/X9R37.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they close the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <5.7>." + } + ] + }, + { + "video_id": "X9R37", + "question_id": "X9R37", + "video_file": "Charades/videos/X9R37.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turn off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.3> to <9.4>." + } + ] + }, + { + "video_id": "E0ZBC", + "question_id": "E0ZBC", + "video_file": "Charades/videos/E0ZBC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person pours a cup of coffee from a thermos' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <28.7>." + } + ] + }, + { + "video_id": "E0ZBC", + "question_id": "E0ZBC", + "video_file": "Charades/videos/E0ZBC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is pouring coffee into a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <28.7>." + } + ] + }, + { + "video_id": "JKZUM", + "question_id": "JKZUM", + "video_file": "Charades/videos/JKZUM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is opening a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "TRUIW", + "question_id": "TRUIW", + "video_file": "Charades/videos/TRUIW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they begin opening the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <29.2>." + } + ] + }, + { + "video_id": "TRUIW", + "question_id": "TRUIW", + "video_file": "Charades/videos/TRUIW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes a dosage of medicine' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.3> to <13.5>." + } + ] + }, + { + "video_id": "TRUIW", + "question_id": "TRUIW", + "video_file": "Charades/videos/TRUIW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a door'." + }, + { + "from": "gpt", + "value": "From <19.0> to <29.2>." + } + ] + }, + { + "video_id": "XIVBU", + "question_id": "XIVBU", + "video_file": "Charades/videos/XIVBU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "6PZJ3", + "question_id": "6PZJ3", + "video_file": "Charades/videos/6PZJ3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts some food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <14.8>." + } + ] + }, + { + "video_id": "I1CS8", + "question_id": "I1CS8", + "video_file": "Charades/videos/I1CS8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looks out a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.1> to <31.0>." + } + ] + }, + { + "video_id": "I1CS8", + "question_id": "I1CS8", + "video_file": "Charades/videos/I1CS8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person stand up'." + }, + { + "from": "gpt", + "value": "From <21.3> to <27.5>." + } + ] + }, + { + "video_id": "I1CS8", + "question_id": "I1CS8", + "video_file": "Charades/videos/I1CS8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person looks out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.1> to <31.0>." + } + ] + }, + { + "video_id": "I1CS8", + "question_id": "I1CS8", + "video_file": "Charades/videos/I1CS8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person look out the window'." + }, + { + "from": "gpt", + "value": "From <23.1> to <31.0>." + } + ] + }, + { + "video_id": "6YN3J", + "question_id": "6YN3J", + "video_file": "Charades/videos/6YN3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the sandwich on the coffee table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.8> to <13.1>." + } + ] + }, + { + "video_id": "6YN3J", + "question_id": "6YN3J", + "video_file": "Charades/videos/6YN3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "6YN3J", + "question_id": "6YN3J", + "video_file": "Charades/videos/6YN3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person stand up' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.5> to <23.8>." + } + ] + }, + { + "video_id": "6YN3J", + "question_id": "6YN3J", + "video_file": "Charades/videos/6YN3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting on the couch eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "FM078", + "question_id": "FM078", + "video_file": "Charades/videos/FM078.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person tidys up the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.2> to <12.4>." + } + ] + }, + { + "video_id": "FM078", + "question_id": "FM078", + "video_file": "Charades/videos/FM078.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts a box in the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <19.4>." + } + ] + }, + { + "video_id": "G99VH", + "question_id": "G99VH", + "video_file": "Charades/videos/G99VH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts down the camera' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <17.9>." + } + ] + }, + { + "video_id": "G99VH", + "question_id": "G99VH", + "video_file": "Charades/videos/G99VH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a picture of themselves' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.4> to <33.4>." + } + ] + }, + { + "video_id": "Q37OS", + "question_id": "Q37OS", + "video_file": "Charades/videos/Q37OS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person runs down holding a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.4> to <18.9>." + } + ] + }, + { + "video_id": "48WOM", + "question_id": "48WOM", + "video_file": "Charades/videos/48WOM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes out a box'." + }, + { + "from": "gpt", + "value": "From <8.1> to <15.7>." + } + ] + }, + { + "video_id": "48WOM", + "question_id": "48WOM", + "video_file": "Charades/videos/48WOM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person she opens the box'." + }, + { + "from": "gpt", + "value": "From <4.2> to <9.5>." + } + ] + }, + { + "video_id": "48WOM", + "question_id": "48WOM", + "video_file": "Charades/videos/48WOM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a medicine cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.7> to <8.8>." + } + ] + }, + { + "video_id": "IFO3G", + "question_id": "IFO3G", + "video_file": "Charades/videos/IFO3G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins laughing as they open a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.5> to <37.0>." + } + ] + }, + { + "video_id": "IFO3G", + "question_id": "IFO3G", + "video_file": "Charades/videos/IFO3G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open a cabinet door'." + }, + { + "from": "gpt", + "value": "From <30.5> to <37.0>." + } + ] + }, + { + "video_id": "IFO3G", + "question_id": "IFO3G", + "video_file": "Charades/videos/IFO3G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting on a bed they' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.2>." + } + ] + }, + { + "video_id": "IFO3G", + "question_id": "IFO3G", + "video_file": "Charades/videos/IFO3G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open up the cabinet to look inside' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.5> to <37.0>." + } + ] + }, + { + "video_id": "55BOW", + "question_id": "55BOW", + "video_file": "Charades/videos/55BOW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walks out the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.1> to <29.4>." + } + ] + }, + { + "video_id": "UPL9S", + "question_id": "UPL9S", + "video_file": "Charades/videos/UPL9S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks in holding laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <57.8> to <76.0>." + } + ] + }, + { + "video_id": "3L7Y9", + "question_id": "3L7Y9", + "video_file": "Charades/videos/3L7Y9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking a cup of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.9>." + } + ] + }, + { + "video_id": "WF7TY", + "question_id": "WF7TY", + "video_file": "Charades/videos/WF7TY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <12.0>." + } + ] + }, + { + "video_id": "OFB7A", + "question_id": "OFB7A", + "video_file": "Charades/videos/OFB7A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looks at a picture'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <21.5>." + } + ] + }, + { + "video_id": "OFB7A", + "question_id": "OFB7A", + "video_file": "Charades/videos/OFB7A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is drinking a cup of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.3> to <22.1>." + } + ] + }, + { + "video_id": "YXQWH", + "question_id": "YXQWH", + "video_file": "Charades/videos/YXQWH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts the groceries into a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.3> to <30.2>." + } + ] + }, + { + "video_id": "QGLQ0", + "question_id": "QGLQ0", + "video_file": "Charades/videos/QGLQ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they stand up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "QGLQ0", + "question_id": "QGLQ0", + "video_file": "Charades/videos/QGLQ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts a canvas bag into a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.1> to <10.6>." + } + ] + }, + { + "video_id": "7UGF7", + "question_id": "7UGF7", + "video_file": "Charades/videos/7UGF7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person wards they start eating a snack' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.8> to <48.1>." + } + ] + }, + { + "video_id": "7UGF7", + "question_id": "7UGF7", + "video_file": "Charades/videos/7UGF7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens'." + }, + { + "from": "gpt", + "value": "From <3.8> to <16.8>." + } + ] + }, + { + "video_id": "7UGF7", + "question_id": "7UGF7", + "video_file": "Charades/videos/7UGF7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is awakening in a office' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.8> to <16.8>." + } + ] + }, + { + "video_id": "7UGF7", + "question_id": "7UGF7", + "video_file": "Charades/videos/7UGF7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins to eat food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.8> to <48.1>." + } + ] + }, + { + "video_id": "7UGF7", + "question_id": "7UGF7", + "video_file": "Charades/videos/7UGF7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eat a snack'." + }, + { + "from": "gpt", + "value": "From <27.8> to <48.1>." + } + ] + }, + { + "video_id": "7UGF7", + "question_id": "7UGF7", + "video_file": "Charades/videos/7UGF7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drink from a cup'." + }, + { + "from": "gpt", + "value": "From <13.5> to <30.6>." + } + ] + }, + { + "video_id": "106AY", + "question_id": "106AY", + "video_file": "Charades/videos/106AY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <29.1> to <35.0>." + } + ] + }, + { + "video_id": "106AY", + "question_id": "106AY", + "video_file": "Charades/videos/106AY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person is turning light on' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.1> to <9.1>." + } + ] + }, + { + "video_id": "106AY", + "question_id": "106AY", + "video_file": "Charades/videos/106AY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.1> to <35.0>." + } + ] + }, + { + "video_id": "106AY", + "question_id": "106AY", + "video_file": "Charades/videos/106AY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting shoes on' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.9> to <20.6>." + } + ] + }, + { + "video_id": "KZN8C", + "question_id": "KZN8C", + "video_file": "Charades/videos/KZN8C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pours a beverage in a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.6> to <22.9>." + } + ] + }, + { + "video_id": "KZN8C", + "question_id": "KZN8C", + "video_file": "Charades/videos/KZN8C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the camera on the shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <26.0>." + } + ] + }, + { + "video_id": "KZN8C", + "question_id": "KZN8C", + "video_file": "Charades/videos/KZN8C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pours water into a glass on a washing machine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.6> to <22.9>." + } + ] + }, + { + "video_id": "KZN8C", + "question_id": "KZN8C", + "video_file": "Charades/videos/KZN8C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the glass on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.9> to <26.0>." + } + ] + }, + { + "video_id": "WCY52", + "question_id": "WCY52", + "video_file": "Charades/videos/WCY52.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is cooking on a stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.3> to <17.0>." + } + ] + }, + { + "video_id": "WCY52", + "question_id": "WCY52", + "video_file": "Charades/videos/WCY52.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking at a stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.3> to <17.0>." + } + ] + }, + { + "video_id": "WCY52", + "question_id": "WCY52", + "video_file": "Charades/videos/WCY52.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws some clothes that they spot on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.7> to <27.6>." + } + ] + }, + { + "video_id": "Z5AZW", + "question_id": "Z5AZW", + "video_file": "Charades/videos/Z5AZW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person in the hallway holding a towel' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.6>." + } + ] + }, + { + "video_id": "92AO4", + "question_id": "92AO4", + "video_file": "Charades/videos/92AO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person runs in' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "92AO4", + "question_id": "92AO4", + "video_file": "Charades/videos/92AO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a towel'." + }, + { + "from": "gpt", + "value": "From <0.8> to <9.5>." + } + ] + }, + { + "video_id": "92AO4", + "question_id": "92AO4", + "video_file": "Charades/videos/92AO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs into a room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "HEOJM", + "question_id": "HEOJM", + "video_file": "Charades/videos/HEOJM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is throwing a pillow on the floor'." + }, + { + "from": "gpt", + "value": "From <2.7> to <9.9>." + } + ] + }, + { + "video_id": "HEOJM", + "question_id": "HEOJM", + "video_file": "Charades/videos/HEOJM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a pillow off of the bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.8> to <7.4>." + } + ] + }, + { + "video_id": "HEOJM", + "question_id": "HEOJM", + "video_file": "Charades/videos/HEOJM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws the pillow on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.7> to <9.9>." + } + ] + }, + { + "video_id": "8MX1T", + "question_id": "8MX1T", + "video_file": "Charades/videos/8MX1T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinks from cup taken from a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.1> to <29.1>." + } + ] + }, + { + "video_id": "9YI82", + "question_id": "9YI82", + "video_file": "Charades/videos/9YI82.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person reading a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <21.6>." + } + ] + }, + { + "video_id": "9YI82", + "question_id": "9YI82", + "video_file": "Charades/videos/9YI82.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the book down on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.7> to <27.5>." + } + ] + }, + { + "video_id": "9YI82", + "question_id": "9YI82", + "video_file": "Charades/videos/9YI82.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is smiling as they read a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <21.6>." + } + ] + }, + { + "video_id": "9YI82", + "question_id": "9YI82", + "video_file": "Charades/videos/9YI82.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person laughs' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.9> to <26.1>." + } + ] + }, + { + "video_id": "ULAWO", + "question_id": "ULAWO", + "video_file": "Charades/videos/ULAWO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a camera' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.2> to <15.6>." + } + ] + }, + { + "video_id": "ULAWO", + "question_id": "ULAWO", + "video_file": "Charades/videos/ULAWO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes a picture of a pile of clothes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.4> to <15.6>." + } + ] + }, + { + "video_id": "A5XR5", + "question_id": "A5XR5", + "video_file": "Charades/videos/A5XR5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is working on their laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.0>." + } + ] + }, + { + "video_id": "A5XR5", + "question_id": "A5XR5", + "video_file": "Charades/videos/A5XR5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes their shoes off' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <18.3>." + } + ] + }, + { + "video_id": "IET7H", + "question_id": "IET7H", + "video_file": "Charades/videos/IET7H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they are opening the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <17.2>." + } + ] + }, + { + "video_id": "539EH", + "question_id": "539EH", + "video_file": "Charades/videos/539EH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes picture of shoes with camera'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.5> to <17.2>." + } + ] + }, + { + "video_id": "539EH", + "question_id": "539EH", + "video_file": "Charades/videos/539EH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <10.1>." + } + ] + }, + { + "video_id": "539EH", + "question_id": "539EH", + "video_file": "Charades/videos/539EH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts opening box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.7> to <10.1>." + } + ] + }, + { + "video_id": "539EH", + "question_id": "539EH", + "video_file": "Charades/videos/539EH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts box with shoes on table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <18.2>." + } + ] + }, + { + "video_id": "MT6C2", + "question_id": "MT6C2", + "video_file": "Charades/videos/MT6C2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open a window'." + }, + { + "from": "gpt", + "value": "From <2.8> to <10.5>." + } + ] + }, + { + "video_id": "MT6C2", + "question_id": "MT6C2", + "video_file": "Charades/videos/MT6C2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they put the pillow on a counter' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.7> to <5.6>." + } + ] + }, + { + "video_id": "MT6C2", + "question_id": "MT6C2", + "video_file": "Charades/videos/MT6C2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they open a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <10.5>." + } + ] + }, + { + "video_id": "MT6C2", + "question_id": "MT6C2", + "video_file": "Charades/videos/MT6C2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs in the room carrying a pillow'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "MT6C2", + "question_id": "MT6C2", + "video_file": "Charades/videos/MT6C2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <10.5>." + } + ] + }, + { + "video_id": "MT6C2", + "question_id": "MT6C2", + "video_file": "Charades/videos/MT6C2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs into the kitchen holding a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "OX3BI", + "question_id": "OX3BI", + "video_file": "Charades/videos/OX3BI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.7> to <9.8>." + } + ] + }, + { + "video_id": "OX3BI", + "question_id": "OX3BI", + "video_file": "Charades/videos/OX3BI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is putting their homework on the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.6> to <12.5>." + } + ] + }, + { + "video_id": "QKP9V", + "question_id": "QKP9V", + "video_file": "Charades/videos/QKP9V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a door to another room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.1> to <22.2>." + } + ] + }, + { + "video_id": "QKP9V", + "question_id": "QKP9V", + "video_file": "Charades/videos/QKP9V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is laughing by themselves' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <8.5>." + } + ] + }, + { + "video_id": "QKP9V", + "question_id": "QKP9V", + "video_file": "Charades/videos/QKP9V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the door'." + }, + { + "from": "gpt", + "value": "From <14.1> to <22.2>." + } + ] + }, + { + "video_id": "UMUBT", + "question_id": "UMUBT", + "video_file": "Charades/videos/UMUBT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is tidying a blanket on the bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "M79TJ", + "question_id": "M79TJ", + "video_file": "Charades/videos/M79TJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes some dishes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.8> to <11.2>." + } + ] + }, + { + "video_id": "186EF", + "question_id": "186EF", + "video_file": "Charades/videos/186EF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person is holding a broom'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "186EF", + "question_id": "186EF", + "video_file": "Charades/videos/186EF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person holds a broom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "KFHLP", + "question_id": "KFHLP", + "video_file": "Charades/videos/KFHLP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <7.7>." + } + ] + }, + { + "video_id": "KFHLP", + "question_id": "KFHLP", + "video_file": "Charades/videos/KFHLP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs into the room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.9>." + } + ] + }, + { + "video_id": "KFHLP", + "question_id": "KFHLP", + "video_file": "Charades/videos/KFHLP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts a plate of food on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "KFHLP", + "question_id": "KFHLP", + "video_file": "Charades/videos/KFHLP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they run back' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.9>." + } + ] + }, + { + "video_id": "BAGJW", + "question_id": "BAGJW", + "video_file": "Charades/videos/BAGJW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts a paper onto a table'." + }, + { + "from": "gpt", + "value": "From <5.8> to <13.8>." + } + ] + }, + { + "video_id": "RO429", + "question_id": "RO429", + "video_file": "Charades/videos/RO429.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "RO429", + "question_id": "RO429", + "video_file": "Charades/videos/RO429.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts a bag inside' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "LG73V", + "question_id": "LG73V", + "video_file": "Charades/videos/LG73V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they take off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.1> to <24.0>." + } + ] + }, + { + "video_id": "LG73V", + "question_id": "LG73V", + "video_file": "Charades/videos/LG73V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person puts a book into a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.3> to <15.0>." + } + ] + }, + { + "video_id": "ACAK4", + "question_id": "ACAK4", + "video_file": "Charades/videos/ACAK4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begin sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <24.6>." + } + ] + }, + { + "video_id": "ACAK4", + "question_id": "ACAK4", + "video_file": "Charades/videos/ACAK4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes a towel from a chair'." + }, + { + "from": "gpt", + "value": "From <13.6> to <22.1>." + } + ] + }, + { + "video_id": "ACAK4", + "question_id": "ACAK4", + "video_file": "Charades/videos/ACAK4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sits at a table working on their laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "SVQFY", + "question_id": "SVQFY", + "video_file": "Charades/videos/SVQFY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts the cup in the sink'." + }, + { + "from": "gpt", + "value": "From <27.0> to <34.0>." + } + ] + }, + { + "video_id": "SVQFY", + "question_id": "SVQFY", + "video_file": "Charades/videos/SVQFY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <8.1> to <22.3>." + } + ] + }, + { + "video_id": "SVQFY", + "question_id": "SVQFY", + "video_file": "Charades/videos/SVQFY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating a sandwich starts tidying up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <22.3>." + } + ] + }, + { + "video_id": "SVQFY", + "question_id": "SVQFY", + "video_file": "Charades/videos/SVQFY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating something'." + }, + { + "from": "gpt", + "value": "From <14.3> to <24.9>." + } + ] + }, + { + "video_id": "M7BD4", + "question_id": "M7BD4", + "video_file": "Charades/videos/M7BD4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they open the closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.0> to <23.5>." + } + ] + }, + { + "video_id": "M7BD4", + "question_id": "M7BD4", + "video_file": "Charades/videos/M7BD4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person close their laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.2> to <26.2>." + } + ] + }, + { + "video_id": "M7BD4", + "question_id": "M7BD4", + "video_file": "Charades/videos/M7BD4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open a closet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.1> to <23.5>." + } + ] + }, + { + "video_id": "M7BD4", + "question_id": "M7BD4", + "video_file": "Charades/videos/M7BD4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person tidy up some clothes on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <10.9>." + } + ] + }, + { + "video_id": "WMD0E", + "question_id": "WMD0E", + "video_file": "Charades/videos/WMD0E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person the puts the phone away' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.9> to <23.8>." + } + ] + }, + { + "video_id": "WMD0E", + "question_id": "WMD0E", + "video_file": "Charades/videos/WMD0E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts their phone away' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.9> to <23.8>." + } + ] + }, + { + "video_id": "R3Z2O", + "question_id": "R3Z2O", + "video_file": "Charades/videos/R3Z2O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person began playing with his phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.0> to <10.4>." + } + ] + }, + { + "video_id": "F796T", + "question_id": "F796T", + "video_file": "Charades/videos/F796T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.6> to <13.6>." + } + ] + }, + { + "video_id": "41FNM", + "question_id": "41FNM", + "video_file": "Charades/videos/41FNM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.3> to <29.2>." + } + ] + }, + { + "video_id": "41FNM", + "question_id": "41FNM", + "video_file": "Charades/videos/41FNM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person dresses' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <10.5>." + } + ] + }, + { + "video_id": "41FNM", + "question_id": "41FNM", + "video_file": "Charades/videos/41FNM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person one is getting dressed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.0> to <10.5>." + } + ] + }, + { + "video_id": "41FNM", + "question_id": "41FNM", + "video_file": "Charades/videos/41FNM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the first person opens a closet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.3> to <29.2>." + } + ] + }, + { + "video_id": "XJ4T3", + "question_id": "XJ4T3", + "video_file": "Charades/videos/XJ4T3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opens a box in the closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.0> to <11.4>." + } + ] + }, + { + "video_id": "XJ4T3", + "question_id": "XJ4T3", + "video_file": "Charades/videos/XJ4T3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening a boxed game' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <11.4>." + } + ] + }, + { + "video_id": "XJ4T3", + "question_id": "XJ4T3", + "video_file": "Charades/videos/XJ4T3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is opening the door to the closer'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "XJ4T3", + "question_id": "XJ4T3", + "video_file": "Charades/videos/XJ4T3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "XJ4T3", + "question_id": "XJ4T3", + "video_file": "Charades/videos/XJ4T3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person holds the towel close to their face' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.4> to <25.0>." + } + ] + }, + { + "video_id": "V2DAA", + "question_id": "V2DAA", + "video_file": "Charades/videos/V2DAA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put a pillow in the chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.1> to <23.2>." + } + ] + }, + { + "video_id": "V2DAA", + "question_id": "V2DAA", + "video_file": "Charades/videos/V2DAA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put a pillow behind their back' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <23.2>." + } + ] + }, + { + "video_id": "70PUI", + "question_id": "70PUI", + "video_file": "Charades/videos/70PUI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person gets dressed next to a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "70PUI", + "question_id": "70PUI", + "video_file": "Charades/videos/70PUI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.7> to <16.7>." + } + ] + }, + { + "video_id": "70PUI", + "question_id": "70PUI", + "video_file": "Charades/videos/70PUI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is dressing in front of the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "A2JV3", + "question_id": "A2JV3", + "video_file": "Charades/videos/A2JV3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the dishes onto the shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.9> to <15.7>." + } + ] + }, + { + "video_id": "P73KX", + "question_id": "P73KX", + "video_file": "Charades/videos/P73KX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits on a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.3> to <16.3>." + } + ] + }, + { + "video_id": "P73KX", + "question_id": "P73KX", + "video_file": "Charades/videos/P73KX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.3> to <9.4>." + } + ] + }, + { + "video_id": "P73KX", + "question_id": "P73KX", + "video_file": "Charades/videos/P73KX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes off their shoes'." + }, + { + "from": "gpt", + "value": "From <1.3> to <9.4>." + } + ] + }, + { + "video_id": "P31E5", + "question_id": "P31E5", + "video_file": "Charades/videos/P31E5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a glass of a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.0> to <9.0>." + } + ] + }, + { + "video_id": "P31E5", + "question_id": "P31E5", + "video_file": "Charades/videos/P31E5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinks from the glass'." + }, + { + "from": "gpt", + "value": "From <10.4> to <19.7>." + } + ] + }, + { + "video_id": "E1TJI", + "question_id": "E1TJI", + "video_file": "Charades/videos/E1TJI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person is smiling at the picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.7> to <33.7>." + } + ] + }, + { + "video_id": "E1TJI", + "question_id": "E1TJI", + "video_file": "Charades/videos/E1TJI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person starts sneezing'." + }, + { + "from": "gpt", + "value": "From <22.5> to <27.0>." + } + ] + }, + { + "video_id": "5S1SL", + "question_id": "5S1SL", + "video_file": "Charades/videos/5S1SL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts something on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <11.4>." + } + ] + }, + { + "video_id": "95M38", + "question_id": "95M38", + "video_file": "Charades/videos/95M38.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put their coffee cup down a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.6> to <26.1>." + } + ] + }, + { + "video_id": "95M38", + "question_id": "95M38", + "video_file": "Charades/videos/95M38.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is drinking a cup of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.7> to <14.3>." + } + ] + }, + { + "video_id": "E7ESG", + "question_id": "E7ESG", + "video_file": "Charades/videos/E7ESG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person open a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.3> to <32.0>." + } + ] + }, + { + "video_id": "E7ESG", + "question_id": "E7ESG", + "video_file": "Charades/videos/E7ESG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they awaken' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.9> to <20.4>." + } + ] + }, + { + "video_id": "E7ESG", + "question_id": "E7ESG", + "video_file": "Charades/videos/E7ESG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits down to look at a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.2> to <31.9>." + } + ] + }, + { + "video_id": "GHJPQ", + "question_id": "GHJPQ", + "video_file": "Charades/videos/GHJPQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.8> to <21.7>." + } + ] + }, + { + "video_id": "GHJPQ", + "question_id": "GHJPQ", + "video_file": "Charades/videos/GHJPQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pull out a camera to take a picture'." + }, + { + "from": "gpt", + "value": "From <23.2> to <31.0>." + } + ] + }, + { + "video_id": "GHJPQ", + "question_id": "GHJPQ", + "video_file": "Charades/videos/GHJPQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they decide to sit on the chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.8> to <21.7>." + } + ] + }, + { + "video_id": "PQH3Q", + "question_id": "PQH3Q", + "video_file": "Charades/videos/PQH3Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they stand up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.4> to <20.3>." + } + ] + }, + { + "video_id": "PQH3Q", + "question_id": "PQH3Q", + "video_file": "Charades/videos/PQH3Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person is watching themselves in the mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.1> to <30.0>." + } + ] + }, + { + "video_id": "PQH3Q", + "question_id": "PQH3Q", + "video_file": "Charades/videos/PQH3Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person fixes their hair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.3> to <30.0>." + } + ] + }, + { + "video_id": "PQH3Q", + "question_id": "PQH3Q", + "video_file": "Charades/videos/PQH3Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person fix their hair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.3> to <30.0>." + } + ] + }, + { + "video_id": "ZCG3J", + "question_id": "ZCG3J", + "video_file": "Charades/videos/ZCG3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.9> to <23.2>." + } + ] + }, + { + "video_id": "ZCG3J", + "question_id": "ZCG3J", + "video_file": "Charades/videos/ZCG3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits down at a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <26.0>." + } + ] + }, + { + "video_id": "ZCG3J", + "question_id": "ZCG3J", + "video_file": "Charades/videos/ZCG3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <15.0>." + } + ] + }, + { + "video_id": "ZCG3J", + "question_id": "ZCG3J", + "video_file": "Charades/videos/ZCG3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a couple of cabinets' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.1> to <15.0>." + } + ] + }, + { + "video_id": "ZCG3J", + "question_id": "ZCG3J", + "video_file": "Charades/videos/ZCG3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens up a cabinet'." + }, + { + "from": "gpt", + "value": "From <4.1> to <15.0>." + } + ] + }, + { + "video_id": "ZCG3J", + "question_id": "ZCG3J", + "video_file": "Charades/videos/ZCG3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sit down at a table with it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <26.0>." + } + ] + }, + { + "video_id": "ZCG3J", + "question_id": "ZCG3J", + "video_file": "Charades/videos/ZCG3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'smiling the person takes the laptop off the shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.3> to <16.2>." + } + ] + }, + { + "video_id": "DEXMO", + "question_id": "DEXMO", + "video_file": "Charades/videos/DEXMO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they take some clothes from the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.1> to <25.0>." + } + ] + }, + { + "video_id": "J8JOQ", + "question_id": "J8JOQ", + "video_file": "Charades/videos/J8JOQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the phone down'." + }, + { + "from": "gpt", + "value": "From <29.1> to <32.0>." + } + ] + }, + { + "video_id": "J8JOQ", + "question_id": "J8JOQ", + "video_file": "Charades/videos/J8JOQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the phone back on the base' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.1> to <32.0>." + } + ] + }, + { + "video_id": "PFTHZ", + "question_id": "PFTHZ", + "video_file": "Charades/videos/PFTHZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'cooking in a recreation room / man cave' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.2> to <25.0>." + } + ] + }, + { + "video_id": "UVMZ8", + "question_id": "UVMZ8", + "video_file": "Charades/videos/UVMZ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person holding a phone turns on a light switch'." + }, + { + "from": "gpt", + "value": "From <0.1> to <6.7>." + } + ] + }, + { + "video_id": "UVMZ8", + "question_id": "UVMZ8", + "video_file": "Charades/videos/UVMZ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts drinking something from a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <14.8>." + } + ] + }, + { + "video_id": "UVMZ8", + "question_id": "UVMZ8", + "video_file": "Charades/videos/UVMZ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <17.6>." + } + ] + }, + { + "video_id": "UVMZ8", + "question_id": "UVMZ8", + "video_file": "Charades/videos/UVMZ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person turns on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.0> to <17.6>." + } + ] + }, + { + "video_id": "03EW0", + "question_id": "03EW0", + "video_file": "Charades/videos/03EW0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is throwing a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.6>." + } + ] + }, + { + "video_id": "UQJ58", + "question_id": "UQJ58", + "video_file": "Charades/videos/UQJ58.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <10.4>." + } + ] + }, + { + "video_id": "UQJ58", + "question_id": "UQJ58", + "video_file": "Charades/videos/UQJ58.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.6> to <32.4>." + } + ] + }, + { + "video_id": "UQJ58", + "question_id": "UQJ58", + "video_file": "Charades/videos/UQJ58.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is looking out a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "UQJ58", + "question_id": "UQJ58", + "video_file": "Charades/videos/UQJ58.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.6> to <32.4>." + } + ] + }, + { + "video_id": "6UMSJ", + "question_id": "6UMSJ", + "video_file": "Charades/videos/6UMSJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes food out of a bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "S8PVE", + "question_id": "S8PVE", + "video_file": "Charades/videos/S8PVE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is smiling' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.0> to <29.3>." + } + ] + }, + { + "video_id": "4J494", + "question_id": "4J494", + "video_file": "Charades/videos/4J494.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person takes some clothes out of a washing machine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <14.0>." + } + ] + }, + { + "video_id": "4J494", + "question_id": "4J494", + "video_file": "Charades/videos/4J494.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.8> to <31.0>." + } + ] + }, + { + "video_id": "MNXXD", + "question_id": "MNXXD", + "video_file": "Charades/videos/MNXXD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drink a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.0> to <28.1>." + } + ] + }, + { + "video_id": "2U903", + "question_id": "2U903", + "video_file": "Charades/videos/2U903.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is watching the mirror'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.0>." + } + ] + }, + { + "video_id": "9OMY1", + "question_id": "9OMY1", + "video_file": "Charades/videos/9OMY1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks to the couch to sit down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.0> to <28.0>." + } + ] + }, + { + "video_id": "9OMY1", + "question_id": "9OMY1", + "video_file": "Charades/videos/9OMY1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.2> to <6.3>." + } + ] + }, + { + "video_id": "9OMY1", + "question_id": "9OMY1", + "video_file": "Charades/videos/9OMY1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they open the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.2> to <6.3>." + } + ] + }, + { + "video_id": "HOLED", + "question_id": "HOLED", + "video_file": "Charades/videos/HOLED.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pours out some water into a cup'." + }, + { + "from": "gpt", + "value": "From <1.9> to <8.0>." + } + ] + }, + { + "video_id": "HOLED", + "question_id": "HOLED", + "video_file": "Charades/videos/HOLED.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person takes a bottle' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "XTC0F", + "question_id": "XTC0F", + "video_file": "Charades/videos/XTC0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens up plastic bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <16.0>." + } + ] + }, + { + "video_id": "XTC0F", + "question_id": "XTC0F", + "video_file": "Charades/videos/XTC0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person opens the pantry door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "XTC0F", + "question_id": "XTC0F", + "video_file": "Charades/videos/XTC0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <16.0>." + } + ] + }, + { + "video_id": "XTC0F", + "question_id": "XTC0F", + "video_file": "Charades/videos/XTC0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "XTC0F", + "question_id": "XTC0F", + "video_file": "Charades/videos/XTC0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person walks into a room holding a towel'." + }, + { + "from": "gpt", + "value": "From <12.5> to <26.0>." + } + ] + }, + { + "video_id": "57N4Z", + "question_id": "57N4Z", + "video_file": "Charades/videos/57N4Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turning off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.7> to <27.6>." + } + ] + }, + { + "video_id": "57N4Z", + "question_id": "57N4Z", + "video_file": "Charades/videos/57N4Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.7> to <27.6>." + } + ] + }, + { + "video_id": "1M89F", + "question_id": "1M89F", + "video_file": "Charades/videos/1M89F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is pouring a drink into a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.4> to <16.7>." + } + ] + }, + { + "video_id": "1M89F", + "question_id": "1M89F", + "video_file": "Charades/videos/1M89F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinks some coffee from the glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.4> to <16.7>." + } + ] + }, + { + "video_id": "19PNV", + "question_id": "19PNV", + "video_file": "Charades/videos/19PNV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes a blanket' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.7> to <21.3>." + } + ] + }, + { + "video_id": "Z2MY2", + "question_id": "Z2MY2", + "video_file": "Charades/videos/Z2MY2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks into bedroom sits down on their bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.2> to <6.0>." + } + ] + }, + { + "video_id": "Z2MY2", + "question_id": "Z2MY2", + "video_file": "Charades/videos/Z2MY2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person their phone they sit on the bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.2> to <6.0>." + } + ] + }, + { + "video_id": "QNHM7", + "question_id": "QNHM7", + "video_file": "Charades/videos/QNHM7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting it on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.3> to <27.0>." + } + ] + }, + { + "video_id": "QNHM7", + "question_id": "QNHM7", + "video_file": "Charades/videos/QNHM7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person is running to get coffee from shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "QNHM7", + "question_id": "QNHM7", + "video_file": "Charades/videos/QNHM7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts cup of coffee on table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.3> to <27.0>." + } + ] + }, + { + "video_id": "QNHM7", + "question_id": "QNHM7", + "video_file": "Charades/videos/QNHM7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is running in a garage' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "VOX2S", + "question_id": "VOX2S", + "video_file": "Charades/videos/VOX2S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person open a refrigerator door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.3> to <19.4>." + } + ] + }, + { + "video_id": "VOX2S", + "question_id": "VOX2S", + "video_file": "Charades/videos/VOX2S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the second person takes the food to a sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <26.8>." + } + ] + }, + { + "video_id": "V3B51", + "question_id": "V3B51", + "video_file": "Charades/videos/V3B51.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a man is reading a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "V3B51", + "question_id": "V3B51", + "video_file": "Charades/videos/V3B51.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person he turns on the lights' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.2> to <8.8>." + } + ] + }, + { + "video_id": "V3B51", + "question_id": "V3B51", + "video_file": "Charades/videos/V3B51.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person turns on the stairway light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.2> to <8.8>." + } + ] + }, + { + "video_id": "QX33M", + "question_id": "QX33M", + "video_file": "Charades/videos/QX33M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they close the door behind them'." + }, + { + "from": "gpt", + "value": "From <27.4> to <32.0>." + } + ] + }, + { + "video_id": "QX33M", + "question_id": "QX33M", + "video_file": "Charades/videos/QX33M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they stand up to grab a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.3> to <25.3>." + } + ] + }, + { + "video_id": "XZLNT", + "question_id": "XZLNT", + "video_file": "Charades/videos/XZLNT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take some dishes off a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <21.1>." + } + ] + }, + { + "video_id": "QB7ZN", + "question_id": "QB7ZN", + "video_file": "Charades/videos/QB7ZN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they are sitting down on a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.1> to <36.0>." + } + ] + }, + { + "video_id": "QB7ZN", + "question_id": "QB7ZN", + "video_file": "Charades/videos/QB7ZN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits in chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <36.0>." + } + ] + }, + { + "video_id": "FKJ9L", + "question_id": "FKJ9L", + "video_file": "Charades/videos/FKJ9L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person puts a box on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.1> to <7.8>." + } + ] + }, + { + "video_id": "NVKDC", + "question_id": "NVKDC", + "video_file": "Charades/videos/NVKDC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts the bag onto a shelf'." + }, + { + "from": "gpt", + "value": "From <8.8> to <14.0>." + } + ] + }, + { + "video_id": "NVKDC", + "question_id": "NVKDC", + "video_file": "Charades/videos/NVKDC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person puts a bag down on a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <14.0>." + } + ] + }, + { + "video_id": "NVKDC", + "question_id": "NVKDC", + "video_file": "Charades/videos/NVKDC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <13.2>." + } + ] + }, + { + "video_id": "8TNSB", + "question_id": "8TNSB", + "video_file": "Charades/videos/8TNSB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <10.1>." + } + ] + }, + { + "video_id": "8TNSB", + "question_id": "8TNSB", + "video_file": "Charades/videos/8TNSB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a second person runs through the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <12.2>." + } + ] + }, + { + "video_id": "8TNSB", + "question_id": "8TNSB", + "video_file": "Charades/videos/8TNSB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.8> to <10.1>." + } + ] + }, + { + "video_id": "1OVB8", + "question_id": "1OVB8", + "video_file": "Charades/videos/1OVB8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns off the lights' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.6> to <24.0>." + } + ] + }, + { + "video_id": "UDAQU", + "question_id": "UDAQU", + "video_file": "Charades/videos/UDAQU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sneezing multiple times' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "UDAQU", + "question_id": "UDAQU", + "video_file": "Charades/videos/UDAQU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sneezing on their sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "UDAQU", + "question_id": "UDAQU", + "video_file": "Charades/videos/UDAQU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person puts a sandwich on top of a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.3> to <20.4>." + } + ] + }, + { + "video_id": "WEVQH", + "question_id": "WEVQH", + "video_file": "Charades/videos/WEVQH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they begin to throw clothes from a bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.7> to <26.0>." + } + ] + }, + { + "video_id": "WEVQH", + "question_id": "WEVQH", + "video_file": "Charades/videos/WEVQH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes clothing out from the bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.6> to <10.9>." + } + ] + }, + { + "video_id": "Q24OC", + "question_id": "Q24OC", + "video_file": "Charades/videos/Q24OC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins putting dishes on a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.8> to <30.0>." + } + ] + }, + { + "video_id": "GADLP", + "question_id": "GADLP", + "video_file": "Charades/videos/GADLP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they are smiling at the camera' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.5> to <32.0>." + } + ] + }, + { + "video_id": "X91NG", + "question_id": "X91NG", + "video_file": "Charades/videos/X91NG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person sweeps dirt out of an open door'." + }, + { + "from": "gpt", + "value": "From <22.5> to <27.8>." + } + ] + }, + { + "video_id": "X91NG", + "question_id": "X91NG", + "video_file": "Charades/videos/X91NG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.6> to <38.1>." + } + ] + }, + { + "video_id": "X91NG", + "question_id": "X91NG", + "video_file": "Charades/videos/X91NG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the door twice' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.5> to <27.8>." + } + ] + }, + { + "video_id": "EMF9Q", + "question_id": "EMF9Q", + "video_file": "Charades/videos/EMF9Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they open a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.7> to <30.2>." + } + ] + }, + { + "video_id": "GQGAE", + "question_id": "GQGAE", + "video_file": "Charades/videos/GQGAE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "GQGAE", + "question_id": "GQGAE", + "video_file": "Charades/videos/GQGAE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is opening a refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "GQGAE", + "question_id": "GQGAE", + "video_file": "Charades/videos/GQGAE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes out medicine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.0> to <32.0>." + } + ] + }, + { + "video_id": "GQGAE", + "question_id": "GQGAE", + "video_file": "Charades/videos/GQGAE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a cabinet'." + }, + { + "from": "gpt", + "value": "From <14.2> to <25.4>." + } + ] + }, + { + "video_id": "V2BCZ", + "question_id": "V2BCZ", + "video_file": "Charades/videos/V2BCZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins to put clothe in a bag'." + }, + { + "from": "gpt", + "value": "From <17.4> to <28.0>." + } + ] + }, + { + "video_id": "V2BCZ", + "question_id": "V2BCZ", + "video_file": "Charades/videos/V2BCZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put that bag into a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.4> to <28.0>." + } + ] + }, + { + "video_id": "WFD5Z", + "question_id": "WFD5Z", + "video_file": "Charades/videos/WFD5Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person closes the laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <11.7>." + } + ] + }, + { + "video_id": "WFD5Z", + "question_id": "WFD5Z", + "video_file": "Charades/videos/WFD5Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a bite of a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <18.8>." + } + ] + }, + { + "video_id": "WFD5Z", + "question_id": "WFD5Z", + "video_file": "Charades/videos/WFD5Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person close the laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <11.7>." + } + ] + }, + { + "video_id": "WFD5Z", + "question_id": "WFD5Z", + "video_file": "Charades/videos/WFD5Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <11.7>." + } + ] + }, + { + "video_id": "WFD5Z", + "question_id": "WFD5Z", + "video_file": "Charades/videos/WFD5Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person is sitting on chair playing games on laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "WFD5Z", + "question_id": "WFD5Z", + "video_file": "Charades/videos/WFD5Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person proceed to eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <27.2>." + } + ] + }, + { + "video_id": "WFD5Z", + "question_id": "WFD5Z", + "video_file": "Charades/videos/WFD5Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts eating sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <27.2>." + } + ] + }, + { + "video_id": "WFD5Z", + "question_id": "WFD5Z", + "video_file": "Charades/videos/WFD5Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is playing on their laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "EJFBM", + "question_id": "EJFBM", + "video_file": "Charades/videos/EJFBM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinks from cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.9> to <15.1>." + } + ] + }, + { + "video_id": "6RXCP", + "question_id": "6RXCP", + "video_file": "Charades/videos/6RXCP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person washing the dishes in the sink' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.1>." + } + ] + }, + { + "video_id": "6RXCP", + "question_id": "6RXCP", + "video_file": "Charades/videos/6RXCP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person washes dishes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.1>." + } + ] + }, + { + "video_id": "6RXCP", + "question_id": "6RXCP", + "video_file": "Charades/videos/6RXCP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opened the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.2> to <28.6>." + } + ] + }, + { + "video_id": "PV745", + "question_id": "PV745", + "video_file": "Charades/videos/PV745.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts a box on their lap' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.9> to <41.9>." + } + ] + }, + { + "video_id": "PV745", + "question_id": "PV745", + "video_file": "Charades/videos/PV745.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put on a pair of shoes'." + }, + { + "from": "gpt", + "value": "From <33.9> to <46.7>." + } + ] + }, + { + "video_id": "PV745", + "question_id": "PV745", + "video_file": "Charades/videos/PV745.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put the box on the ground' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.9> to <41.9>." + } + ] + }, + { + "video_id": "PV745", + "question_id": "PV745", + "video_file": "Charades/videos/PV745.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sits down in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <28.0> to <56.9>." + } + ] + }, + { + "video_id": "PV745", + "question_id": "PV745", + "video_file": "Charades/videos/PV745.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts their shoes on'." + }, + { + "from": "gpt", + "value": "From <33.9> to <46.7>." + } + ] + }, + { + "video_id": "PV745", + "question_id": "PV745", + "video_file": "Charades/videos/PV745.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts on their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.9> to <46.7>." + } + ] + }, + { + "video_id": "MIRSU", + "question_id": "MIRSU", + "video_file": "Charades/videos/MIRSU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is undressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <11.4>." + } + ] + }, + { + "video_id": "MIRSU", + "question_id": "MIRSU", + "video_file": "Charades/videos/MIRSU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing in the living room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <11.4>." + } + ] + }, + { + "video_id": "QBIOX", + "question_id": "QBIOX", + "video_file": "Charades/videos/QBIOX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts running in the garage' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.7> to <31.9>." + } + ] + }, + { + "video_id": "QBIOX", + "question_id": "QBIOX", + "video_file": "Charades/videos/QBIOX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs out' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.7> to <31.9>." + } + ] + }, + { + "video_id": "QBIOX", + "question_id": "QBIOX", + "video_file": "Charades/videos/QBIOX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person runs out of the room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.7> to <31.9>." + } + ] + }, + { + "video_id": "VTYXW", + "question_id": "VTYXW", + "video_file": "Charades/videos/VTYXW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is opening the refrigerator door'." + }, + { + "from": "gpt", + "value": "From <4.4> to <10.6>." + } + ] + }, + { + "video_id": "VTYXW", + "question_id": "VTYXW", + "video_file": "Charades/videos/VTYXW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opens the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.4> to <10.6>." + } + ] + }, + { + "video_id": "VTYXW", + "question_id": "VTYXW", + "video_file": "Charades/videos/VTYXW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the refrigerator again' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <10.6>." + } + ] + }, + { + "video_id": "VTYXW", + "question_id": "VTYXW", + "video_file": "Charades/videos/VTYXW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person stir something cooking in a pot' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <39.3> to <65.8>." + } + ] + }, + { + "video_id": "VTYXW", + "question_id": "VTYXW", + "video_file": "Charades/videos/VTYXW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pouring themselves a glass of water'." + }, + { + "from": "gpt", + "value": "From <26.5> to <38.4>." + } + ] + }, + { + "video_id": "VTYXW", + "question_id": "VTYXW", + "video_file": "Charades/videos/VTYXW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours some into a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.5> to <38.4>." + } + ] + }, + { + "video_id": "VTYXW", + "question_id": "VTYXW", + "video_file": "Charades/videos/VTYXW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person cooking something on the stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <39.3> to <65.8>." + } + ] + }, + { + "video_id": "GSIYV", + "question_id": "GSIYV", + "video_file": "Charades/videos/GSIYV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is putting a towel' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.2> to <12.1>." + } + ] + }, + { + "video_id": "VA8ZR", + "question_id": "VA8ZR", + "video_file": "Charades/videos/VA8ZR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a camera off the shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.7> to <28.4>." + } + ] + }, + { + "video_id": "VA8ZR", + "question_id": "VA8ZR", + "video_file": "Charades/videos/VA8ZR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a door'." + }, + { + "from": "gpt", + "value": "From <4.1> to <14.3>." + } + ] + }, + { + "video_id": "VA8ZR", + "question_id": "VA8ZR", + "video_file": "Charades/videos/VA8ZR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the pantry door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.1> to <14.3>." + } + ] + }, + { + "video_id": "VA8ZR", + "question_id": "VA8ZR", + "video_file": "Charades/videos/VA8ZR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <19.1>." + } + ] + }, + { + "video_id": "VA8ZR", + "question_id": "VA8ZR", + "video_file": "Charades/videos/VA8ZR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns on a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <19.1>." + } + ] + }, + { + "video_id": "4MGUQ", + "question_id": "4MGUQ", + "video_file": "Charades/videos/4MGUQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is washing dishes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "4MGUQ", + "question_id": "4MGUQ", + "video_file": "Charades/videos/4MGUQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinks from a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.4> to <18.0>." + } + ] + }, + { + "video_id": "4MGUQ", + "question_id": "4MGUQ", + "video_file": "Charades/videos/4MGUQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sitting in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "YH70K", + "question_id": "YH70K", + "video_file": "Charades/videos/YH70K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts some items on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.6> to <26.9>." + } + ] + }, + { + "video_id": "55OAL", + "question_id": "55OAL", + "video_file": "Charades/videos/55OAL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a closet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <15.0>." + } + ] + }, + { + "video_id": "55OAL", + "question_id": "55OAL", + "video_file": "Charades/videos/55OAL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a blanket off a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.7> to <16.7>." + } + ] + }, + { + "video_id": "55OAL", + "question_id": "55OAL", + "video_file": "Charades/videos/55OAL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes the blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.7> to <16.7>." + } + ] + }, + { + "video_id": "CIG76", + "question_id": "CIG76", + "video_file": "Charades/videos/CIG76.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking bottle of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <17.5>." + } + ] + }, + { + "video_id": "MLS4T", + "question_id": "MLS4T", + "video_file": "Charades/videos/MLS4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person taking a drink from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <19.5>." + } + ] + }, + { + "video_id": "MLS4T", + "question_id": "MLS4T", + "video_file": "Charades/videos/MLS4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is putting a box on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <10.4>." + } + ] + }, + { + "video_id": "MLS4T", + "question_id": "MLS4T", + "video_file": "Charades/videos/MLS4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person drinks from a glass of iced tea'." + }, + { + "from": "gpt", + "value": "From <7.4> to <19.5>." + } + ] + }, + { + "video_id": "MLS4T", + "question_id": "MLS4T", + "video_file": "Charades/videos/MLS4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person throws a box of food onto the shelf'." + }, + { + "from": "gpt", + "value": "From <2.3> to <7.5>." + } + ] + }, + { + "video_id": "PK2F0", + "question_id": "PK2F0", + "video_file": "Charades/videos/PK2F0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they close the door'." + }, + { + "from": "gpt", + "value": "From <11.9> to <19.2>." + } + ] + }, + { + "video_id": "QWTGQ", + "question_id": "QWTGQ", + "video_file": "Charades/videos/QWTGQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person turns on the light in the hallway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.2> to <7.6>." + } + ] + }, + { + "video_id": "WKP5N", + "question_id": "WKP5N", + "video_file": "Charades/videos/WKP5N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes it with a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <19.1>." + } + ] + }, + { + "video_id": "WKP5N", + "question_id": "WKP5N", + "video_file": "Charades/videos/WKP5N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes out some medicine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <15.9>." + } + ] + }, + { + "video_id": "WKP5N", + "question_id": "WKP5N", + "video_file": "Charades/videos/WKP5N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes some medicine with some water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <15.9>." + } + ] + }, + { + "video_id": "WKP5N", + "question_id": "WKP5N", + "video_file": "Charades/videos/WKP5N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person opens a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <25.8>." + } + ] + }, + { + "video_id": "WKP5N", + "question_id": "WKP5N", + "video_file": "Charades/videos/WKP5N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person someone is sitting at a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "00HFP", + "question_id": "00HFP", + "video_file": "Charades/videos/00HFP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throwing them on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.6> to <21.9>." + } + ] + }, + { + "video_id": "00HFP", + "question_id": "00HFP", + "video_file": "Charades/videos/00HFP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws them down on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.6> to <21.9>." + } + ] + }, + { + "video_id": "N59H5", + "question_id": "N59H5", + "video_file": "Charades/videos/N59H5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person runs out of the room'." + }, + { + "from": "gpt", + "value": "From <16.6> to <22.0>." + } + ] + }, + { + "video_id": "N59H5", + "question_id": "N59H5", + "video_file": "Charades/videos/N59H5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person running out of the room'." + }, + { + "from": "gpt", + "value": "From <16.6> to <22.0>." + } + ] + }, + { + "video_id": "16D5O", + "question_id": "16D5O", + "video_file": "Charades/videos/16D5O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws a pillow onto it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <9.1>." + } + ] + }, + { + "video_id": "16D5O", + "question_id": "16D5O", + "video_file": "Charades/videos/16D5O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks into the pantry holding a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "16D5O", + "question_id": "16D5O", + "video_file": "Charades/videos/16D5O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person throws the pillow into the cabinet'." + }, + { + "from": "gpt", + "value": "From <1.5> to <9.1>." + } + ] + }, + { + "video_id": "V62NW", + "question_id": "V62NW", + "video_file": "Charades/videos/V62NW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes a book from a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <7.0>." + } + ] + }, + { + "video_id": "6VM42", + "question_id": "6VM42", + "video_file": "Charades/videos/6VM42.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a laptop from the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.8> to <16.8>." + } + ] + }, + { + "video_id": "RZY2I", + "question_id": "RZY2I", + "video_file": "Charades/videos/RZY2I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a window'." + }, + { + "from": "gpt", + "value": "From <5.0> to <10.0>." + } + ] + }, + { + "video_id": "RZY2I", + "question_id": "RZY2I", + "video_file": "Charades/videos/RZY2I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person open a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.0> to <10.0>." + } + ] + }, + { + "video_id": "RZY2I", + "question_id": "RZY2I", + "video_file": "Charades/videos/RZY2I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is standing in their bedroom undressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "BLIFO", + "question_id": "BLIFO", + "video_file": "Charades/videos/BLIFO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs in'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <7.5>." + } + ] + }, + { + "video_id": "BLIFO", + "question_id": "BLIFO", + "video_file": "Charades/videos/BLIFO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <9.1>." + } + ] + }, + { + "video_id": "BLIFO", + "question_id": "BLIFO", + "video_file": "Charades/videos/BLIFO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is running to the cabinet'." + }, + { + "from": "gpt", + "value": "From <0.4> to <7.5>." + } + ] + }, + { + "video_id": "MLLVP", + "question_id": "MLLVP", + "video_file": "Charades/videos/MLLVP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sneezing on themselves'." + }, + { + "from": "gpt", + "value": "From <10.7> to <15.9>." + } + ] + }, + { + "video_id": "MLLVP", + "question_id": "MLLVP", + "video_file": "Charades/videos/MLLVP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinks from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <10.5>." + } + ] + }, + { + "video_id": "2XLVN", + "question_id": "2XLVN", + "video_file": "Charades/videos/2XLVN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they open the box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.7> to <14.5>." + } + ] + }, + { + "video_id": "ZZ2SF", + "question_id": "ZZ2SF", + "video_file": "Charades/videos/ZZ2SF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinks some water from a bottle' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.2> to <32.0>." + } + ] + }, + { + "video_id": "YP1HM", + "question_id": "YP1HM", + "video_file": "Charades/videos/YP1HM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes notebook'." + }, + { + "from": "gpt", + "value": "From <18.6> to <24.0>." + } + ] + }, + { + "video_id": "YP1HM", + "question_id": "YP1HM", + "video_file": "Charades/videos/YP1HM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.7> to <18.6>." + } + ] + }, + { + "video_id": "A21M5", + "question_id": "A21M5", + "video_file": "Charades/videos/A21M5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walks out of the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <31.0>." + } + ] + }, + { + "video_id": "A21M5", + "question_id": "A21M5", + "video_file": "Charades/videos/A21M5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person walks back in the doorway with the broom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <31.0>." + } + ] + }, + { + "video_id": "A21M5", + "question_id": "A21M5", + "video_file": "Charades/videos/A21M5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens the box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "A21M5", + "question_id": "A21M5", + "video_file": "Charades/videos/A21M5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.9> to <8.8>." + } + ] + }, + { + "video_id": "A21M5", + "question_id": "A21M5", + "video_file": "Charades/videos/A21M5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.7> to <19.8>." + } + ] + }, + { + "video_id": "A21M5", + "question_id": "A21M5", + "video_file": "Charades/videos/A21M5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens the book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.6> to <7.3>." + } + ] + }, + { + "video_id": "I8WPI", + "question_id": "I8WPI", + "video_file": "Charades/videos/I8WPI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person washes a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.7>." + } + ] + }, + { + "video_id": "I8WPI", + "question_id": "I8WPI", + "video_file": "Charades/videos/I8WPI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person that they turn off the lights' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.7> to <27.7>." + } + ] + }, + { + "video_id": "I8WPI", + "question_id": "I8WPI", + "video_file": "Charades/videos/I8WPI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.7> to <27.7>." + } + ] + }, + { + "video_id": "1ROX7", + "question_id": "1ROX7", + "video_file": "Charades/videos/1ROX7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.4>." + } + ] + }, + { + "video_id": "1ROX7", + "question_id": "1ROX7", + "video_file": "Charades/videos/1ROX7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting at a table eating some food'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.4>." + } + ] + }, + { + "video_id": "1ROX7", + "question_id": "1ROX7", + "video_file": "Charades/videos/1ROX7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting at a table eating' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.4>." + } + ] + }, + { + "video_id": "CMFMH", + "question_id": "CMFMH", + "video_file": "Charades/videos/CMFMH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is fixing a door'." + }, + { + "from": "gpt", + "value": "From <17.4> to <33.0>." + } + ] + }, + { + "video_id": "CMFMH", + "question_id": "CMFMH", + "video_file": "Charades/videos/CMFMH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is working on eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "CMFMH", + "question_id": "CMFMH", + "video_file": "Charades/videos/CMFMH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "CMFMH", + "question_id": "CMFMH", + "video_file": "Charades/videos/CMFMH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person is fixing a cabinet door with a screwdriver'." + }, + { + "from": "gpt", + "value": "From <17.4> to <33.0>." + } + ] + }, + { + "video_id": "B6487", + "question_id": "B6487", + "video_file": "Charades/videos/B6487.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a refrigerator'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "J7TT5", + "question_id": "J7TT5", + "video_file": "Charades/videos/J7TT5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting on the stairs eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.6>." + } + ] + }, + { + "video_id": "J7TT5", + "question_id": "J7TT5", + "video_file": "Charades/videos/J7TT5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.6>." + } + ] + }, + { + "video_id": "293L4", + "question_id": "293L4", + "video_file": "Charades/videos/293L4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking on a stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <19.6>." + } + ] + }, + { + "video_id": "JBHOL", + "question_id": "JBHOL", + "video_file": "Charades/videos/JBHOL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a laughing person runs into their hallway grasping dishes'." + }, + { + "from": "gpt", + "value": "From <5.4> to <13.5>." + } + ] + }, + { + "video_id": "U3VOM", + "question_id": "U3VOM", + "video_file": "Charades/videos/U3VOM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is laying in their bed watching television' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.1> to <32.0>." + } + ] + }, + { + "video_id": "U3VOM", + "question_id": "U3VOM", + "video_file": "Charades/videos/U3VOM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws a pile of clothes on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.4> to <20.2>." + } + ] + }, + { + "video_id": "GC7L9", + "question_id": "GC7L9", + "video_file": "Charades/videos/GC7L9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <14.5>." + } + ] + }, + { + "video_id": "2ZTNY", + "question_id": "2ZTNY", + "video_file": "Charades/videos/2ZTNY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person puts a blanket onto the stairs' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.1>." + } + ] + }, + { + "video_id": "CP07P", + "question_id": "CP07P", + "video_file": "Charades/videos/CP07P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person is standing their smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <11.8>." + } + ] + }, + { + "video_id": "PTA97", + "question_id": "PTA97", + "video_file": "Charades/videos/PTA97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is seen fixing their hair at their closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.7>." + } + ] + }, + { + "video_id": "KXIMH", + "question_id": "KXIMH", + "video_file": "Charades/videos/KXIMH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting on a couch opening a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <8.0>." + } + ] + }, + { + "video_id": "KXIMH", + "question_id": "KXIMH", + "video_file": "Charades/videos/KXIMH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person in a recreational room is opening a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.8> to <8.0>." + } + ] + }, + { + "video_id": "KXIMH", + "question_id": "KXIMH", + "video_file": "Charades/videos/KXIMH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person opens the box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <8.0>." + } + ] + }, + { + "video_id": "RUG1J", + "question_id": "RUG1J", + "video_file": "Charades/videos/RUG1J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sneezing into a phone in a basement'." + }, + { + "from": "gpt", + "value": "From <10.0> to <14.4>." + } + ] + }, + { + "video_id": "TIJOY", + "question_id": "TIJOY", + "video_file": "Charades/videos/TIJOY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins eating' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.3> to <14.6>." + } + ] + }, + { + "video_id": "TIJOY", + "question_id": "TIJOY", + "video_file": "Charades/videos/TIJOY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes some food out' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.9> to <13.5>." + } + ] + }, + { + "video_id": "TIJOY", + "question_id": "TIJOY", + "video_file": "Charades/videos/TIJOY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.1> to <11.9>." + } + ] + }, + { + "video_id": "41SB2", + "question_id": "41SB2", + "video_file": "Charades/videos/41SB2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is standing on a chair changing a lightbulb'." + }, + { + "from": "gpt", + "value": "From <2.0> to <11.0>." + } + ] + }, + { + "video_id": "41SB2", + "question_id": "41SB2", + "video_file": "Charades/videos/41SB2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <16.0>." + } + ] + }, + { + "video_id": "41SB2", + "question_id": "41SB2", + "video_file": "Charades/videos/41SB2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person fixes a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <10.0>." + } + ] + }, + { + "video_id": "41SB2", + "question_id": "41SB2", + "video_file": "Charades/videos/41SB2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person stand up on the chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.0> to <11.0>." + } + ] + }, + { + "video_id": "41SB2", + "question_id": "41SB2", + "video_file": "Charades/videos/41SB2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes a cabinet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <16.0>." + } + ] + }, + { + "video_id": "41SB2", + "question_id": "41SB2", + "video_file": "Charades/videos/41SB2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person fix the light properly'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <10.0>." + } + ] + }, + { + "video_id": "7HVU8", + "question_id": "7HVU8", + "video_file": "Charades/videos/7HVU8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person tidying up by putting some clothes in a bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.0>." + } + ] + }, + { + "video_id": "UKOAU", + "question_id": "UKOAU", + "video_file": "Charades/videos/UKOAU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person is talking on the phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.7> to <31.0>." + } + ] + }, + { + "video_id": "UKOAU", + "question_id": "UKOAU", + "video_file": "Charades/videos/UKOAU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person running in place in the garage' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <23.9>." + } + ] + }, + { + "video_id": "CUD1T", + "question_id": "CUD1T", + "video_file": "Charades/videos/CUD1T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person laughing at their phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <23.2>." + } + ] + }, + { + "video_id": "CUD1T", + "question_id": "CUD1T", + "video_file": "Charades/videos/CUD1T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person laughs at something on their phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <23.2>." + } + ] + }, + { + "video_id": "CUD1T", + "question_id": "CUD1T", + "video_file": "Charades/videos/CUD1T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts laughing'." + }, + { + "from": "gpt", + "value": "From <11.1> to <23.2>." + } + ] + }, + { + "video_id": "Y8ON2", + "question_id": "Y8ON2", + "video_file": "Charades/videos/Y8ON2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <17.0>." + } + ] + }, + { + "video_id": "Y8ON2", + "question_id": "Y8ON2", + "video_file": "Charades/videos/Y8ON2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws a pillow on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <17.0>." + } + ] + }, + { + "video_id": "Y8ON2", + "question_id": "Y8ON2", + "video_file": "Charades/videos/Y8ON2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throw the pillow on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.0> to <17.0>." + } + ] + }, + { + "video_id": "Y8ON2", + "question_id": "Y8ON2", + "video_file": "Charades/videos/Y8ON2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person awakens on their bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <15.0>." + } + ] + }, + { + "video_id": "94YTY", + "question_id": "94YTY", + "video_file": "Charades/videos/94YTY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes off clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.8> to <19.0>." + } + ] + }, + { + "video_id": "94YTY", + "question_id": "94YTY", + "video_file": "Charades/videos/94YTY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person fixes their hair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <7.0>." + } + ] + }, + { + "video_id": "94YTY", + "question_id": "94YTY", + "video_file": "Charades/videos/94YTY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opening their front door begins undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.5> to <19.0>." + } + ] + }, + { + "video_id": "94YTY", + "question_id": "94YTY", + "video_file": "Charades/videos/94YTY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the door once again' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.5> to <22.3>." + } + ] + }, + { + "video_id": "94YTY", + "question_id": "94YTY", + "video_file": "Charades/videos/94YTY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <22.3>." + } + ] + }, + { + "video_id": "94YTY", + "question_id": "94YTY", + "video_file": "Charades/videos/94YTY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the clothes on a cabinet'." + }, + { + "from": "gpt", + "value": "From <17.7> to <30.3>." + } + ] + }, + { + "video_id": "94YTY", + "question_id": "94YTY", + "video_file": "Charades/videos/94YTY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person opens the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <7.5>." + } + ] + }, + { + "video_id": "NLWBW", + "question_id": "NLWBW", + "video_file": "Charades/videos/NLWBW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put away the groceries'." + }, + { + "from": "gpt", + "value": "From <28.6> to <35.2>." + } + ] + }, + { + "video_id": "NLWBW", + "question_id": "NLWBW", + "video_file": "Charades/videos/NLWBW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns off a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.2> to <21.6>." + } + ] + }, + { + "video_id": "NLWBW", + "question_id": "NLWBW", + "video_file": "Charades/videos/NLWBW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they go turn off the light in the garage' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <21.6>." + } + ] + }, + { + "video_id": "NLWBW", + "question_id": "NLWBW", + "video_file": "Charades/videos/NLWBW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put a bag of groceries on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.2> to <7.6>." + } + ] + }, + { + "video_id": "NLWBW", + "question_id": "NLWBW", + "video_file": "Charades/videos/NLWBW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person closed a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "E0AEN", + "question_id": "E0AEN", + "video_file": "Charades/videos/E0AEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts clothes from a bag in it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <12.2>." + } + ] + }, + { + "video_id": "E0AEN", + "question_id": "E0AEN", + "video_file": "Charades/videos/E0AEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens the doorknob to the closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "E0AEN", + "question_id": "E0AEN", + "video_file": "Charades/videos/E0AEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens the door to their wardrobe' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "E0AEN", + "question_id": "E0AEN", + "video_file": "Charades/videos/E0AEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts clothes away on the shelves'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <12.2>." + } + ] + }, + { + "video_id": "E0AEN", + "question_id": "E0AEN", + "video_file": "Charades/videos/E0AEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "WJEVU", + "question_id": "WJEVU", + "video_file": "Charades/videos/WJEVU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they take a drink from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.7> to <30.6>." + } + ] + }, + { + "video_id": "WJEVU", + "question_id": "WJEVU", + "video_file": "Charades/videos/WJEVU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person reaches to turn the light off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <44.4> to <51.0>." + } + ] + }, + { + "video_id": "WJEVU", + "question_id": "WJEVU", + "video_file": "Charades/videos/WJEVU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person snuggle with a blanket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <47.9> to <57.2>." + } + ] + }, + { + "video_id": "0JHMW", + "question_id": "0JHMW", + "video_file": "Charades/videos/0JHMW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts the glass down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.3> to <9.2>." + } + ] + }, + { + "video_id": "88TGX", + "question_id": "88TGX", + "video_file": "Charades/videos/88TGX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.3> to <8.2>." + } + ] + }, + { + "video_id": "88TGX", + "question_id": "88TGX", + "video_file": "Charades/videos/88TGX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening the refrigerator door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.3> to <8.2>." + } + ] + }, + { + "video_id": "WZRLQ", + "question_id": "WZRLQ", + "video_file": "Charades/videos/WZRLQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a glass of water from a nearby shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.7> to <34.1>." + } + ] + }, + { + "video_id": "T4HF9", + "question_id": "T4HF9", + "video_file": "Charades/videos/T4HF9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person open a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <21.7>." + } + ] + }, + { + "video_id": "T4HF9", + "question_id": "T4HF9", + "video_file": "Charades/videos/T4HF9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <21.7>." + } + ] + }, + { + "video_id": "T4HF9", + "question_id": "T4HF9", + "video_file": "Charades/videos/T4HF9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening a doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.3> to <21.7>." + } + ] + }, + { + "video_id": "29LL7", + "question_id": "29LL7", + "video_file": "Charades/videos/29LL7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a laughing person runs around the room grasping shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.6> to <16.2>." + } + ] + }, + { + "video_id": "29LL7", + "question_id": "29LL7", + "video_file": "Charades/videos/29LL7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person leaves laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.6> to <16.2>." + } + ] + }, + { + "video_id": "C5KNM", + "question_id": "C5KNM", + "video_file": "Charades/videos/C5KNM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks through the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.5> to <37.0>." + } + ] + }, + { + "video_id": "ZSJ2I", + "question_id": "ZSJ2I", + "video_file": "Charades/videos/ZSJ2I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drink water from glasses' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <17.7>." + } + ] + }, + { + "video_id": "ZSJ2I", + "question_id": "ZSJ2I", + "video_file": "Charades/videos/ZSJ2I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person b are sitting at a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "ZSJ2I", + "question_id": "ZSJ2I", + "video_file": "Charades/videos/ZSJ2I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walks through a doorway to another room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.7> to <23.7>." + } + ] + }, + { + "video_id": "ZSJ2I", + "question_id": "ZSJ2I", + "video_file": "Charades/videos/ZSJ2I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.4> to <32.0>." + } + ] + }, + { + "video_id": "ZSJ2I", + "question_id": "ZSJ2I", + "video_file": "Charades/videos/ZSJ2I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person two people sit at a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "ZSJ2I", + "question_id": "ZSJ2I", + "video_file": "Charades/videos/ZSJ2I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person another was closing the door in the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.4> to <32.0>." + } + ] + }, + { + "video_id": "ZSJ2I", + "question_id": "ZSJ2I", + "video_file": "Charades/videos/ZSJ2I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person a closes the front door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.4> to <32.0>." + } + ] + }, + { + "video_id": "WYM1M", + "question_id": "WYM1M", + "video_file": "Charades/videos/WYM1M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes the medicine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.9> to <31.0>." + } + ] + }, + { + "video_id": "WYM1M", + "question_id": "WYM1M", + "video_file": "Charades/videos/WYM1M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person playing on a mobile phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <19.8>." + } + ] + }, + { + "video_id": "WYM1M", + "question_id": "WYM1M", + "video_file": "Charades/videos/WYM1M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins playing a game on their cell phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <19.8>." + } + ] + }, + { + "video_id": "0WV3L", + "question_id": "0WV3L", + "video_file": "Charades/videos/0WV3L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.5> to <32.4>." + } + ] + }, + { + "video_id": "0WV3L", + "question_id": "0WV3L", + "video_file": "Charades/videos/0WV3L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the cabinet'." + }, + { + "from": "gpt", + "value": "From <27.3> to <35.4>." + } + ] + }, + { + "video_id": "0WV3L", + "question_id": "0WV3L", + "video_file": "Charades/videos/0WV3L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.5> to <32.4>." + } + ] + }, + { + "video_id": "C6B3T", + "question_id": "C6B3T", + "video_file": "Charades/videos/C6B3T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <10.8>." + } + ] + }, + { + "video_id": "C6B3T", + "question_id": "C6B3T", + "video_file": "Charades/videos/C6B3T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person sits at a desk working on a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "C6B3T", + "question_id": "C6B3T", + "video_file": "Charades/videos/C6B3T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begin laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.1> to <29.0>." + } + ] + }, + { + "video_id": "OL2QI", + "question_id": "OL2QI", + "video_file": "Charades/videos/OL2QI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws it on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <15.9>." + } + ] + }, + { + "video_id": "K95WN", + "question_id": "K95WN", + "video_file": "Charades/videos/K95WN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is talking on a phone'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.9>." + } + ] + }, + { + "video_id": "XS6PF", + "question_id": "XS6PF", + "video_file": "Charades/videos/XS6PF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person leaves the door open' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.4> to <7.8>." + } + ] + }, + { + "video_id": "XS6PF", + "question_id": "XS6PF", + "video_file": "Charades/videos/XS6PF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person flings open the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.4> to <7.8>." + } + ] + }, + { + "video_id": "XS6PF", + "question_id": "XS6PF", + "video_file": "Charades/videos/XS6PF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits on the couch' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.7> to <22.5>." + } + ] + }, + { + "video_id": "XS6PF", + "question_id": "XS6PF", + "video_file": "Charades/videos/XS6PF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <7.8>." + } + ] + }, + { + "video_id": "XS6PF", + "question_id": "XS6PF", + "video_file": "Charades/videos/XS6PF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person bursts into the front door holding a blanket' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.8> to <9.8>." + } + ] + }, + { + "video_id": "XS6PF", + "question_id": "XS6PF", + "video_file": "Charades/videos/XS6PF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.7> to <13.9>." + } + ] + }, + { + "video_id": "XS6PF", + "question_id": "XS6PF", + "video_file": "Charades/videos/XS6PF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws the blanket on the couch' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.4> to <10.0>." + } + ] + }, + { + "video_id": "78R4Y", + "question_id": "78R4Y", + "video_file": "Charades/videos/78R4Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens the refrigerator door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <5.4>." + } + ] + }, + { + "video_id": "78R4Y", + "question_id": "78R4Y", + "video_file": "Charades/videos/78R4Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is putting away groceries in refrigerator'." + }, + { + "from": "gpt", + "value": "From <1.6> to <6.7>." + } + ] + }, + { + "video_id": "78R4Y", + "question_id": "78R4Y", + "video_file": "Charades/videos/78R4Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.0> to <17.0>." + } + ] + }, + { + "video_id": "78R4Y", + "question_id": "78R4Y", + "video_file": "Charades/videos/78R4Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person open the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "78R4Y", + "question_id": "78R4Y", + "video_file": "Charades/videos/78R4Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing its door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.0> to <17.0>." + } + ] + }, + { + "video_id": "A10NL", + "question_id": "A10NL", + "video_file": "Charades/videos/A10NL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes a laptop from a desk' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <19.6>." + } + ] + }, + { + "video_id": "1OF1O", + "question_id": "1OF1O", + "video_file": "Charades/videos/1OF1O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person walked into the doorway'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "1OF1O", + "question_id": "1OF1O", + "video_file": "Charades/videos/1OF1O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turned on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.2> to <5.4>." + } + ] + }, + { + "video_id": "1OF1O", + "question_id": "1OF1O", + "video_file": "Charades/videos/1OF1O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person turns on a light in a bathroom'." + }, + { + "from": "gpt", + "value": "From <0.2> to <5.4>." + } + ] + }, + { + "video_id": "ZKC7M", + "question_id": "ZKC7M", + "video_file": "Charades/videos/ZKC7M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person is awakening in the bright light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <24.8>." + } + ] + }, + { + "video_id": "DIB6D", + "question_id": "DIB6D", + "video_file": "Charades/videos/DIB6D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinks a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.9> to <24.4>." + } + ] + }, + { + "video_id": "DIB6D", + "question_id": "DIB6D", + "video_file": "Charades/videos/DIB6D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person in the bathroom is smiling in the mirror'." + }, + { + "from": "gpt", + "value": "From <9.2> to <18.1>." + } + ] + }, + { + "video_id": "16IDQ", + "question_id": "16IDQ", + "video_file": "Charades/videos/16IDQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person pushes open a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "16IDQ", + "question_id": "16IDQ", + "video_file": "Charades/videos/16IDQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "16IDQ", + "question_id": "16IDQ", + "video_file": "Charades/videos/16IDQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sneezing into a camera' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <20.5>." + } + ] + }, + { + "video_id": "G9JMG", + "question_id": "G9JMG", + "video_file": "Charades/videos/G9JMG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the pillow in their lap' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <25.2>." + } + ] + }, + { + "video_id": "G9JMG", + "question_id": "G9JMG", + "video_file": "Charades/videos/G9JMG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "G9JMG", + "question_id": "G9JMG", + "video_file": "Charades/videos/G9JMG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the pillow down'." + }, + { + "from": "gpt", + "value": "From <20.1> to <25.2>." + } + ] + }, + { + "video_id": "G9JMG", + "question_id": "G9JMG", + "video_file": "Charades/videos/G9JMG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the pillow on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <25.2>." + } + ] + }, + { + "video_id": "QGIP7", + "question_id": "QGIP7", + "video_file": "Charades/videos/QGIP7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <21.0>." + } + ] + }, + { + "video_id": "QGIP7", + "question_id": "QGIP7", + "video_file": "Charades/videos/QGIP7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks over to a shelf smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.2> to <14.3>." + } + ] + }, + { + "video_id": "WW4MY", + "question_id": "WW4MY", + "video_file": "Charades/videos/WW4MY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they close the box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.5> to <24.2>." + } + ] + }, + { + "video_id": "WW4MY", + "question_id": "WW4MY", + "video_file": "Charades/videos/WW4MY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting on the couch' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.7> to <15.6>." + } + ] + }, + { + "video_id": "WW4MY", + "question_id": "WW4MY", + "video_file": "Charades/videos/WW4MY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting on the sofa drinking' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.7> to <15.6>." + } + ] + }, + { + "video_id": "WW4MY", + "question_id": "WW4MY", + "video_file": "Charades/videos/WW4MY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a drink sits the glass back down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <27.6>." + } + ] + }, + { + "video_id": "WW4MY", + "question_id": "WW4MY", + "video_file": "Charades/videos/WW4MY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating chips' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.0> to <9.4>." + } + ] + }, + { + "video_id": "WW4MY", + "question_id": "WW4MY", + "video_file": "Charades/videos/WW4MY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person a half eaten sandwich on one' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <9.4>." + } + ] + }, + { + "video_id": "X6JKA", + "question_id": "X6JKA", + "video_file": "Charades/videos/X6JKA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "N915C", + "question_id": "N915C", + "video_file": "Charades/videos/N915C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person gets undressed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <16.6>." + } + ] + }, + { + "video_id": "IGEU5", + "question_id": "IGEU5", + "video_file": "Charades/videos/IGEU5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person throws the pillow onto the sofa'." + }, + { + "from": "gpt", + "value": "From <16.2> to <21.7>." + } + ] + }, + { + "video_id": "DOL47", + "question_id": "DOL47", + "video_file": "Charades/videos/DOL47.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <25.4>." + } + ] + }, + { + "video_id": "0JB3D", + "question_id": "0JB3D", + "video_file": "Charades/videos/0JB3D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "0JB3D", + "question_id": "0JB3D", + "video_file": "Charades/videos/0JB3D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes out a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.1> to <10.5>." + } + ] + }, + { + "video_id": "0JB3D", + "question_id": "0JB3D", + "video_file": "Charades/videos/0JB3D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "8EU89", + "question_id": "8EU89", + "video_file": "Charades/videos/8EU89.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <7.6>." + } + ] + }, + { + "video_id": "8EU89", + "question_id": "8EU89", + "video_file": "Charades/videos/8EU89.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.5> to <28.2>." + } + ] + }, + { + "video_id": "8EU89", + "question_id": "8EU89", + "video_file": "Charades/videos/8EU89.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person sits on a pillow on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.8> to <11.5>." + } + ] + }, + { + "video_id": "8EU89", + "question_id": "8EU89", + "video_file": "Charades/videos/8EU89.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sitting on a chair'." + }, + { + "from": "gpt", + "value": "From <5.8> to <11.5>." + } + ] + }, + { + "video_id": "TODWV", + "question_id": "TODWV", + "video_file": "Charades/videos/TODWV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakens' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "TODWV", + "question_id": "TODWV", + "video_file": "Charades/videos/TODWV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person dresses in clothes'." + }, + { + "from": "gpt", + "value": "From <21.6> to <32.0>." + } + ] + }, + { + "video_id": "TODWV", + "question_id": "TODWV", + "video_file": "Charades/videos/TODWV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins dressing in clothes in front of a mirror' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <32.0>." + } + ] + }, + { + "video_id": "TODWV", + "question_id": "TODWV", + "video_file": "Charades/videos/TODWV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts on glasses' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <19.5>." + } + ] + }, + { + "video_id": "TODWV", + "question_id": "TODWV", + "video_file": "Charades/videos/TODWV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person lying on a bed awakens' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "LWP5F", + "question_id": "LWP5F", + "video_file": "Charades/videos/LWP5F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person begins opening a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.2> to <20.0>." + } + ] + }, + { + "video_id": "V2MHO", + "question_id": "V2MHO", + "video_file": "Charades/videos/V2MHO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws a blanket across the room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.8> to <17.8>." + } + ] + }, + { + "video_id": "V2MHO", + "question_id": "V2MHO", + "video_file": "Charades/videos/V2MHO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person continue looking out the window'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.4>." + } + ] + }, + { + "video_id": "V2MHO", + "question_id": "V2MHO", + "video_file": "Charades/videos/V2MHO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looking out a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.4>." + } + ] + }, + { + "video_id": "A58GR", + "question_id": "A58GR", + "video_file": "Charades/videos/A58GR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person laugh as they put a towel in it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.0> to <25.4>." + } + ] + }, + { + "video_id": "A58GR", + "question_id": "A58GR", + "video_file": "Charades/videos/A58GR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts clothes in it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <26.1>." + } + ] + }, + { + "video_id": "A58GR", + "question_id": "A58GR", + "video_file": "Charades/videos/A58GR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes a box from a shelf'." + }, + { + "from": "gpt", + "value": "From <1.4> to <9.4>." + } + ] + }, + { + "video_id": "A58GR", + "question_id": "A58GR", + "video_file": "Charades/videos/A58GR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes a box off a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.4> to <9.4>." + } + ] + }, + { + "video_id": "A58GR", + "question_id": "A58GR", + "video_file": "Charades/videos/A58GR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put a towel inside' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.0> to <25.4>." + } + ] + }, + { + "video_id": "0OUEP", + "question_id": "0OUEP", + "video_file": "Charades/videos/0OUEP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.2> to <33.0>." + } + ] + }, + { + "video_id": "0OUEP", + "question_id": "0OUEP", + "video_file": "Charades/videos/0OUEP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person is holding a camera' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.7> to <10.6>." + } + ] + }, + { + "video_id": "0OUEP", + "question_id": "0OUEP", + "video_file": "Charades/videos/0OUEP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes a door'." + }, + { + "from": "gpt", + "value": "From <26.2> to <33.0>." + } + ] + }, + { + "video_id": "CJ0U1", + "question_id": "CJ0U1", + "video_file": "Charades/videos/CJ0U1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person turns the light off' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.8> to <24.3>." + } + ] + }, + { + "video_id": "CJ0U1", + "question_id": "CJ0U1", + "video_file": "Charades/videos/CJ0U1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is opening the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "CJ0U1", + "question_id": "CJ0U1", + "video_file": "Charades/videos/CJ0U1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.8> to <24.3>." + } + ] + }, + { + "video_id": "I68J0", + "question_id": "I68J0", + "video_file": "Charades/videos/I68J0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the book on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.2> to <12.5>." + } + ] + }, + { + "video_id": "I68J0", + "question_id": "I68J0", + "video_file": "Charades/videos/I68J0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a book from the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <9.1>." + } + ] + }, + { + "video_id": "I68J0", + "question_id": "I68J0", + "video_file": "Charades/videos/I68J0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a picture of themselves' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.5> to <35.1>." + } + ] + }, + { + "video_id": "I68J0", + "question_id": "I68J0", + "video_file": "Charades/videos/I68J0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person grabs a book in the closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.7> to <24.4>." + } + ] + }, + { + "video_id": "I68J0", + "question_id": "I68J0", + "video_file": "Charades/videos/I68J0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.1> to <10.8>." + } + ] + }, + { + "video_id": "I68J0", + "question_id": "I68J0", + "video_file": "Charades/videos/I68J0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs into the closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "I68J0", + "question_id": "I68J0", + "video_file": "Charades/videos/I68J0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a book off of a shelf'." + }, + { + "from": "gpt", + "value": "From <3.2> to <9.1>." + } + ] + }, + { + "video_id": "I68J0", + "question_id": "I68J0", + "video_file": "Charades/videos/I68J0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs to a closet'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "GNMBV", + "question_id": "GNMBV", + "video_file": "Charades/videos/GNMBV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holds a medicine bottle' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.5> to <24.0>." + } + ] + }, + { + "video_id": "PEY34", + "question_id": "PEY34", + "video_file": "Charades/videos/PEY34.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is drinking a glass of water'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "PEY34", + "question_id": "PEY34", + "video_file": "Charades/videos/PEY34.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a broom up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.8> to <23.9>." + } + ] + }, + { + "video_id": "PEY34", + "question_id": "PEY34", + "video_file": "Charades/videos/PEY34.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.6> to <17.8>." + } + ] + }, + { + "video_id": "PEY34", + "question_id": "PEY34", + "video_file": "Charades/videos/PEY34.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting a box on top of a shelf'." + }, + { + "from": "gpt", + "value": "From <9.6> to <17.8>." + } + ] + }, + { + "video_id": "PEY34", + "question_id": "PEY34", + "video_file": "Charades/videos/PEY34.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person begins laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.2> to <10.6>." + } + ] + }, + { + "video_id": "D9YWY", + "question_id": "D9YWY", + "video_file": "Charades/videos/D9YWY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting stuff in a fridge from a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <37.7> to <43.6>." + } + ] + }, + { + "video_id": "HSSOZ", + "question_id": "HSSOZ", + "video_file": "Charades/videos/HSSOZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put the dish in the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.3> to <33.9>." + } + ] + }, + { + "video_id": "J867Z", + "question_id": "J867Z", + "video_file": "Charades/videos/J867Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws a book down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.3> to <20.0>." + } + ] + }, + { + "video_id": "J867Z", + "question_id": "J867Z", + "video_file": "Charades/videos/J867Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they throw down the book they were also holding'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.3> to <20.0>." + } + ] + }, + { + "video_id": "ZAJYX", + "question_id": "ZAJYX", + "video_file": "Charades/videos/ZAJYX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is cooking food on a stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.2>." + } + ] + }, + { + "video_id": "ZAJYX", + "question_id": "ZAJYX", + "video_file": "Charades/videos/ZAJYX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is cooking some food on a stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.2>." + } + ] + }, + { + "video_id": "55NNF", + "question_id": "55NNF", + "video_file": "Charades/videos/55NNF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins opening a refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.8> to <21.3>." + } + ] + }, + { + "video_id": "55NNF", + "question_id": "55NNF", + "video_file": "Charades/videos/55NNF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open the refrigerator'." + }, + { + "from": "gpt", + "value": "From <14.8> to <21.3>." + } + ] + }, + { + "video_id": "55NNF", + "question_id": "55NNF", + "video_file": "Charades/videos/55NNF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is cooking'." + }, + { + "from": "gpt", + "value": "From <2.6> to <17.9>." + } + ] + }, + { + "video_id": "55NNF", + "question_id": "55NNF", + "video_file": "Charades/videos/55NNF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person cooks food on a stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.6> to <17.9>." + } + ] + }, + { + "video_id": "8ZO48", + "question_id": "8ZO48", + "video_file": "Charades/videos/8ZO48.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits in a chair'." + }, + { + "from": "gpt", + "value": "From <16.5> to <23.4>." + } + ] + }, + { + "video_id": "8ZO48", + "question_id": "8ZO48", + "video_file": "Charades/videos/8ZO48.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throw the pillow aside'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.7> to <11.3>." + } + ] + }, + { + "video_id": "8ZO48", + "question_id": "8ZO48", + "video_file": "Charades/videos/8ZO48.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.5> to <23.4>." + } + ] + }, + { + "video_id": "8ZO48", + "question_id": "8ZO48", + "video_file": "Charades/videos/8ZO48.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they throw down a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.7> to <11.3>." + } + ] + }, + { + "video_id": "8ZO48", + "question_id": "8ZO48", + "video_file": "Charades/videos/8ZO48.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.7> to <11.3>." + } + ] + }, + { + "video_id": "8ZO48", + "question_id": "8ZO48", + "video_file": "Charades/videos/8ZO48.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is awakened on the bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <13.3>." + } + ] + }, + { + "video_id": "5L09N", + "question_id": "5L09N", + "video_file": "Charades/videos/5L09N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a picture'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.1> to <40.0>." + } + ] + }, + { + "video_id": "DVDU2", + "question_id": "DVDU2", + "video_file": "Charades/videos/DVDU2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats snack food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.7> to <30.2>." + } + ] + }, + { + "video_id": "DVDU2", + "question_id": "DVDU2", + "video_file": "Charades/videos/DVDU2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.7> to <30.2>." + } + ] + }, + { + "video_id": "RV48T", + "question_id": "RV48T", + "video_file": "Charades/videos/RV48T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <11.0>." + } + ] + }, + { + "video_id": "RV48T", + "question_id": "RV48T", + "video_file": "Charades/videos/RV48T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'closes the door.the person stops' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <11.0>." + } + ] + }, + { + "video_id": "TZ548", + "question_id": "TZ548", + "video_file": "Charades/videos/TZ548.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is walking through a doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <12.8>." + } + ] + }, + { + "video_id": "TZ548", + "question_id": "TZ548", + "video_file": "Charades/videos/TZ548.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks through a doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.4> to <12.8>." + } + ] + }, + { + "video_id": "FDYQZ", + "question_id": "FDYQZ", + "video_file": "Charades/videos/FDYQZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person one is undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.1> to <16.1>." + } + ] + }, + { + "video_id": "FDYQZ", + "question_id": "FDYQZ", + "video_file": "Charades/videos/FDYQZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person standing at a sink starts to undress' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.1> to <16.1>." + } + ] + }, + { + "video_id": "FDYQZ", + "question_id": "FDYQZ", + "video_file": "Charades/videos/FDYQZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person is undressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.1> to <16.1>." + } + ] + }, + { + "video_id": "1HL2K", + "question_id": "1HL2K", + "video_file": "Charades/videos/1HL2K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person goes back to looking out the window'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "1HL2K", + "question_id": "1HL2K", + "video_file": "Charades/videos/1HL2K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sitting looking out the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "TKAFY", + "question_id": "TKAFY", + "video_file": "Charades/videos/TKAFY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.8>." + } + ] + }, + { + "video_id": "TKAFY", + "question_id": "TKAFY", + "video_file": "Charades/videos/TKAFY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person get dressed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <20.1>." + } + ] + }, + { + "video_id": "TKAFY", + "question_id": "TKAFY", + "video_file": "Charades/videos/TKAFY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins dressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.4> to <20.1>." + } + ] + }, + { + "video_id": "6FYKZ", + "question_id": "6FYKZ", + "video_file": "Charades/videos/6FYKZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person puts away a broom beside the washing machine' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "DN0CB", + "question_id": "DN0CB", + "video_file": "Charades/videos/DN0CB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person an unidentified hand turns on a light switch' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <66.4> to <75.8>." + } + ] + }, + { + "video_id": "DN0CB", + "question_id": "DN0CB", + "video_file": "Charades/videos/DN0CB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person leaves holding the bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.5> to <31.9>." + } + ] + }, + { + "video_id": "DN0CB", + "question_id": "DN0CB", + "video_file": "Charades/videos/DN0CB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person grasps the light switch' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <66.4> to <75.8>." + } + ] + }, + { + "video_id": "DN0CB", + "question_id": "DN0CB", + "video_file": "Charades/videos/DN0CB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <66.8> to <77.0>." + } + ] + }, + { + "video_id": "DN0CB", + "question_id": "DN0CB", + "video_file": "Charades/videos/DN0CB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens the bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.1> to <31.5>." + } + ] + }, + { + "video_id": "DN0CB", + "question_id": "DN0CB", + "video_file": "Charades/videos/DN0CB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a full person enters the room holding a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.5> to <31.9>." + } + ] + }, + { + "video_id": "UJOLG", + "question_id": "UJOLG", + "video_file": "Charades/videos/UJOLG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take off shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.3> to <20.1>." + } + ] + }, + { + "video_id": "UJOLG", + "question_id": "UJOLG", + "video_file": "Charades/videos/UJOLG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "UJOLG", + "question_id": "UJOLG", + "video_file": "Charades/videos/UJOLG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "UJOLG", + "question_id": "UJOLG", + "video_file": "Charades/videos/UJOLG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes off their shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.3> to <20.1>." + } + ] + }, + { + "video_id": "AZZ7V", + "question_id": "AZZ7V", + "video_file": "Charades/videos/AZZ7V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person holds a pillow in bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.9>." + } + ] + }, + { + "video_id": "AZZ7V", + "question_id": "AZZ7V", + "video_file": "Charades/videos/AZZ7V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person lying on a bed snuggling a pillow awakens' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <20.8>." + } + ] + }, + { + "video_id": "M43KR", + "question_id": "M43KR", + "video_file": "Charades/videos/M43KR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a picture of their food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.1> to <21.3>." + } + ] + }, + { + "video_id": "PC3WT", + "question_id": "PC3WT", + "video_file": "Charades/videos/PC3WT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.8> to <25.8>." + } + ] + }, + { + "video_id": "PC3WT", + "question_id": "PC3WT", + "video_file": "Charades/videos/PC3WT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is awakening'." + }, + { + "from": "gpt", + "value": "From <1.8> to <7.9>." + } + ] + }, + { + "video_id": "PC3WT", + "question_id": "PC3WT", + "video_file": "Charades/videos/PC3WT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.8> to <25.8>." + } + ] + }, + { + "video_id": "0ETKN", + "question_id": "0ETKN", + "video_file": "Charades/videos/0ETKN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opening the food bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.2> to <9.3>." + } + ] + }, + { + "video_id": "XS7RN", + "question_id": "XS7RN", + "video_file": "Charades/videos/XS7RN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throw a book on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.4> to <26.9>." + } + ] + }, + { + "video_id": "XS7RN", + "question_id": "XS7RN", + "video_file": "Charades/videos/XS7RN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <19.5>." + } + ] + }, + { + "video_id": "XS7RN", + "question_id": "XS7RN", + "video_file": "Charades/videos/XS7RN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put the box on a washing machine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.0> to <33.0>." + } + ] + }, + { + "video_id": "ZWDVG", + "question_id": "ZWDVG", + "video_file": "Charades/videos/ZWDVG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'awakening another person who starts sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <14.2>." + } + ] + }, + { + "video_id": "ZWDVG", + "question_id": "ZWDVG", + "video_file": "Charades/videos/ZWDVG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person throws a bag down on a small table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <7.4>." + } + ] + }, + { + "video_id": "7L18C", + "question_id": "7L18C", + "video_file": "Charades/videos/7L18C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sitting on the bed looking at his laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.9> to <21.7>." + } + ] + }, + { + "video_id": "7L18C", + "question_id": "7L18C", + "video_file": "Charades/videos/7L18C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person playing a game on their laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.1> to <10.6>." + } + ] + }, + { + "video_id": "O0C2Z", + "question_id": "O0C2Z", + "video_file": "Charades/videos/O0C2Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person turns on a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.4> to <7.8>." + } + ] + }, + { + "video_id": "CEL4V", + "question_id": "CEL4V", + "video_file": "Charades/videos/CEL4V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes a glass from the cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.9> to <6.6>." + } + ] + }, + { + "video_id": "901GH", + "question_id": "901GH", + "video_file": "Charades/videos/901GH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person opens two doors on a wall cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <6.3>." + } + ] + }, + { + "video_id": "901GH", + "question_id": "901GH", + "video_file": "Charades/videos/901GH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'holding a towel the person begins sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.2> to <13.5>." + } + ] + }, + { + "video_id": "901GH", + "question_id": "901GH", + "video_file": "Charades/videos/901GH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening the window in the hallway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "N4SA2", + "question_id": "N4SA2", + "video_file": "Charades/videos/N4SA2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.3> to <39.1>." + } + ] + }, + { + "video_id": "N4SA2", + "question_id": "N4SA2", + "video_file": "Charades/videos/N4SA2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put a towel around their neck' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.8> to <29.8>." + } + ] + }, + { + "video_id": "GC51E", + "question_id": "GC51E", + "video_file": "Charades/videos/GC51E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats the apple that was in the bag'." + }, + { + "from": "gpt", + "value": "From <63.2> to <87.5>." + } + ] + }, + { + "video_id": "GC51E", + "question_id": "GC51E", + "video_file": "Charades/videos/GC51E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eat it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <63.2> to <87.5>." + } + ] + }, + { + "video_id": "GC51E", + "question_id": "GC51E", + "video_file": "Charades/videos/GC51E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person on their stairs takes off their shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <37.4> to <56.7>." + } + ] + }, + { + "video_id": "GC51E", + "question_id": "GC51E", + "video_file": "Charades/videos/GC51E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take their shoes off' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <37.4> to <56.7>." + } + ] + }, + { + "video_id": "JQW5G", + "question_id": "JQW5G", + "video_file": "Charades/videos/JQW5G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting them on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.4> to <31.0>." + } + ] + }, + { + "video_id": "JQW5G", + "question_id": "JQW5G", + "video_file": "Charades/videos/JQW5G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person is snuggling under a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.1> to <15.2>." + } + ] + }, + { + "video_id": "JQW5G", + "question_id": "JQW5G", + "video_file": "Charades/videos/JQW5G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting a bag on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.4> to <31.0>." + } + ] + }, + { + "video_id": "UO5PM", + "question_id": "UO5PM", + "video_file": "Charades/videos/UO5PM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes a towel' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <12.0>." + } + ] + }, + { + "video_id": "UO5PM", + "question_id": "UO5PM", + "video_file": "Charades/videos/UO5PM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "UO5PM", + "question_id": "UO5PM", + "video_file": "Charades/videos/UO5PM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on the lights'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "UO5PM", + "question_id": "UO5PM", + "video_file": "Charades/videos/UO5PM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turning on the light as they enter'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "8175C", + "question_id": "8175C", + "video_file": "Charades/videos/8175C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opening a door'." + }, + { + "from": "gpt", + "value": "From <27.8> to <33.2>." + } + ] + }, + { + "video_id": "YS9RE", + "question_id": "YS9RE", + "video_file": "Charades/videos/YS9RE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they take a picture of themselves' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <29.0>." + } + ] + }, + { + "video_id": "MSQA1", + "question_id": "MSQA1", + "video_file": "Charades/videos/MSQA1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begin to eat the sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <50.0> to <54.0>." + } + ] + }, + { + "video_id": "MSQA1", + "question_id": "MSQA1", + "video_file": "Charades/videos/MSQA1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sit in a chair at the kitchen table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <48.2> to <54.0>." + } + ] + }, + { + "video_id": "MSQA1", + "question_id": "MSQA1", + "video_file": "Charades/videos/MSQA1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.3> to <18.8>." + } + ] + }, + { + "video_id": "8LWV7", + "question_id": "8LWV7", + "video_file": "Charades/videos/8LWV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they awaken' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <13.8>." + } + ] + }, + { + "video_id": "8LWV7", + "question_id": "8LWV7", + "video_file": "Charades/videos/8LWV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person awakens on a bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <13.8>." + } + ] + }, + { + "video_id": "60W9B", + "question_id": "60W9B", + "video_file": "Charades/videos/60W9B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.8> to <7.6>." + } + ] + }, + { + "video_id": "E75C2", + "question_id": "E75C2", + "video_file": "Charades/videos/E75C2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person pours a cup of coffee' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <13.8>." + } + ] + }, + { + "video_id": "E75C2", + "question_id": "E75C2", + "video_file": "Charades/videos/E75C2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looks out the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.9> to <32.0>." + } + ] + }, + { + "video_id": "SSUZA", + "question_id": "SSUZA", + "video_file": "Charades/videos/SSUZA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person try to open the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.8> to <19.9>." + } + ] + }, + { + "video_id": "9S0QC", + "question_id": "9S0QC", + "video_file": "Charades/videos/9S0QC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is throwing a bag on the floor'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "9S0QC", + "question_id": "9S0QC", + "video_file": "Charades/videos/9S0QC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting on shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.1> to <21.0>." + } + ] + }, + { + "video_id": "IDACP", + "question_id": "IDACP", + "video_file": "Charades/videos/IDACP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person dresses into a coat' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <14.4>." + } + ] + }, + { + "video_id": "QLQNJ", + "question_id": "QLQNJ", + "video_file": "Charades/videos/QLQNJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs in' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.4> to <15.7>." + } + ] + }, + { + "video_id": "ATI0Z", + "question_id": "ATI0Z", + "video_file": "Charades/videos/ATI0Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.8>." + } + ] + }, + { + "video_id": "ATI0Z", + "question_id": "ATI0Z", + "video_file": "Charades/videos/ATI0Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns on a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <17.1>." + } + ] + }, + { + "video_id": "OJ6SX", + "question_id": "OJ6SX", + "video_file": "Charades/videos/OJ6SX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking from a cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.7> to <20.8>." + } + ] + }, + { + "video_id": "OJ6SX", + "question_id": "OJ6SX", + "video_file": "Charades/videos/OJ6SX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <17.6>." + } + ] + }, + { + "video_id": "EWH3D", + "question_id": "EWH3D", + "video_file": "Charades/videos/EWH3D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person leaving a table that they were sitting at' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "PAYYN", + "question_id": "PAYYN", + "video_file": "Charades/videos/PAYYN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person playing on their laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.9> to <23.0>." + } + ] + }, + { + "video_id": "EDPTO", + "question_id": "EDPTO", + "video_file": "Charades/videos/EDPTO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person looks out a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <22.8>." + } + ] + }, + { + "video_id": "EDPTO", + "question_id": "EDPTO", + "video_file": "Charades/videos/EDPTO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person picking up a sandwich to eat'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.9> to <37.3>." + } + ] + }, + { + "video_id": "EDPTO", + "question_id": "EDPTO", + "video_file": "Charades/videos/EDPTO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person holding a blanket looks out of a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.8> to <22.8>." + } + ] + }, + { + "video_id": "QZP8N", + "question_id": "QZP8N", + "video_file": "Charades/videos/QZP8N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person smiles when they look in the mirror'." + }, + { + "from": "gpt", + "value": "From <3.1> to <10.3>." + } + ] + }, + { + "video_id": "WIYE2", + "question_id": "WIYE2", + "video_file": "Charades/videos/WIYE2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts the water down on the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.1> to <14.3>." + } + ] + }, + { + "video_id": "29757", + "question_id": "29757", + "video_file": "Charades/videos/29757.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person runs to the mirror'." + }, + { + "from": "gpt", + "value": "From <20.1> to <26.9>." + } + ] + }, + { + "video_id": "DVO31", + "question_id": "DVO31", + "video_file": "Charades/videos/DVO31.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person finally drinking from a glass in a bedroom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <34.7> to <43.0>." + } + ] + }, + { + "video_id": "DVO31", + "question_id": "DVO31", + "video_file": "Charades/videos/DVO31.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throwing clothes into a hamper' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.7> to <19.5>." + } + ] + }, + { + "video_id": "W6210", + "question_id": "W6210", + "video_file": "Charades/videos/W6210.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts them on a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.6> to <20.6>." + } + ] + }, + { + "video_id": "W6210", + "question_id": "W6210", + "video_file": "Charades/videos/W6210.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <32.0>." + } + ] + }, + { + "video_id": "W6210", + "question_id": "W6210", + "video_file": "Charades/videos/W6210.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the shoes on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.6> to <20.6>." + } + ] + }, + { + "video_id": "W6210", + "question_id": "W6210", + "video_file": "Charades/videos/W6210.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is putting shoes on the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.6> to <20.6>." + } + ] + }, + { + "video_id": "GKII3", + "question_id": "GKII3", + "video_file": "Charades/videos/GKII3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <28.0>." + } + ] + }, + { + "video_id": "EK19H", + "question_id": "EK19H", + "video_file": "Charades/videos/EK19H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.5> to <15.6>." + } + ] + }, + { + "video_id": "1HAYV", + "question_id": "1HAYV", + "video_file": "Charades/videos/1HAYV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "1HAYV", + "question_id": "1HAYV", + "video_file": "Charades/videos/1HAYV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sitting in a chair eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "1HAYV", + "question_id": "1HAYV", + "video_file": "Charades/videos/1HAYV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking a cup of coffee'." + }, + { + "from": "gpt", + "value": "From <1.6> to <8.2>." + } + ] + }, + { + "video_id": "1HAYV", + "question_id": "1HAYV", + "video_file": "Charades/videos/1HAYV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking out of a glass'." + }, + { + "from": "gpt", + "value": "From <1.6> to <8.2>." + } + ] + }, + { + "video_id": "FCDXN", + "question_id": "FCDXN", + "video_file": "Charades/videos/FCDXN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put a glass on table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <18.8>." + } + ] + }, + { + "video_id": "FL3CG", + "question_id": "FL3CG", + "video_file": "Charades/videos/FL3CG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is smiling at a picture'." + }, + { + "from": "gpt", + "value": "From <1.4> to <9.0>." + } + ] + }, + { + "video_id": "FYR3P", + "question_id": "FYR3P", + "video_file": "Charades/videos/FYR3P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a towel from the bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.5> to <17.8>." + } + ] + }, + { + "video_id": "0W39H", + "question_id": "0W39H", + "video_file": "Charades/videos/0W39H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is pouring liquids between 2 cups' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.2> to <9.8>." + } + ] + }, + { + "video_id": "0W39H", + "question_id": "0W39H", + "video_file": "Charades/videos/0W39H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is pouring fresh coffee into their glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <9.8>." + } + ] + }, + { + "video_id": "6ZSB2", + "question_id": "6ZSB2", + "video_file": "Charades/videos/6ZSB2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs into a hallway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "6ZSB2", + "question_id": "6ZSB2", + "video_file": "Charades/videos/6ZSB2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open a bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.6> to <7.2>." + } + ] + }, + { + "video_id": "MD6C0", + "question_id": "MD6C0", + "video_file": "Charades/videos/MD6C0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a man opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "TU3IG", + "question_id": "TU3IG", + "video_file": "Charades/videos/TU3IG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person eats food out of a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.4> to <11.9>." + } + ] + }, + { + "video_id": "K3193", + "question_id": "K3193", + "video_file": "Charades/videos/K3193.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sneezing by a sofa'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.6> to <19.6>." + } + ] + }, + { + "video_id": "XPDI1", + "question_id": "XPDI1", + "video_file": "Charades/videos/XPDI1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'this person runs into the room with a cup'." + }, + { + "from": "gpt", + "value": "From <2.0> to <9.0>." + } + ] + }, + { + "video_id": "XPDI1", + "question_id": "XPDI1", + "video_file": "Charades/videos/XPDI1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person running through the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <9.0>." + } + ] + }, + { + "video_id": "IV7IZ", + "question_id": "IV7IZ", + "video_file": "Charades/videos/IV7IZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person runs into the basement holding a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <8.9>." + } + ] + }, + { + "video_id": "IV7IZ", + "question_id": "IV7IZ", + "video_file": "Charades/videos/IV7IZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.6> to <12.1>." + } + ] + }, + { + "video_id": "IV7IZ", + "question_id": "IV7IZ", + "video_file": "Charades/videos/IV7IZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs in holding a box with shoes inside'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <8.9>." + } + ] + }, + { + "video_id": "AMX7S", + "question_id": "AMX7S", + "video_file": "Charades/videos/AMX7S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is putting a pillow into a washing machine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.0> to <24.0>." + } + ] + }, + { + "video_id": "Q49E0", + "question_id": "Q49E0", + "video_file": "Charades/videos/Q49E0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'eating a snack as the person talks on the phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.9> to <21.6>." + } + ] + }, + { + "video_id": "Q49E0", + "question_id": "Q49E0", + "video_file": "Charades/videos/Q49E0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person at the same time they are also eating something'." + }, + { + "from": "gpt", + "value": "From <6.9> to <21.6>." + } + ] + }, + { + "video_id": "Q49E0", + "question_id": "Q49E0", + "video_file": "Charades/videos/Q49E0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person look out a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.8> to <33.0>." + } + ] + }, + { + "video_id": "Q49E0", + "question_id": "Q49E0", + "video_file": "Charades/videos/Q49E0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating something'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.9> to <21.6>." + } + ] + }, + { + "video_id": "Q49E0", + "question_id": "Q49E0", + "video_file": "Charades/videos/Q49E0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person look out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.8> to <33.0>." + } + ] + }, + { + "video_id": "HTCKH", + "question_id": "HTCKH", + "video_file": "Charades/videos/HTCKH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put it on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.7> to <18.2>." + } + ] + }, + { + "video_id": "DVHIB", + "question_id": "DVHIB", + "video_file": "Charades/videos/DVHIB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.0> to <27.6>." + } + ] + }, + { + "video_id": "DVHIB", + "question_id": "DVHIB", + "video_file": "Charades/videos/DVHIB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a book from a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <19.2>." + } + ] + }, + { + "video_id": "MUVB8", + "question_id": "MUVB8", + "video_file": "Charades/videos/MUVB8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person gets up turn on a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <22.9>." + } + ] + }, + { + "video_id": "MUVB8", + "question_id": "MUVB8", + "video_file": "Charades/videos/MUVB8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person reaching to turn on the light on the table'." + }, + { + "from": "gpt", + "value": "From <18.0> to <22.9>." + } + ] + }, + { + "video_id": "DN6QU", + "question_id": "DN6QU", + "video_file": "Charades/videos/DN6QU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts it back on the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <9.5>." + } + ] + }, + { + "video_id": "DN6QU", + "question_id": "DN6QU", + "video_file": "Charades/videos/DN6QU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person tries to take a picture' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.3> to <22.7>." + } + ] + }, + { + "video_id": "6B93R", + "question_id": "6B93R", + "video_file": "Charades/videos/6B93R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is eating food from various dishes'." + }, + { + "from": "gpt", + "value": "From <12.8> to <21.9>." + } + ] + }, + { + "video_id": "6B93R", + "question_id": "6B93R", + "video_file": "Charades/videos/6B93R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <38.0> to <46.4>." + } + ] + }, + { + "video_id": "6B93R", + "question_id": "6B93R", + "video_file": "Charades/videos/6B93R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <21.9>." + } + ] + }, + { + "video_id": "6B93R", + "question_id": "6B93R", + "video_file": "Charades/videos/6B93R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes a drink from a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <38.0> to <46.4>." + } + ] + }, + { + "video_id": "A54DW", + "question_id": "A54DW", + "video_file": "Charades/videos/A54DW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sitting on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.5> to <32.0>." + } + ] + }, + { + "video_id": "V58VE", + "question_id": "V58VE", + "video_file": "Charades/videos/V58VE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they are smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.6> to <23.2>." + } + ] + }, + { + "video_id": "ABXKN", + "question_id": "ABXKN", + "video_file": "Charades/videos/ABXKN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they take the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.1> to <16.5>." + } + ] + }, + { + "video_id": "ABXKN", + "question_id": "ABXKN", + "video_file": "Charades/videos/ABXKN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they take a box of tissues from the sink' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.1> to <16.5>." + } + ] + }, + { + "video_id": "5JSFE", + "question_id": "5JSFE", + "video_file": "Charades/videos/5JSFE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person grabs doorknob to open door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.6> to <28.8>." + } + ] + }, + { + "video_id": "VNVRA", + "question_id": "VNVRA", + "video_file": "Charades/videos/VNVRA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a door takes a box out' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <5.8>." + } + ] + }, + { + "video_id": "VNVRA", + "question_id": "VNVRA", + "video_file": "Charades/videos/VNVRA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.1> to <18.6>." + } + ] + }, + { + "video_id": "VNVRA", + "question_id": "VNVRA", + "video_file": "Charades/videos/VNVRA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts it on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.8> to <9.6>." + } + ] + }, + { + "video_id": "8PXPO", + "question_id": "8PXPO", + "video_file": "Charades/videos/8PXPO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <30.2> to <33.0>." + } + ] + }, + { + "video_id": "8PXPO", + "question_id": "8PXPO", + "video_file": "Charades/videos/8PXPO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts some dishes away'." + }, + { + "from": "gpt", + "value": "From <8.3> to <16.5>." + } + ] + }, + { + "video_id": "8PXPO", + "question_id": "8PXPO", + "video_file": "Charades/videos/8PXPO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'that same person starts putting dishes in a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.3> to <16.5>." + } + ] + }, + { + "video_id": "OC6YU", + "question_id": "OC6YU", + "video_file": "Charades/videos/OC6YU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pouring some into a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.8> to <14.3>." + } + ] + }, + { + "video_id": "OC6YU", + "question_id": "OC6YU", + "video_file": "Charades/videos/OC6YU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits in a chair'." + }, + { + "from": "gpt", + "value": "From <9.6> to <18.5>." + } + ] + }, + { + "video_id": "PZD7Z", + "question_id": "PZD7Z", + "video_file": "Charades/videos/PZD7Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting in a chair holding a broom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "PZD7Z", + "question_id": "PZD7Z", + "video_file": "Charades/videos/PZD7Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person the put the broom down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.5> to <17.6>." + } + ] + }, + { + "video_id": "PZD7Z", + "question_id": "PZD7Z", + "video_file": "Charades/videos/PZD7Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts down the broom'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <17.6>." + } + ] + }, + { + "video_id": "KL68V", + "question_id": "KL68V", + "video_file": "Charades/videos/KL68V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person getting dressed puts on their shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.2>." + } + ] + }, + { + "video_id": "5N26T", + "question_id": "5N26T", + "video_file": "Charades/videos/5N26T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens up the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.1> to <37.0>." + } + ] + }, + { + "video_id": "I4IFT", + "question_id": "I4IFT", + "video_file": "Charades/videos/I4IFT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting on their shoes'." + }, + { + "from": "gpt", + "value": "From <2.1> to <15.4>." + } + ] + }, + { + "video_id": "CC6JC", + "question_id": "CC6JC", + "video_file": "Charades/videos/CC6JC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is smiling'." + }, + { + "from": "gpt", + "value": "From <4.4> to <12.1>." + } + ] + }, + { + "video_id": "CC6JC", + "question_id": "CC6JC", + "video_file": "Charades/videos/CC6JC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting on a bed eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "CC6JC", + "question_id": "CC6JC", + "video_file": "Charades/videos/CC6JC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting on a bed eating some food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "15OH5", + "question_id": "15OH5", + "video_file": "Charades/videos/15OH5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person washes the mirror with a towel' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.3> to <36.5>." + } + ] + }, + { + "video_id": "J8SHW", + "question_id": "J8SHW", + "video_file": "Charades/videos/J8SHW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins to make a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.9> to <40.0>." + } + ] + }, + { + "video_id": "F1GWF", + "question_id": "F1GWF", + "video_file": "Charades/videos/F1GWF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "F1GWF", + "question_id": "F1GWF", + "video_file": "Charades/videos/F1GWF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.3> to <10.1>." + } + ] + }, + { + "video_id": "WWPEN", + "question_id": "WWPEN", + "video_file": "Charades/videos/WWPEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is playing a game on a smartphone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.1>." + } + ] + }, + { + "video_id": "WWPEN", + "question_id": "WWPEN", + "video_file": "Charades/videos/WWPEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person playing a game on a cell phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.1>." + } + ] + }, + { + "video_id": "WWPEN", + "question_id": "WWPEN", + "video_file": "Charades/videos/WWPEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.9> to <24.2>." + } + ] + }, + { + "video_id": "WWPEN", + "question_id": "WWPEN", + "video_file": "Charades/videos/WWPEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a man laying in bed playing on a phone'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.1>." + } + ] + }, + { + "video_id": "STG2D", + "question_id": "STG2D", + "video_file": "Charades/videos/STG2D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a cabinet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "6QW3Y", + "question_id": "6QW3Y", + "video_file": "Charades/videos/6QW3Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person throws their bag down on the stairs' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "6QW3Y", + "question_id": "6QW3Y", + "video_file": "Charades/videos/6QW3Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person throws a bag downstairs' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "MMY3N", + "question_id": "MMY3N", + "video_file": "Charades/videos/MMY3N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person puts a box into the garbage'." + }, + { + "from": "gpt", + "value": "From <3.1> to <11.2>." + } + ] + }, + { + "video_id": "N1RJW", + "question_id": "N1RJW", + "video_file": "Charades/videos/N1RJW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door to the room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.9> to <36.2>." + } + ] + }, + { + "video_id": "N1RJW", + "question_id": "N1RJW", + "video_file": "Charades/videos/N1RJW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.9> to <36.2>." + } + ] + }, + { + "video_id": "C9ISQ", + "question_id": "C9ISQ", + "video_file": "Charades/videos/C9ISQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person was closing the laundry detergent cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <13.7>." + } + ] + }, + { + "video_id": "DC3PF", + "question_id": "DC3PF", + "video_file": "Charades/videos/DC3PF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person another is sneezing at the groceries' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.2> to <28.9>." + } + ] + }, + { + "video_id": "DC3PF", + "question_id": "DC3PF", + "video_file": "Charades/videos/DC3PF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.2> to <28.9>." + } + ] + }, + { + "video_id": "ZSHWE", + "question_id": "ZSHWE", + "video_file": "Charades/videos/ZSHWE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person stops at the bottom of the stairs to undress' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <28.2>." + } + ] + }, + { + "video_id": "ZSHWE", + "question_id": "ZSHWE", + "video_file": "Charades/videos/ZSHWE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person he undresses from a towel' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <28.2>." + } + ] + }, + { + "video_id": "5HQJ8", + "question_id": "5HQJ8", + "video_file": "Charades/videos/5HQJ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sneezing as they look up at a light'." + }, + { + "from": "gpt", + "value": "From <19.4> to <31.0>." + } + ] + }, + { + "video_id": "417RA", + "question_id": "417RA", + "video_file": "Charades/videos/417RA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drink from a coffee cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.9> to <31.0>." + } + ] + }, + { + "video_id": "417RA", + "question_id": "417RA", + "video_file": "Charades/videos/417RA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person awakens on a bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "417RA", + "question_id": "417RA", + "video_file": "Charades/videos/417RA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person is awakening' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "JX18W", + "question_id": "JX18W", + "video_file": "Charades/videos/JX18W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a closet door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.4> to <23.2>." + } + ] + }, + { + "video_id": "JX18W", + "question_id": "JX18W", + "video_file": "Charades/videos/JX18W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakens' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "JX18W", + "question_id": "JX18W", + "video_file": "Charades/videos/JX18W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open a closet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.4> to <23.2>." + } + ] + }, + { + "video_id": "JX18W", + "question_id": "JX18W", + "video_file": "Charades/videos/JX18W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding a blanket' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.0> to <16.3>." + } + ] + }, + { + "video_id": "JX18W", + "question_id": "JX18W", + "video_file": "Charades/videos/JX18W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person stands holding a blanket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <16.3>." + } + ] + }, + { + "video_id": "WSLMS", + "question_id": "WSLMS", + "video_file": "Charades/videos/WSLMS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes off shoes in a hallway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.4> to <23.3>." + } + ] + }, + { + "video_id": "WSLMS", + "question_id": "WSLMS", + "video_file": "Charades/videos/WSLMS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes off shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.4> to <23.3>." + } + ] + }, + { + "video_id": "WSLMS", + "question_id": "WSLMS", + "video_file": "Charades/videos/WSLMS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts down the glass'." + }, + { + "from": "gpt", + "value": "From <16.4> to <24.9>." + } + ] + }, + { + "video_id": "2MDHE", + "question_id": "2MDHE", + "video_file": "Charades/videos/2MDHE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking out of a glass'." + }, + { + "from": "gpt", + "value": "From <25.0> to <31.2>." + } + ] + }, + { + "video_id": "2MDHE", + "question_id": "2MDHE", + "video_file": "Charades/videos/2MDHE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person is sitting on the floor eating' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.7> to <25.9>." + } + ] + }, + { + "video_id": "2MDHE", + "question_id": "2MDHE", + "video_file": "Charades/videos/2MDHE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person is eating food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.7> to <25.9>." + } + ] + }, + { + "video_id": "PRV4N", + "question_id": "PRV4N", + "video_file": "Charades/videos/PRV4N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the glass on the desk' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <18.3>." + } + ] + }, + { + "video_id": "PRV4N", + "question_id": "PRV4N", + "video_file": "Charades/videos/PRV4N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a glass off a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <8.7>." + } + ] + }, + { + "video_id": "N7AOK", + "question_id": "N7AOK", + "video_file": "Charades/videos/N7AOK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes a pillow from the chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.0> to <31.0>." + } + ] + }, + { + "video_id": "FE6RN", + "question_id": "FE6RN", + "video_file": "Charades/videos/FE6RN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a cup of coffee'." + }, + { + "from": "gpt", + "value": "From <4.7> to <11.7>." + } + ] + }, + { + "video_id": "KFGXC", + "question_id": "KFGXC", + "video_file": "Charades/videos/KFGXC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding a dish' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.8>." + } + ] + }, + { + "video_id": "KFGXC", + "question_id": "KFGXC", + "video_file": "Charades/videos/KFGXC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person fix their hair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.5> to <30.5>." + } + ] + }, + { + "video_id": "XNT6F", + "question_id": "XNT6F", + "video_file": "Charades/videos/XNT6F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throw it on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.6> to <13.2>." + } + ] + }, + { + "video_id": "C8KER", + "question_id": "C8KER", + "video_file": "Charades/videos/C8KER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.8> to <43.3>." + } + ] + }, + { + "video_id": "C8KER", + "question_id": "C8KER", + "video_file": "Charades/videos/C8KER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a cabinet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "C8KER", + "question_id": "C8KER", + "video_file": "Charades/videos/C8KER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens another cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "JSD0F", + "question_id": "JSD0F", + "video_file": "Charades/videos/JSD0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns off a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.7> to <22.1>." + } + ] + }, + { + "video_id": "JSD0F", + "question_id": "JSD0F", + "video_file": "Charades/videos/JSD0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is holding a bag of groceries'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <14.9>." + } + ] + }, + { + "video_id": "JSD0F", + "question_id": "JSD0F", + "video_file": "Charades/videos/JSD0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they turn off the light'." + }, + { + "from": "gpt", + "value": "From <15.7> to <22.1>." + } + ] + }, + { + "video_id": "JSD0F", + "question_id": "JSD0F", + "video_file": "Charades/videos/JSD0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.1> to <27.6>." + } + ] + }, + { + "video_id": "HHAM7", + "question_id": "HHAM7", + "video_file": "Charades/videos/HHAM7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opens the closet door'." + }, + { + "from": "gpt", + "value": "From <18.4> to <23.0>." + } + ] + }, + { + "video_id": "HHAM7", + "question_id": "HHAM7", + "video_file": "Charades/videos/HHAM7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is opening their closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.4> to <16.3>." + } + ] + }, + { + "video_id": "HHAM7", + "question_id": "HHAM7", + "video_file": "Charades/videos/HHAM7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the closet door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.4> to <23.0>." + } + ] + }, + { + "video_id": "4WMDU", + "question_id": "4WMDU", + "video_file": "Charades/videos/4WMDU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person holding a glass drinking' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.8> to <30.5>." + } + ] + }, + { + "video_id": "4WMDU", + "question_id": "4WMDU", + "video_file": "Charades/videos/4WMDU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking from a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <7.8>." + } + ] + }, + { + "video_id": "4WMDU", + "question_id": "4WMDU", + "video_file": "Charades/videos/4WMDU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'carrying a bag of food.the person opens the bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <12.7>." + } + ] + }, + { + "video_id": "4WMDU", + "question_id": "4WMDU", + "video_file": "Charades/videos/4WMDU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts glass on counter' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.6> to <7.9>." + } + ] + }, + { + "video_id": "4WMDU", + "question_id": "4WMDU", + "video_file": "Charades/videos/4WMDU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinks out of glass again' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <7.8>." + } + ] + }, + { + "video_id": "WO0OM", + "question_id": "WO0OM", + "video_file": "Charades/videos/WO0OM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.3> to <32.0>." + } + ] + }, + { + "video_id": "WO0OM", + "question_id": "WO0OM", + "video_file": "Charades/videos/WO0OM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person start eating the sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.3> to <32.0>." + } + ] + }, + { + "video_id": "WO0OM", + "question_id": "WO0OM", + "video_file": "Charades/videos/WO0OM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <17.4>." + } + ] + }, + { + "video_id": "WO0OM", + "question_id": "WO0OM", + "video_file": "Charades/videos/WO0OM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person laughing at their phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <17.4>." + } + ] + }, + { + "video_id": "47532", + "question_id": "47532", + "video_file": "Charades/videos/47532.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person looking at a picture' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.4> to <23.6>." + } + ] + }, + { + "video_id": "CRJGZ", + "question_id": "CRJGZ", + "video_file": "Charades/videos/CRJGZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.6> to <25.9>." + } + ] + }, + { + "video_id": "CRJGZ", + "question_id": "CRJGZ", + "video_file": "Charades/videos/CRJGZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <13.5>." + } + ] + }, + { + "video_id": "CRJGZ", + "question_id": "CRJGZ", + "video_file": "Charades/videos/CRJGZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits in chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.6> to <25.9>." + } + ] + }, + { + "video_id": "CRJGZ", + "question_id": "CRJGZ", + "video_file": "Charades/videos/CRJGZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts keys on a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.1> to <18.2>." + } + ] + }, + { + "video_id": "CRJGZ", + "question_id": "CRJGZ", + "video_file": "Charades/videos/CRJGZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person closes the door'." + }, + { + "from": "gpt", + "value": "From <7.0> to <13.5>." + } + ] + }, + { + "video_id": "CRJGZ", + "question_id": "CRJGZ", + "video_file": "Charades/videos/CRJGZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens door to entryway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <10.3>." + } + ] + }, + { + "video_id": "SFWBN", + "question_id": "SFWBN", + "video_file": "Charades/videos/SFWBN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they sit down in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <10.1>." + } + ] + }, + { + "video_id": "SFWBN", + "question_id": "SFWBN", + "video_file": "Charades/videos/SFWBN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put their phone back down on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.5> to <10.8>." + } + ] + }, + { + "video_id": "SFWBN", + "question_id": "SFWBN", + "video_file": "Charades/videos/SFWBN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is dressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "SFWBN", + "question_id": "SFWBN", + "video_file": "Charades/videos/SFWBN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes their phone off of a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.5> to <7.4>." + } + ] + }, + { + "video_id": "SFWBN", + "question_id": "SFWBN", + "video_file": "Charades/videos/SFWBN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throwing their clothes onto a blanket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.6> to <5.3>." + } + ] + }, + { + "video_id": "SFWBN", + "question_id": "SFWBN", + "video_file": "Charades/videos/SFWBN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is dressing under the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "6C65Z", + "question_id": "6C65Z", + "video_file": "Charades/videos/6C65Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the refrigerator door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.2> to <8.0>." + } + ] + }, + { + "video_id": "6C65Z", + "question_id": "6C65Z", + "video_file": "Charades/videos/6C65Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is closing the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.2> to <8.0>." + } + ] + }, + { + "video_id": "QQUI7", + "question_id": "QQUI7", + "video_file": "Charades/videos/QQUI7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person #1 is watching a video on their laptop laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.8> to <32.0>." + } + ] + }, + { + "video_id": "QQUI7", + "question_id": "QQUI7", + "video_file": "Charades/videos/QQUI7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person running in circles holding a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "QQUI7", + "question_id": "QQUI7", + "video_file": "Charades/videos/QQUI7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person runs into the kitchen with a box laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "L4DUU", + "question_id": "L4DUU", + "video_file": "Charades/videos/L4DUU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.4> to <24.9>." + } + ] + }, + { + "video_id": "L4DUU", + "question_id": "L4DUU", + "video_file": "Charades/videos/L4DUU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the camera down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <27.5>." + } + ] + }, + { + "video_id": "L4DUU", + "question_id": "L4DUU", + "video_file": "Charades/videos/L4DUU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <18.4> to <24.9>." + } + ] + }, + { + "video_id": "KGMA5", + "question_id": "KGMA5", + "video_file": "Charades/videos/KGMA5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "RF39W", + "question_id": "RF39W", + "video_file": "Charades/videos/RF39W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person was putting food away on shelves' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.8> to <15.0>." + } + ] + }, + { + "video_id": "QO7SM", + "question_id": "QO7SM", + "video_file": "Charades/videos/QO7SM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person talking on the phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <25.9>." + } + ] + }, + { + "video_id": "RXLKF", + "question_id": "RXLKF", + "video_file": "Charades/videos/RXLKF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.4> to <15.2>." + } + ] + }, + { + "video_id": "RXLKF", + "question_id": "RXLKF", + "video_file": "Charades/videos/RXLKF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes book'." + }, + { + "from": "gpt", + "value": "From <23.8> to <28.7>." + } + ] + }, + { + "video_id": "RXLKF", + "question_id": "RXLKF", + "video_file": "Charades/videos/RXLKF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a cup of coffee'." + }, + { + "from": "gpt", + "value": "From <27.0> to <37.3>." + } + ] + }, + { + "video_id": "RXLKF", + "question_id": "RXLKF", + "video_file": "Charades/videos/RXLKF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person she puts down the book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.5> to <30.2>." + } + ] + }, + { + "video_id": "RXLKF", + "question_id": "RXLKF", + "video_file": "Charades/videos/RXLKF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they put the book down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.5> to <30.2>." + } + ] + }, + { + "video_id": "LTQS9", + "question_id": "LTQS9", + "video_file": "Charades/videos/LTQS9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting on a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.1> to <27.0>." + } + ] + }, + { + "video_id": "8N939", + "question_id": "8N939", + "video_file": "Charades/videos/8N939.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person runs down the hallway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.8> to <31.0>." + } + ] + }, + { + "video_id": "8N939", + "question_id": "8N939", + "video_file": "Charades/videos/8N939.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person drinks from the glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <7.6>." + } + ] + }, + { + "video_id": "8N939", + "question_id": "8N939", + "video_file": "Charades/videos/8N939.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes another drink from the glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.4> to <7.6>." + } + ] + }, + { + "video_id": "8N939", + "question_id": "8N939", + "video_file": "Charades/videos/8N939.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person run out of the room'." + }, + { + "from": "gpt", + "value": "From <23.8> to <31.0>." + } + ] + }, + { + "video_id": "2JYRQ", + "question_id": "2JYRQ", + "video_file": "Charades/videos/2JYRQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person washes the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.4> to <23.8>." + } + ] + }, + { + "video_id": "FZ0LU", + "question_id": "FZ0LU", + "video_file": "Charades/videos/FZ0LU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walk out the doorway with it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <10.2>." + } + ] + }, + { + "video_id": "DFG9M", + "question_id": "DFG9M", + "video_file": "Charades/videos/DFG9M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is snuggling under a blanket in the bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.5>." + } + ] + }, + { + "video_id": "5N7DP", + "question_id": "5N7DP", + "video_file": "Charades/videos/5N7DP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is throwing shoes down the stairs' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.6>." + } + ] + }, + { + "video_id": "5N7DP", + "question_id": "5N7DP", + "video_file": "Charades/videos/5N7DP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts talking on the phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.1> to <34.0>." + } + ] + }, + { + "video_id": "5N7DP", + "question_id": "5N7DP", + "video_file": "Charades/videos/5N7DP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws his shoes off their feet'." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.6>." + } + ] + }, + { + "video_id": "MJ1T7", + "question_id": "MJ1T7", + "video_file": "Charades/videos/MJ1T7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person walks to another room holding a bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.5> to <31.0>." + } + ] + }, + { + "video_id": "H5V2Y", + "question_id": "H5V2Y", + "video_file": "Charades/videos/H5V2Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person was eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.3> to <13.7>." + } + ] + }, + { + "video_id": "H5V2Y", + "question_id": "H5V2Y", + "video_file": "Charades/videos/H5V2Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.3> to <13.7>." + } + ] + }, + { + "video_id": "H5V2Y", + "question_id": "H5V2Y", + "video_file": "Charades/videos/H5V2Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking a cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <13.9>." + } + ] + }, + { + "video_id": "H5V2Y", + "question_id": "H5V2Y", + "video_file": "Charades/videos/H5V2Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking from a cup that are on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.5> to <13.9>." + } + ] + }, + { + "video_id": "OCVY4", + "question_id": "OCVY4", + "video_file": "Charades/videos/OCVY4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts pan of food on the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <10.0>." + } + ] + }, + { + "video_id": "OCVY4", + "question_id": "OCVY4", + "video_file": "Charades/videos/OCVY4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put food in it on a stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <10.0>." + } + ] + }, + { + "video_id": "IBX56", + "question_id": "IBX56", + "video_file": "Charades/videos/IBX56.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person eats a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <11.5>." + } + ] + }, + { + "video_id": "IBX56", + "question_id": "IBX56", + "video_file": "Charades/videos/IBX56.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <11.5>." + } + ] + }, + { + "video_id": "IBX56", + "question_id": "IBX56", + "video_file": "Charades/videos/IBX56.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person in the recreation room is eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <11.5>." + } + ] + }, + { + "video_id": "86X97", + "question_id": "86X97", + "video_file": "Charades/videos/86X97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'watching out the window as the person does so' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <11.2>." + } + ] + }, + { + "video_id": "M1GW9", + "question_id": "M1GW9", + "video_file": "Charades/videos/M1GW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throwing the pillow onto the bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.5> to <23.5>." + } + ] + }, + { + "video_id": "M1GW9", + "question_id": "M1GW9", + "video_file": "Charades/videos/M1GW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws the pillow on the bed'." + }, + { + "from": "gpt", + "value": "From <13.5> to <23.5>." + } + ] + }, + { + "video_id": "M1GW9", + "question_id": "M1GW9", + "video_file": "Charades/videos/M1GW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opening the door'." + }, + { + "from": "gpt", + "value": "From <3.8> to <9.9>." + } + ] + }, + { + "video_id": "DFVXX", + "question_id": "DFVXX", + "video_file": "Charades/videos/DFVXX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a towel with them' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.0> to <31.0>." + } + ] + }, + { + "video_id": "P4UXI", + "question_id": "P4UXI", + "video_file": "Charades/videos/P4UXI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pours a cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.0> to <31.9>." + } + ] + }, + { + "video_id": "XWVDK", + "question_id": "XWVDK", + "video_file": "Charades/videos/XWVDK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person throws a bag of groceries onto the sofa' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.2> to <12.4>." + } + ] + }, + { + "video_id": "XWVDK", + "question_id": "XWVDK", + "video_file": "Charades/videos/XWVDK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they throw the bag on a couch'." + }, + { + "from": "gpt", + "value": "From <7.2> to <12.4>." + } + ] + }, + { + "video_id": "SNK5Z", + "question_id": "SNK5Z", + "video_file": "Charades/videos/SNK5Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person fixing a hinge on a door'." + }, + { + "from": "gpt", + "value": "From <2.3> to <8.7>." + } + ] + }, + { + "video_id": "SNK5Z", + "question_id": "SNK5Z", + "video_file": "Charades/videos/SNK5Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eat it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.1> to <20.9>." + } + ] + }, + { + "video_id": "SNK5Z", + "question_id": "SNK5Z", + "video_file": "Charades/videos/SNK5Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is fixing the door to their pantry' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <8.7>." + } + ] + }, + { + "video_id": "SNK5Z", + "question_id": "SNK5Z", + "video_file": "Charades/videos/SNK5Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person was eating food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.1> to <20.9>." + } + ] + }, + { + "video_id": "4Y1AW", + "question_id": "4Y1AW", + "video_file": "Charades/videos/4Y1AW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take a picture of themself with a camera' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.3> to <35.0>." + } + ] + }, + { + "video_id": "N513D", + "question_id": "N513D", + "video_file": "Charades/videos/N513D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person two is holding a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.9> to <21.0>." + } + ] + }, + { + "video_id": "N513D", + "question_id": "N513D", + "video_file": "Charades/videos/N513D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person looks at a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.9> to <18.4>." + } + ] + }, + { + "video_id": "N513D", + "question_id": "N513D", + "video_file": "Charades/videos/N513D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is putting a book in a hall cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <13.9>." + } + ] + }, + { + "video_id": "N513D", + "question_id": "N513D", + "video_file": "Charades/videos/N513D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the book into a storage table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <13.9>." + } + ] + }, + { + "video_id": "IA7X9", + "question_id": "IA7X9", + "video_file": "Charades/videos/IA7X9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts on shoes'." + }, + { + "from": "gpt", + "value": "From <7.7> to <19.5>." + } + ] + }, + { + "video_id": "IA7X9", + "question_id": "IA7X9", + "video_file": "Charades/videos/IA7X9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting shoes on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.7> to <19.5>." + } + ] + }, + { + "video_id": "IA7X9", + "question_id": "IA7X9", + "video_file": "Charades/videos/IA7X9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.3> to <21.1>." + } + ] + }, + { + "video_id": "IA7X9", + "question_id": "IA7X9", + "video_file": "Charades/videos/IA7X9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is opening a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <21.1>." + } + ] + }, + { + "video_id": "APH4V", + "question_id": "APH4V", + "video_file": "Charades/videos/APH4V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they are done they close their closet door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <19.1>." + } + ] + }, + { + "video_id": "APH4V", + "question_id": "APH4V", + "video_file": "Charades/videos/APH4V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open the nearby window'." + }, + { + "from": "gpt", + "value": "From <16.8> to <28.8>." + } + ] + }, + { + "video_id": "APH4V", + "question_id": "APH4V", + "video_file": "Charades/videos/APH4V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is seen putting clothes in their closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.3> to <17.6>." + } + ] + }, + { + "video_id": "APH4V", + "question_id": "APH4V", + "video_file": "Charades/videos/APH4V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person looks out a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <32.0>." + } + ] + }, + { + "video_id": "APH4V", + "question_id": "APH4V", + "video_file": "Charades/videos/APH4V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the clothes in a closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.3> to <17.6>." + } + ] + }, + { + "video_id": "APH4V", + "question_id": "APH4V", + "video_file": "Charades/videos/APH4V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the closet door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <19.1>." + } + ] + }, + { + "video_id": "D1O6N", + "question_id": "D1O6N", + "video_file": "Charades/videos/D1O6N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a laughing person begins talking to a window'." + }, + { + "from": "gpt", + "value": "From <19.2> to <29.5>." + } + ] + }, + { + "video_id": "D1O6N", + "question_id": "D1O6N", + "video_file": "Charades/videos/D1O6N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person another is laughing under a blanket on the sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.2> to <29.5>." + } + ] + }, + { + "video_id": "D1O6N", + "question_id": "D1O6N", + "video_file": "Charades/videos/D1O6N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person walks to the window smiling' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.2> to <29.3>." + } + ] + }, + { + "video_id": "08QQS", + "question_id": "08QQS", + "video_file": "Charades/videos/08QQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.7> to <30.1>." + } + ] + }, + { + "video_id": "08QQS", + "question_id": "08QQS", + "video_file": "Charades/videos/08QQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opening the door'." + }, + { + "from": "gpt", + "value": "From <23.7> to <30.1>." + } + ] + }, + { + "video_id": "X2QWR", + "question_id": "X2QWR", + "video_file": "Charades/videos/X2QWR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they stand up'." + }, + { + "from": "gpt", + "value": "From <23.4> to <30.1>." + } + ] + }, + { + "video_id": "X2QWR", + "question_id": "X2QWR", + "video_file": "Charades/videos/X2QWR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person leaving turn off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.7> to <31.0>." + } + ] + }, + { + "video_id": "MV0TX", + "question_id": "MV0TX", + "video_file": "Charades/videos/MV0TX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts down the sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.2> to <26.1>." + } + ] + }, + { + "video_id": "MV0TX", + "question_id": "MV0TX", + "video_file": "Charades/videos/MV0TX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.8> to <14.1>." + } + ] + }, + { + "video_id": "MV0TX", + "question_id": "MV0TX", + "video_file": "Charades/videos/MV0TX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts to eat a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <14.1>." + } + ] + }, + { + "video_id": "G2VOP", + "question_id": "G2VOP", + "video_file": "Charades/videos/G2VOP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person stands in a doorway sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "G2VOP", + "question_id": "G2VOP", + "video_file": "Charades/videos/G2VOP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put clothes onto a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.8> to <26.0>." + } + ] + }, + { + "video_id": "OJGGZ", + "question_id": "OJGGZ", + "video_file": "Charades/videos/OJGGZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they take off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <38.3> to <45.0>." + } + ] + }, + { + "video_id": "8ENLP", + "question_id": "8ENLP", + "video_file": "Charades/videos/8ENLP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts towel on head'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <12.8>." + } + ] + }, + { + "video_id": "8ENLP", + "question_id": "8ENLP", + "video_file": "Charades/videos/8ENLP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes off shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.1> to <21.4>." + } + ] + }, + { + "video_id": "8ENLP", + "question_id": "8ENLP", + "video_file": "Charades/videos/8ENLP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts a towel over their head'." + }, + { + "from": "gpt", + "value": "From <6.0> to <12.8>." + } + ] + }, + { + "video_id": "E9FY2", + "question_id": "E9FY2", + "video_file": "Charades/videos/E9FY2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the sandwich back' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.4> to <7.2>." + } + ] + }, + { + "video_id": "E9FY2", + "question_id": "E9FY2", + "video_file": "Charades/videos/E9FY2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "E9FY2", + "question_id": "E9FY2", + "video_file": "Charades/videos/E9FY2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'this person eats some bread' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "E9FY2", + "question_id": "E9FY2", + "video_file": "Charades/videos/E9FY2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person finally smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.6> to <30.3>." + } + ] + }, + { + "video_id": "E9FY2", + "question_id": "E9FY2", + "video_file": "Charades/videos/E9FY2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats the sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "E9FY2", + "question_id": "E9FY2", + "video_file": "Charades/videos/E9FY2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes a sandwich from the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "1FPW9", + "question_id": "1FPW9", + "video_file": "Charades/videos/1FPW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person awakens'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <12.4>." + } + ] + }, + { + "video_id": "1FPW9", + "question_id": "1FPW9", + "video_file": "Charades/videos/1FPW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits back on the bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.1> to <31.0>." + } + ] + }, + { + "video_id": "1FPW9", + "question_id": "1FPW9", + "video_file": "Charades/videos/1FPW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits on bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.1> to <31.0>." + } + ] + }, + { + "video_id": "1FPW9", + "question_id": "1FPW9", + "video_file": "Charades/videos/1FPW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinks from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.8> to <25.4>." + } + ] + }, + { + "video_id": "Y7MSY", + "question_id": "Y7MSY", + "video_file": "Charades/videos/Y7MSY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting the pillow on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.0> to <20.2>." + } + ] + }, + { + "video_id": "Y7MSY", + "question_id": "Y7MSY", + "video_file": "Charades/videos/Y7MSY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the pillow on the floor'." + }, + { + "from": "gpt", + "value": "From <15.0> to <20.2>." + } + ] + }, + { + "video_id": "3D0WV", + "question_id": "3D0WV", + "video_file": "Charades/videos/3D0WV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person laughing as he grabs the doorknob' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <30.4>." + } + ] + }, + { + "video_id": "QDOXA", + "question_id": "QDOXA", + "video_file": "Charades/videos/QDOXA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts sneezing'." + }, + { + "from": "gpt", + "value": "From <22.9> to <29.7>." + } + ] + }, + { + "video_id": "QDOXA", + "question_id": "QDOXA", + "video_file": "Charades/videos/QDOXA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is undressing in a garage' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.8> to <15.6>." + } + ] + }, + { + "video_id": "MFOVZ", + "question_id": "MFOVZ", + "video_file": "Charades/videos/MFOVZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person walks through a doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.8> to <32.0>." + } + ] + }, + { + "video_id": "MFOVZ", + "question_id": "MFOVZ", + "video_file": "Charades/videos/MFOVZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "MFOVZ", + "question_id": "MFOVZ", + "video_file": "Charades/videos/MFOVZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person walking out the doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.8> to <32.0>." + } + ] + }, + { + "video_id": "MFOVZ", + "question_id": "MFOVZ", + "video_file": "Charades/videos/MFOVZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person a persno is seen eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "E18UD", + "question_id": "E18UD", + "video_file": "Charades/videos/E18UD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pour beer onto a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.6> to <17.0>." + } + ] + }, + { + "video_id": "WFWTU", + "question_id": "WFWTU", + "video_file": "Charades/videos/WFWTU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs into the garage' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "WFWTU", + "question_id": "WFWTU", + "video_file": "Charades/videos/WFWTU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is running into a garage eating some food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "WFWTU", + "question_id": "WFWTU", + "video_file": "Charades/videos/WFWTU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person close the door behind them' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.3> to <23.0>." + } + ] + }, + { + "video_id": "WFWTU", + "question_id": "WFWTU", + "video_file": "Charades/videos/WFWTU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the door behind them' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.3> to <23.0>." + } + ] + }, + { + "video_id": "WFWTU", + "question_id": "WFWTU", + "video_file": "Charades/videos/WFWTU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person runs out' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "FIAJP", + "question_id": "FIAJP", + "video_file": "Charades/videos/FIAJP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is in a hallway undressing by a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "RPPCV", + "question_id": "RPPCV", + "video_file": "Charades/videos/RPPCV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walk through a doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.3> to <32.0>." + } + ] + }, + { + "video_id": "LV24Z", + "question_id": "LV24Z", + "video_file": "Charades/videos/LV24Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person after sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "LV24Z", + "question_id": "LV24Z", + "video_file": "Charades/videos/LV24Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a laptop from a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <28.0>." + } + ] + }, + { + "video_id": "YFBA1", + "question_id": "YFBA1", + "video_file": "Charades/videos/YFBA1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens the book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <8.6>." + } + ] + }, + { + "video_id": "LDST7", + "question_id": "LDST7", + "video_file": "Charades/videos/LDST7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they take a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <11.1>." + } + ] + }, + { + "video_id": "LDST7", + "question_id": "LDST7", + "video_file": "Charades/videos/LDST7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is opening their medicine cabinet in their bathroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "EGPAO", + "question_id": "EGPAO", + "video_file": "Charades/videos/EGPAO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <11.6>." + } + ] + }, + { + "video_id": "2FLXE", + "question_id": "2FLXE", + "video_file": "Charades/videos/2FLXE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the dryer door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.6> to <13.2>." + } + ] + }, + { + "video_id": "2FLXE", + "question_id": "2FLXE", + "video_file": "Charades/videos/2FLXE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the pillow on the window sill' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.7> to <16.0>." + } + ] + }, + { + "video_id": "ZOK5T", + "question_id": "ZOK5T", + "video_file": "Charades/videos/ZOK5T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs down a hallway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "ZOK5T", + "question_id": "ZOK5T", + "video_file": "Charades/videos/ZOK5T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person runs out of the hallway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "ZOK5T", + "question_id": "ZOK5T", + "video_file": "Charades/videos/ZOK5T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs down the hallway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "OEUGC", + "question_id": "OEUGC", + "video_file": "Charades/videos/OEUGC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take a drink from a medicine bottle' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.8> to <37.0>." + } + ] + }, + { + "video_id": "OEUGC", + "question_id": "OEUGC", + "video_file": "Charades/videos/OEUGC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the laptop down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <18.3>." + } + ] + }, + { + "video_id": "OEUGC", + "question_id": "OEUGC", + "video_file": "Charades/videos/OEUGC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person dresses in warmer clothing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.1> to <30.2>." + } + ] + }, + { + "video_id": "LQGMM", + "question_id": "LQGMM", + "video_file": "Charades/videos/LQGMM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person drinks something from a bottle' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.0> to <13.8>." + } + ] + }, + { + "video_id": "LQGMM", + "question_id": "LQGMM", + "video_file": "Charades/videos/LQGMM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is cooking' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <39.3> to <55.4>." + } + ] + }, + { + "video_id": "LQGMM", + "question_id": "LQGMM", + "video_file": "Charades/videos/LQGMM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts cooking on the stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <39.3> to <55.4>." + } + ] + }, + { + "video_id": "LQGMM", + "question_id": "LQGMM", + "video_file": "Charades/videos/LQGMM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person cooking on a stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <39.3> to <55.4>." + } + ] + }, + { + "video_id": "F7OOB", + "question_id": "F7OOB", + "video_file": "Charades/videos/F7OOB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes the picture down from the wall' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.0> to <33.0>." + } + ] + }, + { + "video_id": "F7OOB", + "question_id": "F7OOB", + "video_file": "Charades/videos/F7OOB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looks a picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.1> to <33.0>." + } + ] + }, + { + "video_id": "F7OOB", + "question_id": "F7OOB", + "video_file": "Charades/videos/F7OOB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a picture from a coatrack' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.0> to <33.0>." + } + ] + }, + { + "video_id": "98GQH", + "question_id": "98GQH", + "video_file": "Charades/videos/98GQH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looking at a picture'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "98GQH", + "question_id": "98GQH", + "video_file": "Charades/videos/98GQH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is holding a picture'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "ZSZS4", + "question_id": "ZSZS4", + "video_file": "Charades/videos/ZSZS4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits on the sofa' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.1> to <26.2>." + } + ] + }, + { + "video_id": "RNCN8", + "question_id": "RNCN8", + "video_file": "Charades/videos/RNCN8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <26.1>." + } + ] + }, + { + "video_id": "RNCN8", + "question_id": "RNCN8", + "video_file": "Charades/videos/RNCN8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes their laptop from a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <19.3>." + } + ] + }, + { + "video_id": "RNCN8", + "question_id": "RNCN8", + "video_file": "Charades/videos/RNCN8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person lay down on the floor with it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.3> to <30.0>." + } + ] + }, + { + "video_id": "RNCN8", + "question_id": "RNCN8", + "video_file": "Charades/videos/RNCN8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they open up the laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.2> to <26.1>." + } + ] + }, + { + "video_id": "JSLW5", + "question_id": "JSLW5", + "video_file": "Charades/videos/JSLW5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person in their closet opens a wardrobe door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "JSLW5", + "question_id": "JSLW5", + "video_file": "Charades/videos/JSLW5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person runs away'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <14.3>." + } + ] + }, + { + "video_id": "8VDZ7", + "question_id": "8VDZ7", + "video_file": "Charades/videos/8VDZ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits on the floor near a drying machine'." + }, + { + "from": "gpt", + "value": "From <15.1> to <28.3>." + } + ] + }, + { + "video_id": "70TVS", + "question_id": "70TVS", + "video_file": "Charades/videos/70TVS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes off his shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.0> to <14.9>." + } + ] + }, + { + "video_id": "70TVS", + "question_id": "70TVS", + "video_file": "Charades/videos/70TVS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing in a home office' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.4> to <26.7>." + } + ] + }, + { + "video_id": "70TVS", + "question_id": "70TVS", + "video_file": "Charades/videos/70TVS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes off shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.0> to <14.9>." + } + ] + }, + { + "video_id": "70TVS", + "question_id": "70TVS", + "video_file": "Charades/videos/70TVS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person watches television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.4> to <39.6>." + } + ] + }, + { + "video_id": "4I00F", + "question_id": "4I00F", + "video_file": "Charades/videos/4I00F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person in a closet takes out some food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.3>." + } + ] + }, + { + "video_id": "8J2K5", + "question_id": "8J2K5", + "video_file": "Charades/videos/8J2K5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.0> to <28.3>." + } + ] + }, + { + "video_id": "8J2K5", + "question_id": "8J2K5", + "video_file": "Charades/videos/8J2K5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closing doors as they vacuum' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <30.2> to <36.0>." + } + ] + }, + { + "video_id": "I6AZD", + "question_id": "I6AZD", + "video_file": "Charades/videos/I6AZD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.0> to <19.4>." + } + ] + }, + { + "video_id": "OO6BH", + "question_id": "OO6BH", + "video_file": "Charades/videos/OO6BH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put it on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <18.8>." + } + ] + }, + { + "video_id": "OO6BH", + "question_id": "OO6BH", + "video_file": "Charades/videos/OO6BH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they open the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <13.7>." + } + ] + }, + { + "video_id": "6B328", + "question_id": "6B328", + "video_file": "Charades/videos/6B328.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes some medicine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <29.3>." + } + ] + }, + { + "video_id": "6B328", + "question_id": "6B328", + "video_file": "Charades/videos/6B328.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes some medicine from the desk' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.7> to <29.3>." + } + ] + }, + { + "video_id": "LBT7P", + "question_id": "LBT7P", + "video_file": "Charades/videos/LBT7P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is throwing a pillow at a picture'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <12.5>." + } + ] + }, + { + "video_id": "LBT7P", + "question_id": "LBT7P", + "video_file": "Charades/videos/LBT7P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person throws a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <12.5>." + } + ] + }, + { + "video_id": "A2EY4", + "question_id": "A2EY4", + "video_file": "Charades/videos/A2EY4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person snuggles in the blanket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.8> to <24.4>." + } + ] + }, + { + "video_id": "A2EY4", + "question_id": "A2EY4", + "video_file": "Charades/videos/A2EY4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opens a refrigerator door'." + }, + { + "from": "gpt", + "value": "From <1.5> to <6.9>." + } + ] + }, + { + "video_id": "C9FMC", + "question_id": "C9FMC", + "video_file": "Charades/videos/C9FMC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pouring soda into their cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.8> to <37.3>." + } + ] + }, + { + "video_id": "C9FMC", + "question_id": "C9FMC", + "video_file": "Charades/videos/C9FMC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is closing the kitchen door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.4> to <29.4>." + } + ] + }, + { + "video_id": "C9FMC", + "question_id": "C9FMC", + "video_file": "Charades/videos/C9FMC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <6.7>." + } + ] + }, + { + "video_id": "C9FMC", + "question_id": "C9FMC", + "video_file": "Charades/videos/C9FMC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <21.4> to <29.4>." + } + ] + }, + { + "video_id": "C9FMC", + "question_id": "C9FMC", + "video_file": "Charades/videos/C9FMC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is putting groceries into a refrigerator'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.4>." + } + ] + }, + { + "video_id": "C9FMC", + "question_id": "C9FMC", + "video_file": "Charades/videos/C9FMC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts away groceries'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.4>." + } + ] + }, + { + "video_id": "CG9ZL", + "question_id": "CG9ZL", + "video_file": "Charades/videos/CG9ZL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks through a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <10.2>." + } + ] + }, + { + "video_id": "IUW9N", + "question_id": "IUW9N", + "video_file": "Charades/videos/IUW9N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door behind them' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.1> to <5.9>." + } + ] + }, + { + "video_id": "IUW9N", + "question_id": "IUW9N", + "video_file": "Charades/videos/IUW9N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the door behind them' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.1> to <5.9>." + } + ] + }, + { + "video_id": "X28IF", + "question_id": "X28IF", + "video_file": "Charades/videos/X28IF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours a cup of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.4> to <14.6>." + } + ] + }, + { + "video_id": "X28IF", + "question_id": "X28IF", + "video_file": "Charades/videos/X28IF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person awakens in a chair in their home office' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "X28IF", + "question_id": "X28IF", + "video_file": "Charades/videos/X28IF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they pour a cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.4> to <14.6>." + } + ] + }, + { + "video_id": "8TRIW", + "question_id": "8TRIW", + "video_file": "Charades/videos/8TRIW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes some medicine with a glass of water'." + }, + { + "from": "gpt", + "value": "From <4.1> to <14.1>." + } + ] + }, + { + "video_id": "XECIL", + "question_id": "XECIL", + "video_file": "Charades/videos/XECIL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.0> to <8.8>." + } + ] + }, + { + "video_id": "XECIL", + "question_id": "XECIL", + "video_file": "Charades/videos/XECIL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts food away' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <10.1>." + } + ] + }, + { + "video_id": "XECIL", + "question_id": "XECIL", + "video_file": "Charades/videos/XECIL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <14.5>." + } + ] + }, + { + "video_id": "XECIL", + "question_id": "XECIL", + "video_file": "Charades/videos/XECIL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a bag of chips from the cabinet'." + }, + { + "from": "gpt", + "value": "From <12.0> to <16.7>." + } + ] + }, + { + "video_id": "XECIL", + "question_id": "XECIL", + "video_file": "Charades/videos/XECIL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is eating a sandwich in the pantry' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.0> to <8.8>." + } + ] + }, + { + "video_id": "PCNB7", + "question_id": "PCNB7", + "video_file": "Charades/videos/PCNB7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is fixing their hair under the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.6>." + } + ] + }, + { + "video_id": "PCNB7", + "question_id": "PCNB7", + "video_file": "Charades/videos/PCNB7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person fixing hair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.6>." + } + ] + }, + { + "video_id": "BU2R4", + "question_id": "BU2R4", + "video_file": "Charades/videos/BU2R4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts it on the table'." + }, + { + "from": "gpt", + "value": "From <20.3> to <27.3>." + } + ] + }, + { + "video_id": "96GIQ", + "question_id": "96GIQ", + "video_file": "Charades/videos/96GIQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <18.1>." + } + ] + }, + { + "video_id": "96GIQ", + "question_id": "96GIQ", + "video_file": "Charades/videos/96GIQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person throws a broom on the floor'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "XTACZ", + "question_id": "XTACZ", + "video_file": "Charades/videos/XTACZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turn off the light'." + }, + { + "from": "gpt", + "value": "From <18.2> to <24.0>." + } + ] + }, + { + "video_id": "XTACZ", + "question_id": "XTACZ", + "video_file": "Charades/videos/XTACZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts undressing'." + }, + { + "from": "gpt", + "value": "From <5.0> to <19.5>." + } + ] + }, + { + "video_id": "XTACZ", + "question_id": "XTACZ", + "video_file": "Charades/videos/XTACZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns off a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <24.0>." + } + ] + }, + { + "video_id": "XTACZ", + "question_id": "XTACZ", + "video_file": "Charades/videos/XTACZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes off shoes'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.2>." + } + ] + }, + { + "video_id": "S0Z00", + "question_id": "S0Z00", + "video_file": "Charades/videos/S0Z00.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person laughs' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.4> to <31.0>." + } + ] + }, + { + "video_id": "S0Z00", + "question_id": "S0Z00", + "video_file": "Charades/videos/S0Z00.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person puts down a phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.4> to <9.1>." + } + ] + }, + { + "video_id": "S0Z00", + "question_id": "S0Z00", + "video_file": "Charades/videos/S0Z00.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.4> to <31.0>." + } + ] + }, + { + "video_id": "CNVZX", + "question_id": "CNVZX", + "video_file": "Charades/videos/CNVZX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'this person awakens'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.9> to <12.8>." + } + ] + }, + { + "video_id": "CNVZX", + "question_id": "CNVZX", + "video_file": "Charades/videos/CNVZX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns on the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "CNVZX", + "question_id": "CNVZX", + "video_file": "Charades/videos/CNVZX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "MYB7P", + "question_id": "MYB7P", + "video_file": "Charades/videos/MYB7P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.6> to <15.0>." + } + ] + }, + { + "video_id": "MYB7P", + "question_id": "MYB7P", + "video_file": "Charades/videos/MYB7P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes a picture with their phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <10.0>." + } + ] + }, + { + "video_id": "HQZ29", + "question_id": "HQZ29", + "video_file": "Charades/videos/HQZ29.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts a dish into the sink' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.1> to <26.3>." + } + ] + }, + { + "video_id": "3NOML", + "question_id": "3NOML", + "video_file": "Charades/videos/3NOML.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.8> to <31.0>." + } + ] + }, + { + "video_id": "L0YQM", + "question_id": "L0YQM", + "video_file": "Charades/videos/L0YQM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person laughing at the book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.9> to <15.6>." + } + ] + }, + { + "video_id": "L0YQM", + "question_id": "L0YQM", + "video_file": "Charades/videos/L0YQM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take the book out again'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.8>." + } + ] + }, + { + "video_id": "L0YQM", + "question_id": "L0YQM", + "video_file": "Charades/videos/L0YQM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts a book in a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.6> to <20.8>." + } + ] + }, + { + "video_id": "L0YQM", + "question_id": "L0YQM", + "video_file": "Charades/videos/L0YQM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person start laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.9> to <15.6>." + } + ] + }, + { + "video_id": "9N76L", + "question_id": "9N76L", + "video_file": "Charades/videos/9N76L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person eats food from a container' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.7> to <36.8>." + } + ] + }, + { + "video_id": "9N76L", + "question_id": "9N76L", + "video_file": "Charades/videos/9N76L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eat it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.7> to <36.8>." + } + ] + }, + { + "video_id": "H5TGJ", + "question_id": "H5TGJ", + "video_file": "Charades/videos/H5TGJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakens in a bedroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.5> to <11.7>." + } + ] + }, + { + "video_id": "H5TGJ", + "question_id": "H5TGJ", + "video_file": "Charades/videos/H5TGJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person laying on a bed awakens' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.5> to <11.7>." + } + ] + }, + { + "video_id": "IS3OX", + "question_id": "IS3OX", + "video_file": "Charades/videos/IS3OX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits on the floor in the entryway'." + }, + { + "from": "gpt", + "value": "From <29.4> to <37.0>." + } + ] + }, + { + "video_id": "ZI2XH", + "question_id": "ZI2XH", + "video_file": "Charades/videos/ZI2XH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is putting a book on the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <17.6>." + } + ] + }, + { + "video_id": "F31X6", + "question_id": "F31X6", + "video_file": "Charades/videos/F31X6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person eats something off a plate'." + }, + { + "from": "gpt", + "value": "From <20.1> to <25.7>." + } + ] + }, + { + "video_id": "KNX2Y", + "question_id": "KNX2Y", + "video_file": "Charades/videos/KNX2Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding the glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.0> to <21.0>." + } + ] + }, + { + "video_id": "KNX2Y", + "question_id": "KNX2Y", + "video_file": "Charades/videos/KNX2Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they begin laughing holding food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <21.0>." + } + ] + }, + { + "video_id": "KNX2Y", + "question_id": "KNX2Y", + "video_file": "Charades/videos/KNX2Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person holds a camera'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.4>." + } + ] + }, + { + "video_id": "KNX2Y", + "question_id": "KNX2Y", + "video_file": "Charades/videos/KNX2Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they laughed'." + }, + { + "from": "gpt", + "value": "From <13.0> to <21.0>." + } + ] + }, + { + "video_id": "KNX2Y", + "question_id": "KNX2Y", + "video_file": "Charades/videos/KNX2Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person putting it on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <16.8>." + } + ] + }, + { + "video_id": "VEV4O", + "question_id": "VEV4O", + "video_file": "Charades/videos/VEV4O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person throws a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <14.4>." + } + ] + }, + { + "video_id": "VEV4O", + "question_id": "VEV4O", + "video_file": "Charades/videos/VEV4O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws a pillow onto a couch but misses' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <14.4>." + } + ] + }, + { + "video_id": "QC4R2", + "question_id": "QC4R2", + "video_file": "Charades/videos/QC4R2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the picture on the floor'." + }, + { + "from": "gpt", + "value": "From <12.1> to <18.6>." + } + ] + }, + { + "video_id": "QC4R2", + "question_id": "QC4R2", + "video_file": "Charades/videos/QC4R2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begin laughing at it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.6> to <29.8>." + } + ] + }, + { + "video_id": "QC4R2", + "question_id": "QC4R2", + "video_file": "Charades/videos/QC4R2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts laughing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.6> to <29.8>." + } + ] + }, + { + "video_id": "D2GQH", + "question_id": "D2GQH", + "video_file": "Charades/videos/D2GQH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a sneezing person is sitting in their garage' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "D2GQH", + "question_id": "D2GQH", + "video_file": "Charades/videos/D2GQH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person eats a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <21.5>." + } + ] + }, + { + "video_id": "N7130", + "question_id": "N7130", + "video_file": "Charades/videos/N7130.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put their phone in their pocket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <13.5>." + } + ] + }, + { + "video_id": "N7130", + "question_id": "N7130", + "video_file": "Charades/videos/N7130.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "N7130", + "question_id": "N7130", + "video_file": "Charades/videos/N7130.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts a phone in their pocket'." + }, + { + "from": "gpt", + "value": "From <5.6> to <13.5>." + } + ] + }, + { + "video_id": "N7130", + "question_id": "N7130", + "video_file": "Charades/videos/N7130.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person walks through the doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.6> to <7.7>." + } + ] + }, + { + "video_id": "ZFZEN", + "question_id": "ZFZEN", + "video_file": "Charades/videos/ZFZEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins dressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.9> to <40.5>." + } + ] + }, + { + "video_id": "ZFZEN", + "question_id": "ZFZEN", + "video_file": "Charades/videos/ZFZEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is awakening from their bed'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "06CKN", + "question_id": "06CKN", + "video_file": "Charades/videos/06CKN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <17.6> to <26.0>." + } + ] + }, + { + "video_id": "06CKN", + "question_id": "06CKN", + "video_file": "Charades/videos/06CKN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a cabinet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.2> to <14.6>." + } + ] + }, + { + "video_id": "06CKN", + "question_id": "06CKN", + "video_file": "Charades/videos/06CKN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.2> to <14.6>." + } + ] + }, + { + "video_id": "08F85", + "question_id": "08F85", + "video_file": "Charades/videos/08F85.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <15.1>." + } + ] + }, + { + "video_id": "08F85", + "question_id": "08F85", + "video_file": "Charades/videos/08F85.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes some medicine down from a shelf'." + }, + { + "from": "gpt", + "value": "From <14.4> to <23.2>." + } + ] + }, + { + "video_id": "08F85", + "question_id": "08F85", + "video_file": "Charades/videos/08F85.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <15.1>." + } + ] + }, + { + "video_id": "3K6HO", + "question_id": "3K6HO", + "video_file": "Charades/videos/3K6HO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person close the door when exiting the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.6> to <26.0>." + } + ] + }, + { + "video_id": "3K6HO", + "question_id": "3K6HO", + "video_file": "Charades/videos/3K6HO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person throws a blanket on a bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <7.5>." + } + ] + }, + { + "video_id": "3K6HO", + "question_id": "3K6HO", + "video_file": "Charades/videos/3K6HO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes a door'." + }, + { + "from": "gpt", + "value": "From <19.6> to <26.0>." + } + ] + }, + { + "video_id": "X226B", + "question_id": "X226B", + "video_file": "Charades/videos/X226B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is cooking in their kitchen' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "X226B", + "question_id": "X226B", + "video_file": "Charades/videos/X226B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put a pillow behind their back' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.0> to <27.2>." + } + ] + }, + { + "video_id": "X226B", + "question_id": "X226B", + "video_file": "Charades/videos/X226B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is cooking on the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "X226B", + "question_id": "X226B", + "video_file": "Charades/videos/X226B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sit in a chair'." + }, + { + "from": "gpt", + "value": "From <20.7> to <29.6>." + } + ] + }, + { + "video_id": "X226B", + "question_id": "X226B", + "video_file": "Charades/videos/X226B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is cooking something on the stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "X226B", + "question_id": "X226B", + "video_file": "Charades/videos/X226B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits down in a chair adjusting the pillows' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.7> to <29.6>." + } + ] + }, + { + "video_id": "3XKBC", + "question_id": "3XKBC", + "video_file": "Charades/videos/3XKBC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.8> to <28.1>." + } + ] + }, + { + "video_id": "IME5D", + "question_id": "IME5D", + "video_file": "Charades/videos/IME5D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts the food on the shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.7> to <25.4>." + } + ] + }, + { + "video_id": "STAZI", + "question_id": "STAZI", + "video_file": "Charades/videos/STAZI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they start taking out pans to cook it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.8> to <37.0>." + } + ] + }, + { + "video_id": "XXIT4", + "question_id": "XXIT4", + "video_file": "Charades/videos/XXIT4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person still eating' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.9> to <19.3>." + } + ] + }, + { + "video_id": "XXIT4", + "question_id": "XXIT4", + "video_file": "Charades/videos/XXIT4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.9> to <18.9>." + } + ] + }, + { + "video_id": "XXIT4", + "question_id": "XXIT4", + "video_file": "Charades/videos/XXIT4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting in a chair at a desk' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "XXIT4", + "question_id": "XXIT4", + "video_file": "Charades/videos/XXIT4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "5SAAB", + "question_id": "5SAAB", + "video_file": "Charades/videos/5SAAB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take a towel open the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.5> to <22.8>." + } + ] + }, + { + "video_id": "5SAAB", + "question_id": "5SAAB", + "video_file": "Charades/videos/5SAAB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sitting down on the sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.2> to <32.0>." + } + ] + }, + { + "video_id": "5SAAB", + "question_id": "5SAAB", + "video_file": "Charades/videos/5SAAB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts a foot on the coffee table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.8> to <11.5>." + } + ] + }, + { + "video_id": "5SAAB", + "question_id": "5SAAB", + "video_file": "Charades/videos/5SAAB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits on the sofa' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.2> to <32.0>." + } + ] + }, + { + "video_id": "5SAAB", + "question_id": "5SAAB", + "video_file": "Charades/videos/5SAAB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats some' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.2> to <32.0>." + } + ] + }, + { + "video_id": "5SAAB", + "question_id": "5SAAB", + "video_file": "Charades/videos/5SAAB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.2> to <32.0>." + } + ] + }, + { + "video_id": "RJ8WY", + "question_id": "RJ8WY", + "video_file": "Charades/videos/RJ8WY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a cup off of the washer/dryer' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.3> to <17.2>." + } + ] + }, + { + "video_id": "RJ8WY", + "question_id": "RJ8WY", + "video_file": "Charades/videos/RJ8WY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns on the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.3> to <7.1>." + } + ] + }, + { + "video_id": "RJ8WY", + "question_id": "RJ8WY", + "video_file": "Charades/videos/RJ8WY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns on a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <7.1>." + } + ] + }, + { + "video_id": "YXP9P", + "question_id": "YXP9P", + "video_file": "Charades/videos/YXP9P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person pretends to start sneezing'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "Z9W47", + "question_id": "Z9W47", + "video_file": "Charades/videos/Z9W47.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person throws some items in a box the smiles' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.9> to <15.7>." + } + ] + }, + { + "video_id": "Z9W47", + "question_id": "Z9W47", + "video_file": "Charades/videos/Z9W47.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "Z9W47", + "question_id": "Z9W47", + "video_file": "Charades/videos/Z9W47.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person throws some objects in the box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.9> to <15.7>." + } + ] + }, + { + "video_id": "Z9W47", + "question_id": "Z9W47", + "video_file": "Charades/videos/Z9W47.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a closet door with a light on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "Z9W47", + "question_id": "Z9W47", + "video_file": "Charades/videos/Z9W47.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they throw something into a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.9> to <15.7>." + } + ] + }, + { + "video_id": "F44A2", + "question_id": "F44A2", + "video_file": "Charades/videos/F44A2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a picture of themself'." + }, + { + "from": "gpt", + "value": "From <25.2> to <30.9>." + } + ] + }, + { + "video_id": "F44A2", + "question_id": "F44A2", + "video_file": "Charades/videos/F44A2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.2> to <30.9>." + } + ] + }, + { + "video_id": "I3Q65", + "question_id": "I3Q65", + "video_file": "Charades/videos/I3Q65.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sit on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <11.0>." + } + ] + }, + { + "video_id": "I3Q65", + "question_id": "I3Q65", + "video_file": "Charades/videos/I3Q65.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begin to put boxes on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.1> to <17.3>." + } + ] + }, + { + "video_id": "I3Q65", + "question_id": "I3Q65", + "video_file": "Charades/videos/I3Q65.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts putting boxes on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.1> to <17.3>." + } + ] + }, + { + "video_id": "GXIWH", + "question_id": "GXIWH", + "video_file": "Charades/videos/GXIWH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is smiling at the bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "JZDL6", + "question_id": "JZDL6", + "video_file": "Charades/videos/JZDL6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sit on the bed'." + }, + { + "from": "gpt", + "value": "From <9.0> to <14.0>." + } + ] + }, + { + "video_id": "YG2EP", + "question_id": "YG2EP", + "video_file": "Charades/videos/YG2EP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person walks threw the bathroom doorway holding a towel' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.3>." + } + ] + }, + { + "video_id": "8BNBV", + "question_id": "8BNBV", + "video_file": "Charades/videos/8BNBV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person a is walking through an entryway smiling' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.3> to <14.4>." + } + ] + }, + { + "video_id": "8BNBV", + "question_id": "8BNBV", + "video_file": "Charades/videos/8BNBV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person a walks through a doorway'." + }, + { + "from": "gpt", + "value": "From <14.9> to <20.2>." + } + ] + }, + { + "video_id": "VPI6D", + "question_id": "VPI6D", + "video_file": "Charades/videos/VPI6D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes pillows off of the bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.0> to <12.8>." + } + ] + }, + { + "video_id": "VPI6D", + "question_id": "VPI6D", + "video_file": "Charades/videos/VPI6D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a door by turning a doorknob'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "VPI6D", + "question_id": "VPI6D", + "video_file": "Charades/videos/VPI6D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the door to their pantry' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "6C4DV", + "question_id": "6C4DV", + "video_file": "Charades/videos/6C4DV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes a picture' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <15.6>." + } + ] + }, + { + "video_id": "6C4DV", + "question_id": "6C4DV", + "video_file": "Charades/videos/6C4DV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs to the closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "6C4DV", + "question_id": "6C4DV", + "video_file": "Charades/videos/6C4DV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs into a closet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "6C4DV", + "question_id": "6C4DV", + "video_file": "Charades/videos/6C4DV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a picture of it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <15.6>." + } + ] + }, + { + "video_id": "6C4DV", + "question_id": "6C4DV", + "video_file": "Charades/videos/6C4DV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'this person runs into the closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "6C4DV", + "question_id": "6C4DV", + "video_file": "Charades/videos/6C4DV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts a glass on the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <7.3>." + } + ] + }, + { + "video_id": "XZ2QQ", + "question_id": "XZ2QQ", + "video_file": "Charades/videos/XZ2QQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts down the cell phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.9> to <13.6>." + } + ] + }, + { + "video_id": "XZ2QQ", + "question_id": "XZ2QQ", + "video_file": "Charades/videos/XZ2QQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the phone down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.9> to <13.6>." + } + ] + }, + { + "video_id": "78N0Z", + "question_id": "78N0Z", + "video_file": "Charades/videos/78N0Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins eating'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.1> to <32.0>." + } + ] + }, + { + "video_id": "78N0Z", + "question_id": "78N0Z", + "video_file": "Charades/videos/78N0Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts eating bag of chips'." + }, + { + "from": "gpt", + "value": "From <21.1> to <32.0>." + } + ] + }, + { + "video_id": "J7D0D", + "question_id": "J7D0D", + "video_file": "Charades/videos/J7D0D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "J7D0D", + "question_id": "J7D0D", + "video_file": "Charades/videos/J7D0D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person opened a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "KP88I", + "question_id": "KP88I", + "video_file": "Charades/videos/KP88I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a bite of a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.5> to <24.9>." + } + ] + }, + { + "video_id": "KP88I", + "question_id": "KP88I", + "video_file": "Charades/videos/KP88I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats some food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <28.3>." + } + ] + }, + { + "video_id": "KP88I", + "question_id": "KP88I", + "video_file": "Charades/videos/KP88I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they start eating some food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <28.3>." + } + ] + }, + { + "video_id": "G1KTZ", + "question_id": "G1KTZ", + "video_file": "Charades/videos/G1KTZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looked out of a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.9> to <31.6>." + } + ] + }, + { + "video_id": "G1KTZ", + "question_id": "G1KTZ", + "video_file": "Charades/videos/G1KTZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looks out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.9> to <31.6>." + } + ] + }, + { + "video_id": "WR1QE", + "question_id": "WR1QE", + "video_file": "Charades/videos/WR1QE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <12.6>." + } + ] + }, + { + "video_id": "NGZQ9", + "question_id": "NGZQ9", + "video_file": "Charades/videos/NGZQ9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person drinks from a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <6.2>." + } + ] + }, + { + "video_id": "NGZQ9", + "question_id": "NGZQ9", + "video_file": "Charades/videos/NGZQ9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking a glass of water'." + }, + { + "from": "gpt", + "value": "From <0.3> to <6.2>." + } + ] + }, + { + "video_id": "NGZQ9", + "question_id": "NGZQ9", + "video_file": "Charades/videos/NGZQ9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is drinking a cup of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <6.2>." + } + ] + }, + { + "video_id": "CV20O", + "question_id": "CV20O", + "video_file": "Charades/videos/CV20O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person runs into the bedroom eating something'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.1>." + } + ] + }, + { + "video_id": "CV20O", + "question_id": "CV20O", + "video_file": "Charades/videos/CV20O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is eating something' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.1>." + } + ] + }, + { + "video_id": "RRHEV", + "question_id": "RRHEV", + "video_file": "Charades/videos/RRHEV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person run out' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.0> to <31.8>." + } + ] + }, + { + "video_id": "RRHEV", + "question_id": "RRHEV", + "video_file": "Charades/videos/RRHEV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs out the kitchen' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.0> to <31.8>." + } + ] + }, + { + "video_id": "2YO6G", + "question_id": "2YO6G", + "video_file": "Charades/videos/2YO6G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is opening the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "2YO6G", + "question_id": "2YO6G", + "video_file": "Charades/videos/2YO6G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they close the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.3> to <11.0>." + } + ] + }, + { + "video_id": "2YO6G", + "question_id": "2YO6G", + "video_file": "Charades/videos/2YO6G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "VX1GX", + "question_id": "VX1GX", + "video_file": "Charades/videos/VX1GX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts them on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <17.1>." + } + ] + }, + { + "video_id": "VX1GX", + "question_id": "VX1GX", + "video_file": "Charades/videos/VX1GX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.3> to <37.9>." + } + ] + }, + { + "video_id": "17DOT", + "question_id": "17DOT", + "video_file": "Charades/videos/17DOT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walk through a doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.1> to <23.0>." + } + ] + }, + { + "video_id": "VRY89", + "question_id": "VRY89", + "video_file": "Charades/videos/VRY89.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.3> to <32.0>." + } + ] + }, + { + "video_id": "VRY89", + "question_id": "VRY89", + "video_file": "Charades/videos/VRY89.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they close the door behind them'." + }, + { + "from": "gpt", + "value": "From <26.3> to <32.0>." + } + ] + }, + { + "video_id": "VRY89", + "question_id": "VRY89", + "video_file": "Charades/videos/VRY89.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.3> to <32.0>." + } + ] + }, + { + "video_id": "E0U3R", + "question_id": "E0U3R", + "video_file": "Charades/videos/E0U3R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <22.5>." + } + ] + }, + { + "video_id": "E0U3R", + "question_id": "E0U3R", + "video_file": "Charades/videos/E0U3R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sitting in a chair next to the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <22.5>." + } + ] + }, + { + "video_id": "E0U3R", + "question_id": "E0U3R", + "video_file": "Charades/videos/E0U3R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the clothes on the bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <15.2>." + } + ] + }, + { + "video_id": "E0U3R", + "question_id": "E0U3R", + "video_file": "Charades/videos/E0U3R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person tidying up the bedroom by folding clothes'." + }, + { + "from": "gpt", + "value": "From <5.6> to <13.1>." + } + ] + }, + { + "video_id": "T3L2A", + "question_id": "T3L2A", + "video_file": "Charades/videos/T3L2A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is standing in the hallway holding a camera' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.5> to <10.8>." + } + ] + }, + { + "video_id": "EX1PY", + "question_id": "EX1PY", + "video_file": "Charades/videos/EX1PY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes a box off the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <9.8>." + } + ] + }, + { + "video_id": "7DGWM", + "question_id": "7DGWM", + "video_file": "Charades/videos/7DGWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person running up to a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.2> to <8.2>." + } + ] + }, + { + "video_id": "7DGWM", + "question_id": "7DGWM", + "video_file": "Charades/videos/7DGWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holds two pillows' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.1> to <30.3>." + } + ] + }, + { + "video_id": "7DGWM", + "question_id": "7DGWM", + "video_file": "Charades/videos/7DGWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits on his bed on a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <11.2>." + } + ] + }, + { + "video_id": "7DGWM", + "question_id": "7DGWM", + "video_file": "Charades/videos/7DGWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person holds the pillows to their chest' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.1> to <30.3>." + } + ] + }, + { + "video_id": "7DGWM", + "question_id": "7DGWM", + "video_file": "Charades/videos/7DGWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person sits down on a bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <11.2>." + } + ] + }, + { + "video_id": "6PL9Z", + "question_id": "6PL9Z", + "video_file": "Charades/videos/6PL9Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.8> to <13.4>." + } + ] + }, + { + "video_id": "6PL9Z", + "question_id": "6PL9Z", + "video_file": "Charades/videos/6PL9Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person is sitting on a sofa'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "K2L7F", + "question_id": "K2L7F", + "video_file": "Charades/videos/K2L7F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <14.0>." + } + ] + }, + { + "video_id": "RTEWJ", + "question_id": "RTEWJ", + "video_file": "Charades/videos/RTEWJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person runs to the closet'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "RTEWJ", + "question_id": "RTEWJ", + "video_file": "Charades/videos/RTEWJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs into the closet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "RTEWJ", + "question_id": "RTEWJ", + "video_file": "Charades/videos/RTEWJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes a towel' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <9.8>." + } + ] + }, + { + "video_id": "YSZVZ", + "question_id": "YSZVZ", + "video_file": "Charades/videos/YSZVZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take a drink from a cup'." + }, + { + "from": "gpt", + "value": "From <0.5> to <7.9>." + } + ] + }, + { + "video_id": "YSZVZ", + "question_id": "YSZVZ", + "video_file": "Charades/videos/YSZVZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person starts sneezing into a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.7> to <18.2>." + } + ] + }, + { + "video_id": "YSZVZ", + "question_id": "YSZVZ", + "video_file": "Charades/videos/YSZVZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.4> to <17.0>." + } + ] + }, + { + "video_id": "SQBEE", + "question_id": "SQBEE", + "video_file": "Charades/videos/SQBEE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is eating with a blanket draped over them' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.2> to <6.5>." + } + ] + }, + { + "video_id": "SQBEE", + "question_id": "SQBEE", + "video_file": "Charades/videos/SQBEE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person is eating food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.2> to <6.5>." + } + ] + }, + { + "video_id": "SG8ZR", + "question_id": "SG8ZR", + "video_file": "Charades/videos/SG8ZR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person snuggles with a blanket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <19.7>." + } + ] + }, + { + "video_id": "SG8ZR", + "question_id": "SG8ZR", + "video_file": "Charades/videos/SG8ZR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the beer cap into a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <33.9> to <41.4>." + } + ] + }, + { + "video_id": "55MRE", + "question_id": "55MRE", + "video_file": "Charades/videos/55MRE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throwing them on to floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.2>." + } + ] + }, + { + "video_id": "55MRE", + "question_id": "55MRE", + "video_file": "Charades/videos/55MRE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throwing clothes on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.2>." + } + ] + }, + { + "video_id": "55MRE", + "question_id": "55MRE", + "video_file": "Charades/videos/55MRE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person eats' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "55MRE", + "question_id": "55MRE", + "video_file": "Charades/videos/55MRE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating food'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "YNJ86", + "question_id": "YNJ86", + "video_file": "Charades/videos/YNJ86.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a drink from a cup of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.5> to <30.6>." + } + ] + }, + { + "video_id": "YNJ86", + "question_id": "YNJ86", + "video_file": "Charades/videos/YNJ86.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person closes the book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <19.1>." + } + ] + }, + { + "video_id": "51801", + "question_id": "51801", + "video_file": "Charades/videos/51801.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they drink from a cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.4> to <17.3>." + } + ] + }, + { + "video_id": "51801", + "question_id": "51801", + "video_file": "Charades/videos/51801.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.4> to <17.3>." + } + ] + }, + { + "video_id": "51801", + "question_id": "51801", + "video_file": "Charades/videos/51801.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a drink out of a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.4> to <17.3>." + } + ] + }, + { + "video_id": "A8LZE", + "question_id": "A8LZE", + "video_file": "Charades/videos/A8LZE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.1> to <7.5>." + } + ] + }, + { + "video_id": "A8LZE", + "question_id": "A8LZE", + "video_file": "Charades/videos/A8LZE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a smiling person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "A8LZE", + "question_id": "A8LZE", + "video_file": "Charades/videos/A8LZE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <7.5>." + } + ] + }, + { + "video_id": "Y7WEK", + "question_id": "Y7WEK", + "video_file": "Charades/videos/Y7WEK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.6> to <29.2>." + } + ] + }, + { + "video_id": "XLR4B", + "question_id": "XLR4B", + "video_file": "Charades/videos/XLR4B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking in the kitchen' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <13.8>." + } + ] + }, + { + "video_id": "XLR4B", + "question_id": "XLR4B", + "video_file": "Charades/videos/XLR4B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sit in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.4> to <31.0>." + } + ] + }, + { + "video_id": "7NLXI", + "question_id": "7NLXI", + "video_file": "Charades/videos/7NLXI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.4> to <17.5>." + } + ] + }, + { + "video_id": "FSKKQ", + "question_id": "FSKKQ", + "video_file": "Charades/videos/FSKKQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person fix their hair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.6> to <33.8>." + } + ] + }, + { + "video_id": "FSKKQ", + "question_id": "FSKKQ", + "video_file": "Charades/videos/FSKKQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakens in their recreation room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.8> to <19.1>." + } + ] + }, + { + "video_id": "L1S6P", + "question_id": "L1S6P", + "video_file": "Charades/videos/L1S6P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person eats some of the snacks' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.8> to <19.0>." + } + ] + }, + { + "video_id": "L1S6P", + "question_id": "L1S6P", + "video_file": "Charades/videos/L1S6P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs into the kitchen'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "L1S6P", + "question_id": "L1S6P", + "video_file": "Charades/videos/L1S6P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.8> to <19.0>." + } + ] + }, + { + "video_id": "L1S6P", + "question_id": "L1S6P", + "video_file": "Charades/videos/L1S6P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs into the pantry' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "L1S6P", + "question_id": "L1S6P", + "video_file": "Charades/videos/L1S6P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a bag of chips' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.4> to <14.4>." + } + ] + }, + { + "video_id": "I0588", + "question_id": "I0588", + "video_file": "Charades/videos/I0588.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts a box on the counter'." + }, + { + "from": "gpt", + "value": "From <13.0> to <21.5>." + } + ] + }, + { + "video_id": "I0588", + "question_id": "I0588", + "video_file": "Charades/videos/I0588.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person puts a box into the cabinet pantry' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.0> to <21.5>." + } + ] + }, + { + "video_id": "I0588", + "question_id": "I0588", + "video_file": "Charades/videos/I0588.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put it on a shelf in a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.0> to <21.6>." + } + ] + }, + { + "video_id": "BB8Q0", + "question_id": "BB8Q0", + "video_file": "Charades/videos/BB8Q0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a box of crackers'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <25.8>." + } + ] + }, + { + "video_id": "3OQ81", + "question_id": "3OQ81", + "video_file": "Charades/videos/3OQ81.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting the bag on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.9>." + } + ] + }, + { + "video_id": "3OQ81", + "question_id": "3OQ81", + "video_file": "Charades/videos/3OQ81.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts some groceries on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "N8SO1", + "question_id": "N8SO1", + "video_file": "Charades/videos/N8SO1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is putting on shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <14.6>." + } + ] + }, + { + "video_id": "JP60G", + "question_id": "JP60G", + "video_file": "Charades/videos/JP60G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits on the couch'." + }, + { + "from": "gpt", + "value": "From <13.9> to <23.0>." + } + ] + }, + { + "video_id": "JP60G", + "question_id": "JP60G", + "video_file": "Charades/videos/JP60G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.9> to <10.9>." + } + ] + }, + { + "video_id": "JP60G", + "question_id": "JP60G", + "video_file": "Charades/videos/JP60G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes their shoes off'." + }, + { + "from": "gpt", + "value": "From <0.6> to <8.1>." + } + ] + }, + { + "video_id": "JP60G", + "question_id": "JP60G", + "video_file": "Charades/videos/JP60G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take a picture with their phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <18.1>." + } + ] + }, + { + "video_id": "JP60G", + "question_id": "JP60G", + "video_file": "Charades/videos/JP60G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes their shoes off int the kitchen' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.6> to <8.1>." + } + ] + }, + { + "video_id": "72VJD", + "question_id": "72VJD", + "video_file": "Charades/videos/72VJD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a phone to their ear' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "NHUCA", + "question_id": "NHUCA", + "video_file": "Charades/videos/NHUCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours coffee into a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <7.4>." + } + ] + }, + { + "video_id": "NHUCA", + "question_id": "NHUCA", + "video_file": "Charades/videos/NHUCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person pours a cup of coffee' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <7.4>." + } + ] + }, + { + "video_id": "VD8PU", + "question_id": "VD8PU", + "video_file": "Charades/videos/VD8PU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting in a chair holding a bowl' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.2>." + } + ] + }, + { + "video_id": "BRE3P", + "question_id": "BRE3P", + "video_file": "Charades/videos/BRE3P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pours a cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.4> to <27.4>." + } + ] + }, + { + "video_id": "BRE3P", + "question_id": "BRE3P", + "video_file": "Charades/videos/BRE3P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts groceries on the counter' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.6> to <16.1>." + } + ] + }, + { + "video_id": "BRE3P", + "question_id": "BRE3P", + "video_file": "Charades/videos/BRE3P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pours coffee into a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.4> to <27.4>." + } + ] + }, + { + "video_id": "1SKLG", + "question_id": "1SKLG", + "video_file": "Charades/videos/1SKLG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person is smiling with a broom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.0> to <31.0>." + } + ] + }, + { + "video_id": "UFWD5", + "question_id": "UFWD5", + "video_file": "Charades/videos/UFWD5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person the door will not open' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "UFWD5", + "question_id": "UFWD5", + "video_file": "Charades/videos/UFWD5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "UFWD5", + "question_id": "UFWD5", + "video_file": "Charades/videos/UFWD5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sitting on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <20.9>." + } + ] + }, + { + "video_id": "UFWD5", + "question_id": "UFWD5", + "video_file": "Charades/videos/UFWD5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person tries to open the door again' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "7T6MB", + "question_id": "7T6MB", + "video_file": "Charades/videos/7T6MB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes off shoes'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "JDG40", + "question_id": "JDG40", + "video_file": "Charades/videos/JDG40.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sneezing toward the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.2> to <27.3>." + } + ] + }, + { + "video_id": "DC4XY", + "question_id": "DC4XY", + "video_file": "Charades/videos/DC4XY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begin sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.9> to <11.4>." + } + ] + }, + { + "video_id": "334W9", + "question_id": "334W9", + "video_file": "Charades/videos/334W9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person is drinking from a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.6> to <22.0>." + } + ] + }, + { + "video_id": "SJQVM", + "question_id": "SJQVM", + "video_file": "Charades/videos/SJQVM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pouring water into a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <17.8>." + } + ] + }, + { + "video_id": "SJQVM", + "question_id": "SJQVM", + "video_file": "Charades/videos/SJQVM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person snuggles with a blanket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.2>." + } + ] + }, + { + "video_id": "QRM8K", + "question_id": "QRM8K", + "video_file": "Charades/videos/QRM8K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating a sandwich in a bedroom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <12.2>." + } + ] + }, + { + "video_id": "QRM8K", + "question_id": "QRM8K", + "video_file": "Charades/videos/QRM8K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats the sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <12.2>." + } + ] + }, + { + "video_id": "QRM8K", + "question_id": "QRM8K", + "video_file": "Charades/videos/QRM8K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a bit out of a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <7.4>." + } + ] + }, + { + "video_id": "QRM8K", + "question_id": "QRM8K", + "video_file": "Charades/videos/QRM8K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a cup off the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.2> to <24.0>." + } + ] + }, + { + "video_id": "0DY8C", + "question_id": "0DY8C", + "video_file": "Charades/videos/0DY8C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person watching television in a living room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "0DY8C", + "question_id": "0DY8C", + "video_file": "Charades/videos/0DY8C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person watching the television' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "CXGSD", + "question_id": "CXGSD", + "video_file": "Charades/videos/CXGSD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts tidying up stray clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.8> to <29.5>." + } + ] + }, + { + "video_id": "CXGSD", + "question_id": "CXGSD", + "video_file": "Charades/videos/CXGSD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws a pillow toward the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.5> to <13.5>." + } + ] + }, + { + "video_id": "R3ZN9", + "question_id": "R3ZN9", + "video_file": "Charades/videos/R3ZN9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sneezing on a picture'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <13.5>." + } + ] + }, + { + "video_id": "3ZH9Z", + "question_id": "3ZH9Z", + "video_file": "Charades/videos/3ZH9Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is eating at their desk' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "3ZH9Z", + "question_id": "3ZH9Z", + "video_file": "Charades/videos/3ZH9Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting down at a desk eating' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "LNXMJ", + "question_id": "LNXMJ", + "video_file": "Charades/videos/LNXMJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person turns the light off'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.6> to <27.8>." + } + ] + }, + { + "video_id": "LNXMJ", + "question_id": "LNXMJ", + "video_file": "Charades/videos/LNXMJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns off a light'." + }, + { + "from": "gpt", + "value": "From <22.6> to <27.8>." + } + ] + }, + { + "video_id": "B2NAS", + "question_id": "B2NAS", + "video_file": "Charades/videos/B2NAS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person have a bite to eat' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <18.0>." + } + ] + }, + { + "video_id": "B2NAS", + "question_id": "B2NAS", + "video_file": "Charades/videos/B2NAS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they are tidying the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.0> to <25.1>." + } + ] + }, + { + "video_id": "B2NAS", + "question_id": "B2NAS", + "video_file": "Charades/videos/B2NAS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is eating a sandwich in the man cave' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.6> to <16.9>." + } + ] + }, + { + "video_id": "J4RPL", + "question_id": "J4RPL", + "video_file": "Charades/videos/J4RPL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take a towel from their lap'." + }, + { + "from": "gpt", + "value": "From <20.8> to <26.8>." + } + ] + }, + { + "video_id": "J4RPL", + "question_id": "J4RPL", + "video_file": "Charades/videos/J4RPL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take some homework out of a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "WH1PD", + "question_id": "WH1PD", + "video_file": "Charades/videos/WH1PD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts camera down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <19.8>." + } + ] + }, + { + "video_id": "WH1PD", + "question_id": "WH1PD", + "video_file": "Charades/videos/WH1PD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person start drinking from a glass of water'." + }, + { + "from": "gpt", + "value": "From <19.9> to <31.0>." + } + ] + }, + { + "video_id": "JFR2I", + "question_id": "JFR2I", + "video_file": "Charades/videos/JFR2I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person move to pour more win into a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.7> to <33.0>." + } + ] + }, + { + "video_id": "KOHWO", + "question_id": "KOHWO", + "video_file": "Charades/videos/KOHWO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they close the laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.3> to <32.0>." + } + ] + }, + { + "video_id": "29WOY", + "question_id": "29WOY", + "video_file": "Charades/videos/29WOY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is opening a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "NNUCA", + "question_id": "NNUCA", + "video_file": "Charades/videos/NNUCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person close the door'." + }, + { + "from": "gpt", + "value": "From <7.0> to <14.8>." + } + ] + }, + { + "video_id": "NNUCA", + "question_id": "NNUCA", + "video_file": "Charades/videos/NNUCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'starts sneezing.the person proceeds to walk away from the doors' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <19.9>." + } + ] + }, + { + "video_id": "NNUCA", + "question_id": "NNUCA", + "video_file": "Charades/videos/NNUCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts a book down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.5> to <7.4>." + } + ] + }, + { + "video_id": "NNUCA", + "question_id": "NNUCA", + "video_file": "Charades/videos/NNUCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'takes out a pillow.the person closes the doors turn'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <9.8>." + } + ] + }, + { + "video_id": "LFC6D", + "question_id": "LFC6D", + "video_file": "Charades/videos/LFC6D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person start sneezing as they walk by a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.0> to <26.5>." + } + ] + }, + { + "video_id": "5NG8W", + "question_id": "5NG8W", + "video_file": "Charades/videos/5NG8W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a phone'." + }, + { + "from": "gpt", + "value": "From <29.7> to <33.0>." + } + ] + }, + { + "video_id": "5NG8W", + "question_id": "5NG8W", + "video_file": "Charades/videos/5NG8W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person playing with a phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <29.7> to <33.0>." + } + ] + }, + { + "video_id": "I5L3Y", + "question_id": "I5L3Y", + "video_file": "Charades/videos/I5L3Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person the door starts opening' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.4> to <11.7>." + } + ] + }, + { + "video_id": "I5L3Y", + "question_id": "I5L3Y", + "video_file": "Charades/videos/I5L3Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person walks in holding a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.7> to <13.5>." + } + ] + }, + { + "video_id": "I5L3Y", + "question_id": "I5L3Y", + "video_file": "Charades/videos/I5L3Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the laptop down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.8> to <14.0>." + } + ] + }, + { + "video_id": "NL0QB", + "question_id": "NL0QB", + "video_file": "Charades/videos/NL0QB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is playing on their laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.7>." + } + ] + }, + { + "video_id": "NL0QB", + "question_id": "NL0QB", + "video_file": "Charades/videos/NL0QB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closed the laptop'." + }, + { + "from": "gpt", + "value": "From <11.9> to <17.0>." + } + ] + }, + { + "video_id": "H2ZYY", + "question_id": "H2ZYY", + "video_file": "Charades/videos/H2ZYY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.9>." + } + ] + }, + { + "video_id": "H2ZYY", + "question_id": "H2ZYY", + "video_file": "Charades/videos/H2ZYY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws it on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.6> to <19.9>." + } + ] + }, + { + "video_id": "H2ZYY", + "question_id": "H2ZYY", + "video_file": "Charades/videos/H2ZYY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throw that on the floor too'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.6> to <19.9>." + } + ] + }, + { + "video_id": "H2ZYY", + "question_id": "H2ZYY", + "video_file": "Charades/videos/H2ZYY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is opening the cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.9>." + } + ] + }, + { + "video_id": "H2ZYY", + "question_id": "H2ZYY", + "video_file": "Charades/videos/H2ZYY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close the cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.9> to <13.6>." + } + ] + }, + { + "video_id": "YBW3D", + "question_id": "YBW3D", + "video_file": "Charades/videos/YBW3D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding onto a bag of groceries' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <9.4>." + } + ] + }, + { + "video_id": "YBW3D", + "question_id": "YBW3D", + "video_file": "Charades/videos/YBW3D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they awaken to start sweeping the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.8> to <23.0>." + } + ] + }, + { + "video_id": "YBW3D", + "question_id": "YBW3D", + "video_file": "Charades/videos/YBW3D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person runs past' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.4> to <31.4>." + } + ] + }, + { + "video_id": "YBW3D", + "question_id": "YBW3D", + "video_file": "Charades/videos/YBW3D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'awakens when person a walks pass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <23.0>." + } + ] + }, + { + "video_id": "YBW3D", + "question_id": "YBW3D", + "video_file": "Charades/videos/YBW3D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person a child runs past them' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.4> to <31.4>." + } + ] + }, + { + "video_id": "UDGRS", + "question_id": "UDGRS", + "video_file": "Charades/videos/UDGRS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take both the book'." + }, + { + "from": "gpt", + "value": "From <21.0> to <31.0>." + } + ] + }, + { + "video_id": "1YBKW", + "question_id": "1YBKW", + "video_file": "Charades/videos/1YBKW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person getting dressed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.3> to <28.9>." + } + ] + }, + { + "video_id": "1YBKW", + "question_id": "1YBKW", + "video_file": "Charades/videos/1YBKW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begin smiling into the mirror'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "1YBKW", + "question_id": "1YBKW", + "video_file": "Charades/videos/1YBKW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person beings getting dressed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.3> to <28.9>." + } + ] + }, + { + "video_id": "1YBKW", + "question_id": "1YBKW", + "video_file": "Charades/videos/1YBKW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they turn on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <22.0>." + } + ] + }, + { + "video_id": "1YBKW", + "question_id": "1YBKW", + "video_file": "Charades/videos/1YBKW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person turns on a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <22.0>." + } + ] + }, + { + "video_id": "KXCSL", + "question_id": "KXCSL", + "video_file": "Charades/videos/KXCSL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is standing in front of an opening refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "MQRFK", + "question_id": "MQRFK", + "video_file": "Charades/videos/MQRFK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person fixes a light above the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <20.4>." + } + ] + }, + { + "video_id": "MT9NE", + "question_id": "MT9NE", + "video_file": "Charades/videos/MT9NE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person continues to eat from the bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <11.3>." + } + ] + }, + { + "video_id": "MT9NE", + "question_id": "MT9NE", + "video_file": "Charades/videos/MT9NE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating from a bag of chips' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <11.3>." + } + ] + }, + { + "video_id": "MT9NE", + "question_id": "MT9NE", + "video_file": "Charades/videos/MT9NE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <11.3>." + } + ] + }, + { + "video_id": "MT9NE", + "question_id": "MT9NE", + "video_file": "Charades/videos/MT9NE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person continues to eat in bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.1> to <11.3>." + } + ] + }, + { + "video_id": "QHM66", + "question_id": "QHM66", + "video_file": "Charades/videos/QHM66.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person the food inside was taken to the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.5> to <17.4>." + } + ] + }, + { + "video_id": "QHM66", + "question_id": "QHM66", + "video_file": "Charades/videos/QHM66.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person the box was opened'." + }, + { + "from": "gpt", + "value": "From <8.4> to <15.4>." + } + ] + }, + { + "video_id": "QHM66", + "question_id": "QHM66", + "video_file": "Charades/videos/QHM66.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opened a box of fruits on the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.4> to <15.4>." + } + ] + }, + { + "video_id": "00SL4", + "question_id": "00SL4", + "video_file": "Charades/videos/00SL4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is opening a closet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.6>." + } + ] + }, + { + "video_id": "9YRHF", + "question_id": "9YRHF", + "video_file": "Charades/videos/9YRHF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throwing a towel on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.7> to <9.9>." + } + ] + }, + { + "video_id": "9YRHF", + "question_id": "9YRHF", + "video_file": "Charades/videos/9YRHF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "9YRHF", + "question_id": "9YRHF", + "video_file": "Charades/videos/9YRHF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is standing in a closet holding a towel'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "COBEK", + "question_id": "COBEK", + "video_file": "Charades/videos/COBEK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.8> to <13.2>." + } + ] + }, + { + "video_id": "QN4MZ", + "question_id": "QN4MZ", + "video_file": "Charades/videos/QN4MZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person kept sneezing'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.2>." + } + ] + }, + { + "video_id": "1BVUA", + "question_id": "1BVUA", + "video_file": "Charades/videos/1BVUA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a smiling person awakens with a towel under their head'." + }, + { + "from": "gpt", + "value": "From <27.1> to <31.0>." + } + ] + }, + { + "video_id": "3WAWR", + "question_id": "3WAWR", + "video_file": "Charades/videos/3WAWR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes out some medicine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.9> to <24.4>." + } + ] + }, + { + "video_id": "3WAWR", + "question_id": "3WAWR", + "video_file": "Charades/videos/3WAWR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person is drinking from a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.3>." + } + ] + }, + { + "video_id": "05GIA", + "question_id": "05GIA", + "video_file": "Charades/videos/05GIA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts it on a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.2> to <13.0>." + } + ] + }, + { + "video_id": "YMWUS", + "question_id": "YMWUS", + "video_file": "Charades/videos/YMWUS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a man getting dressed inside of a closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.2> to <38.9>." + } + ] + }, + { + "video_id": "YMWUS", + "question_id": "YMWUS", + "video_file": "Charades/videos/YMWUS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is dressing by a wardrobe' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.2> to <38.9>." + } + ] + }, + { + "video_id": "P6P6L", + "question_id": "P6P6L", + "video_file": "Charades/videos/P6P6L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person begins smiling'." + }, + { + "from": "gpt", + "value": "From <19.5> to <26.9>." + } + ] + }, + { + "video_id": "YM213", + "question_id": "YM213", + "video_file": "Charades/videos/YM213.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the door a few more times'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "YM213", + "question_id": "YM213", + "video_file": "Charades/videos/YM213.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes a door from inside a bathroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "YM213", + "question_id": "YM213", + "video_file": "Charades/videos/YM213.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closing the door on themselves' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "MFSGO", + "question_id": "MFSGO", + "video_file": "Charades/videos/MFSGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the window in the bathroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <17.6>." + } + ] + }, + { + "video_id": "LK3BW", + "question_id": "LK3BW", + "video_file": "Charades/videos/LK3BW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns on the entryway light'." + }, + { + "from": "gpt", + "value": "From <6.7> to <12.3>." + } + ] + }, + { + "video_id": "LK3BW", + "question_id": "LK3BW", + "video_file": "Charades/videos/LK3BW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "LK3BW", + "question_id": "LK3BW", + "video_file": "Charades/videos/LK3BW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they turn a light on in the hallway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.7> to <12.3>." + } + ] + }, + { + "video_id": "LK3BW", + "question_id": "LK3BW", + "video_file": "Charades/videos/LK3BW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "LK3BW", + "question_id": "LK3BW", + "video_file": "Charades/videos/LK3BW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they close the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "LK3BW", + "question_id": "LK3BW", + "video_file": "Charades/videos/LK3BW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person walks through a doorway into a room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.2>." + } + ] + }, + { + "video_id": "6E4JA", + "question_id": "6E4JA", + "video_file": "Charades/videos/6E4JA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person in the corner is eating food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "L1BNN", + "question_id": "L1BNN", + "video_file": "Charades/videos/L1BNN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the medicine on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.6> to <28.7>." + } + ] + }, + { + "video_id": "L1BNN", + "question_id": "L1BNN", + "video_file": "Charades/videos/L1BNN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts it back on a shelf'." + }, + { + "from": "gpt", + "value": "From <19.6> to <28.7>." + } + ] + }, + { + "video_id": "LVJ3U", + "question_id": "LVJ3U", + "video_file": "Charades/videos/LVJ3U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walk over to the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.0> to <10.0>." + } + ] + }, + { + "video_id": "LVJ3U", + "question_id": "LVJ3U", + "video_file": "Charades/videos/LVJ3U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person walks to a light switch'." + }, + { + "from": "gpt", + "value": "From <6.0> to <10.0>." + } + ] + }, + { + "video_id": "LVJ3U", + "question_id": "LVJ3U", + "video_file": "Charades/videos/LVJ3U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <10.0>." + } + ] + }, + { + "video_id": "86GSE", + "question_id": "86GSE", + "video_file": "Charades/videos/86GSE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.2> to <19.1>." + } + ] + }, + { + "video_id": "86GSE", + "question_id": "86GSE", + "video_file": "Charades/videos/86GSE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person turns on a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <10.9>." + } + ] + }, + { + "video_id": "86GSE", + "question_id": "86GSE", + "video_file": "Charades/videos/86GSE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person walks over to a light'." + }, + { + "from": "gpt", + "value": "From <5.3> to <10.9>." + } + ] + }, + { + "video_id": "86GSE", + "question_id": "86GSE", + "video_file": "Charades/videos/86GSE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes their shoes off' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.2> to <19.1>." + } + ] + }, + { + "video_id": "K71S1", + "question_id": "K71S1", + "video_file": "Charades/videos/K71S1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is opening a box'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "K71S1", + "question_id": "K71S1", + "video_file": "Charades/videos/K71S1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting clothes into it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.0> to <18.3>." + } + ] + }, + { + "video_id": "EF5XT", + "question_id": "EF5XT", + "video_file": "Charades/videos/EF5XT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins pouring water in a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.8> to <19.7>." + } + ] + }, + { + "video_id": "WX4AP", + "question_id": "WX4AP", + "video_file": "Charades/videos/WX4AP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <16.5>." + } + ] + }, + { + "video_id": "R87CC", + "question_id": "R87CC", + "video_file": "Charades/videos/R87CC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sits on the sofa' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.8>." + } + ] + }, + { + "video_id": "R87CC", + "question_id": "R87CC", + "video_file": "Charades/videos/R87CC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person open a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.6> to <15.0>." + } + ] + }, + { + "video_id": "W20LY", + "question_id": "W20LY", + "video_file": "Charades/videos/W20LY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts a towel around their shoulders' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.2> to <25.6>." + } + ] + }, + { + "video_id": "DS22L", + "question_id": "DS22L", + "video_file": "Charades/videos/DS22L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person throws clothes on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <17.4>." + } + ] + }, + { + "video_id": "DS22L", + "question_id": "DS22L", + "video_file": "Charades/videos/DS22L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person throws the clothes onto the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <17.4>." + } + ] + }, + { + "video_id": "DS22L", + "question_id": "DS22L", + "video_file": "Charades/videos/DS22L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinks from a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <26.3>." + } + ] + }, + { + "video_id": "BGX4T", + "question_id": "BGX4T", + "video_file": "Charades/videos/BGX4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.3> to <16.0>." + } + ] + }, + { + "video_id": "BGX4T", + "question_id": "BGX4T", + "video_file": "Charades/videos/BGX4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they close the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.3> to <16.0>." + } + ] + }, + { + "video_id": "XWY6P", + "question_id": "XWY6P", + "video_file": "Charades/videos/XWY6P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they put the phone down on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.8> to <17.8>." + } + ] + }, + { + "video_id": "XWY6P", + "question_id": "XWY6P", + "video_file": "Charades/videos/XWY6P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is opening the closet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "XWY6P", + "question_id": "XWY6P", + "video_file": "Charades/videos/XWY6P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'this person is laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.1> to <14.4>." + } + ] + }, + { + "video_id": "Z1DBB", + "question_id": "Z1DBB", + "video_file": "Charades/videos/Z1DBB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take clothes out of a hamper' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.3> to <39.3>." + } + ] + }, + { + "video_id": "Z1DBB", + "question_id": "Z1DBB", + "video_file": "Charades/videos/Z1DBB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eat from it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <23.5>." + } + ] + }, + { + "video_id": "Z1DBB", + "question_id": "Z1DBB", + "video_file": "Charades/videos/Z1DBB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person eats from a plate in a bedroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <23.5>." + } + ] + }, + { + "video_id": "1XFDY", + "question_id": "1XFDY", + "video_file": "Charades/videos/1XFDY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person so they start walking towards the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.9> to <32.0>." + } + ] + }, + { + "video_id": "RS3SQ", + "question_id": "RS3SQ", + "video_file": "Charades/videos/RS3SQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'man opens up the refrigerator looking for something' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.7> to <8.1>." + } + ] + }, + { + "video_id": "RS3SQ", + "question_id": "RS3SQ", + "video_file": "Charades/videos/RS3SQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is eating food from the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <20.3>." + } + ] + }, + { + "video_id": "N7IH0", + "question_id": "N7IH0", + "video_file": "Charades/videos/N7IH0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person also opens a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <17.7>." + } + ] + }, + { + "video_id": "N7IH0", + "question_id": "N7IH0", + "video_file": "Charades/videos/N7IH0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes out medicine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.9> to <29.1>." + } + ] + }, + { + "video_id": "ROL2T", + "question_id": "ROL2T", + "video_file": "Charades/videos/ROL2T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person in a towel starts dressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <12.2>." + } + ] + }, + { + "video_id": "ROL2T", + "question_id": "ROL2T", + "video_file": "Charades/videos/ROL2T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts on other clothes'." + }, + { + "from": "gpt", + "value": "From <10.5> to <16.0>." + } + ] + }, + { + "video_id": "ROL2T", + "question_id": "ROL2T", + "video_file": "Charades/videos/ROL2T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person the other takes off some clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <8.5>." + } + ] + }, + { + "video_id": "0WHOH", + "question_id": "0WHOH", + "video_file": "Charades/videos/0WHOH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person with hair in a towel undresses' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.0>." + } + ] + }, + { + "video_id": "6JB5R", + "question_id": "6JB5R", + "video_file": "Charades/videos/6JB5R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they start sneezing'." + }, + { + "from": "gpt", + "value": "From <103.2> to <133.1>." + } + ] + }, + { + "video_id": "6JB5R", + "question_id": "6JB5R", + "video_file": "Charades/videos/6JB5R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sit in the chair eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.7> to <31.1>." + } + ] + }, + { + "video_id": "6JB5R", + "question_id": "6JB5R", + "video_file": "Charades/videos/6JB5R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits down on chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.7> to <31.1>." + } + ] + }, + { + "video_id": "NSKMC", + "question_id": "NSKMC", + "video_file": "Charades/videos/NSKMC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person is drinking a cup of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.1> to <19.9>." + } + ] + }, + { + "video_id": "H8UJ3", + "question_id": "H8UJ3", + "video_file": "Charades/videos/H8UJ3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person turned on the light'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "H8UJ3", + "question_id": "H8UJ3", + "video_file": "Charades/videos/H8UJ3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opened the book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.1> to <23.1>." + } + ] + }, + { + "video_id": "H8UJ3", + "question_id": "H8UJ3", + "video_file": "Charades/videos/H8UJ3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person sits down on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <17.5>." + } + ] + }, + { + "video_id": "H8UJ3", + "question_id": "H8UJ3", + "video_file": "Charades/videos/H8UJ3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person turns on the lights to the dining room'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "NGMK1", + "question_id": "NGMK1", + "video_file": "Charades/videos/NGMK1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person drinks out of a red cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.0> to <10.0>." + } + ] + }, + { + "video_id": "DTJS6", + "question_id": "DTJS6", + "video_file": "Charades/videos/DTJS6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person look out the window'." + }, + { + "from": "gpt", + "value": "From <18.1> to <25.3>." + } + ] + }, + { + "video_id": "DTJS6", + "question_id": "DTJS6", + "video_file": "Charades/videos/DTJS6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person running with a sandwich in the bedroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <28.8> to <35.8>." + } + ] + }, + { + "video_id": "DTJS6", + "question_id": "DTJS6", + "video_file": "Charades/videos/DTJS6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they put the broom down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.8> to <16.1>." + } + ] + }, + { + "video_id": "DTJS6", + "question_id": "DTJS6", + "video_file": "Charades/videos/DTJS6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looks out the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <25.3>." + } + ] + }, + { + "video_id": "A4VK8", + "question_id": "A4VK8", + "video_file": "Charades/videos/A4VK8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <19.0>." + } + ] + }, + { + "video_id": "A4VK8", + "question_id": "A4VK8", + "video_file": "Charades/videos/A4VK8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinks a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <15.3>." + } + ] + }, + { + "video_id": "A4VK8", + "question_id": "A4VK8", + "video_file": "Charades/videos/A4VK8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drink from a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <15.3>." + } + ] + }, + { + "video_id": "A4VK8", + "question_id": "A4VK8", + "video_file": "Charades/videos/A4VK8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person start eating' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <19.0>." + } + ] + }, + { + "video_id": "A4VK8", + "question_id": "A4VK8", + "video_file": "Charades/videos/A4VK8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they put the phone down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.1> to <9.9>." + } + ] + }, + { + "video_id": "K4UOG", + "question_id": "K4UOG", + "video_file": "Charades/videos/K4UOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person lookin at the picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.0>." + } + ] + }, + { + "video_id": "K4UOG", + "question_id": "K4UOG", + "video_file": "Charades/videos/K4UOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats the sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <29.0>." + } + ] + }, + { + "video_id": "K4UOG", + "question_id": "K4UOG", + "video_file": "Charades/videos/K4UOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looking at a picture' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.0>." + } + ] + }, + { + "video_id": "K4UOG", + "question_id": "K4UOG", + "video_file": "Charades/videos/K4UOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a sandwich from the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.1> to <16.8>." + } + ] + }, + { + "video_id": "K4UOG", + "question_id": "K4UOG", + "video_file": "Charades/videos/K4UOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes a sandwich from a dish' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.1> to <16.8>." + } + ] + }, + { + "video_id": "291LH", + "question_id": "291LH", + "video_file": "Charades/videos/291LH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking from a cup'." + }, + { + "from": "gpt", + "value": "From <0.7> to <6.7>." + } + ] + }, + { + "video_id": "291LH", + "question_id": "291LH", + "video_file": "Charades/videos/291LH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking water out of a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.7> to <6.7>." + } + ] + }, + { + "video_id": "0VOQC", + "question_id": "0VOQC", + "video_file": "Charades/videos/0VOQC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person runs in with a phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "0VOQC", + "question_id": "0VOQC", + "video_file": "Charades/videos/0VOQC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs into the room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "16WZF", + "question_id": "16WZF", + "video_file": "Charades/videos/16WZF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a broom'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "16WZF", + "question_id": "16WZF", + "video_file": "Charades/videos/16WZF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the broom down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <8.5>." + } + ] + }, + { + "video_id": "16WZF", + "question_id": "16WZF", + "video_file": "Charades/videos/16WZF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they put the broom aside' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.8> to <8.5>." + } + ] + }, + { + "video_id": "OHBYF", + "question_id": "OHBYF", + "video_file": "Charades/videos/OHBYF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <24.8>." + } + ] + }, + { + "video_id": "HDHWS", + "question_id": "HDHWS", + "video_file": "Charades/videos/HDHWS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes a few notebooks' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.4> to <15.6>." + } + ] + }, + { + "video_id": "HDHWS", + "question_id": "HDHWS", + "video_file": "Charades/videos/HDHWS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a picture of themselves'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "GWQ73", + "question_id": "GWQ73", + "video_file": "Charades/videos/GWQ73.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is eating food slowly'." + }, + { + "from": "gpt", + "value": "From <8.7> to <14.2>." + } + ] + }, + { + "video_id": "2WX2Z", + "question_id": "2WX2Z", + "video_file": "Charades/videos/2WX2Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person leaves smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.6> to <12.7>." + } + ] + }, + { + "video_id": "SQKE4", + "question_id": "SQKE4", + "video_file": "Charades/videos/SQKE4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the cup down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.0> to <9.3>." + } + ] + }, + { + "video_id": "SQKE4", + "question_id": "SQKE4", + "video_file": "Charades/videos/SQKE4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a cup of coffee' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "SQKE4", + "question_id": "SQKE4", + "video_file": "Charades/videos/SQKE4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person watching television'." + }, + { + "from": "gpt", + "value": "From <6.5> to <11.3>." + } + ] + }, + { + "video_id": "SQKE4", + "question_id": "SQKE4", + "video_file": "Charades/videos/SQKE4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting on a couch holding a cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "5LE93", + "question_id": "5LE93", + "video_file": "Charades/videos/5LE93.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinks from a glass'." + }, + { + "from": "gpt", + "value": "From <1.2> to <8.6>." + } + ] + }, + { + "video_id": "5LE93", + "question_id": "5LE93", + "video_file": "Charades/videos/5LE93.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they are smiling'." + }, + { + "from": "gpt", + "value": "From <7.5> to <13.4>." + } + ] + }, + { + "video_id": "VRMNI", + "question_id": "VRMNI", + "video_file": "Charades/videos/VRMNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they wash the dish in the sink'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.4> to <32.8>." + } + ] + }, + { + "video_id": "VRMNI", + "question_id": "VRMNI", + "video_file": "Charades/videos/VRMNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "VRMNI", + "question_id": "VRMNI", + "video_file": "Charades/videos/VRMNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts putting the dishes on the shelf'." + }, + { + "from": "gpt", + "value": "From <29.8> to <35.0>." + } + ] + }, + { + "video_id": "VRMNI", + "question_id": "VRMNI", + "video_file": "Charades/videos/VRMNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they turn on the light in the room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.6> to <11.0>." + } + ] + }, + { + "video_id": "VRMNI", + "question_id": "VRMNI", + "video_file": "Charades/videos/VRMNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person turns on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <11.0>." + } + ] + }, + { + "video_id": "VRMNI", + "question_id": "VRMNI", + "video_file": "Charades/videos/VRMNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person brings dishes to the sink washes them' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.4> to <32.8>." + } + ] + }, + { + "video_id": "VRMNI", + "question_id": "VRMNI", + "video_file": "Charades/videos/VRMNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person enters through a doorway turns on a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <11.0>." + } + ] + }, + { + "video_id": "FQDS1", + "question_id": "FQDS1", + "video_file": "Charades/videos/FQDS1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is reading a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "FQDS1", + "question_id": "FQDS1", + "video_file": "Charades/videos/FQDS1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'there is a person holding a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.2>." + } + ] + }, + { + "video_id": "FQDS1", + "question_id": "FQDS1", + "video_file": "Charades/videos/FQDS1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is in their bathroom reading a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "ZZ89F", + "question_id": "ZZ89F", + "video_file": "Charades/videos/ZZ89F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pour it into a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.2> to <36.0>." + } + ] + }, + { + "video_id": "ZZ89F", + "question_id": "ZZ89F", + "video_file": "Charades/videos/ZZ89F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <14.0>." + } + ] + }, + { + "video_id": "ZZ89F", + "question_id": "ZZ89F", + "video_file": "Charades/videos/ZZ89F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person walks into a kitchen through a doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.6>." + } + ] + }, + { + "video_id": "9WLBQ", + "question_id": "9WLBQ", + "video_file": "Charades/videos/9WLBQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person someone is in their kitchen washing dishes'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.3>." + } + ] + }, + { + "video_id": "9WLBQ", + "question_id": "9WLBQ", + "video_file": "Charades/videos/9WLBQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person washes some dishes in the sink' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.3>." + } + ] + }, + { + "video_id": "9WLBQ", + "question_id": "9WLBQ", + "video_file": "Charades/videos/9WLBQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they open the refrigerator'." + }, + { + "from": "gpt", + "value": "From <12.1> to <19.6>." + } + ] + }, + { + "video_id": "B1FY9", + "question_id": "B1FY9", + "video_file": "Charades/videos/B1FY9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person another is throwing food in a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.3> to <28.6>." + } + ] + }, + { + "video_id": "JF0O8", + "question_id": "JF0O8", + "video_file": "Charades/videos/JF0O8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'this person is sitting on the couch holding a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.8>." + } + ] + }, + { + "video_id": "7B8EQ", + "question_id": "7B8EQ", + "video_file": "Charades/videos/7B8EQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.3> to <13.9>." + } + ] + }, + { + "video_id": "7B8EQ", + "question_id": "7B8EQ", + "video_file": "Charades/videos/7B8EQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person dressing in different clothes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.3> to <13.9>." + } + ] + }, + { + "video_id": "IG1YT", + "question_id": "IG1YT", + "video_file": "Charades/videos/IG1YT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.6> to <20.3>." + } + ] + }, + { + "video_id": "IG1YT", + "question_id": "IG1YT", + "video_file": "Charades/videos/IG1YT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting on a pillow on a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.1>." + } + ] + }, + { + "video_id": "IG1YT", + "question_id": "IG1YT", + "video_file": "Charades/videos/IG1YT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns the lights off' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.8> to <26.9>." + } + ] + }, + { + "video_id": "M5C5L", + "question_id": "M5C5L", + "video_file": "Charades/videos/M5C5L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person proceeds to cook on a stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <31.7>." + } + ] + }, + { + "video_id": "M5C5L", + "question_id": "M5C5L", + "video_file": "Charades/videos/M5C5L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is smiling next to a refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.6> to <13.1>." + } + ] + }, + { + "video_id": "M5C5L", + "question_id": "M5C5L", + "video_file": "Charades/videos/M5C5L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person cooking something on the stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <31.7>." + } + ] + }, + { + "video_id": "BID6Z", + "question_id": "BID6Z", + "video_file": "Charades/videos/BID6Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put the towel on a shelf'." + }, + { + "from": "gpt", + "value": "From <21.4> to <29.4>." + } + ] + }, + { + "video_id": "BID6Z", + "question_id": "BID6Z", + "video_file": "Charades/videos/BID6Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a closet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.8> to <34.0>." + } + ] + }, + { + "video_id": "BID6Z", + "question_id": "BID6Z", + "video_file": "Charades/videos/BID6Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.8> to <34.0>." + } + ] + }, + { + "video_id": "BID6Z", + "question_id": "BID6Z", + "video_file": "Charades/videos/BID6Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the towel in the closet'." + }, + { + "from": "gpt", + "value": "From <21.4> to <29.4>." + } + ] + }, + { + "video_id": "BID6Z", + "question_id": "BID6Z", + "video_file": "Charades/videos/BID6Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person close the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <29.8> to <34.0>." + } + ] + }, + { + "video_id": "BID6Z", + "question_id": "BID6Z", + "video_file": "Charades/videos/BID6Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a door to a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.4> to <25.8>." + } + ] + }, + { + "video_id": "QT4WM", + "question_id": "QT4WM", + "video_file": "Charades/videos/QT4WM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is opening a cabinet in their closet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <9.3>." + } + ] + }, + { + "video_id": "QT4WM", + "question_id": "QT4WM", + "video_file": "Charades/videos/QT4WM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a linen closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.4> to <9.3>." + } + ] + }, + { + "video_id": "785X2", + "question_id": "785X2", + "video_file": "Charades/videos/785X2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes a bottle of medicine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.0> to <39.7>." + } + ] + }, + { + "video_id": "785X2", + "question_id": "785X2", + "video_file": "Charades/videos/785X2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person close a laptop'." + }, + { + "from": "gpt", + "value": "From <39.2> to <55.0>." + } + ] + }, + { + "video_id": "785X2", + "question_id": "785X2", + "video_file": "Charades/videos/785X2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakened' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.5> to <30.7>." + } + ] + }, + { + "video_id": "F8TWE", + "question_id": "F8TWE", + "video_file": "Charades/videos/F8TWE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person is putting a bag down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.4> to <33.0>." + } + ] + }, + { + "video_id": "0TXYV", + "question_id": "0TXYV", + "video_file": "Charades/videos/0TXYV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking a bottle of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <16.3>." + } + ] + }, + { + "video_id": "TNWWG", + "question_id": "TNWWG", + "video_file": "Charades/videos/TNWWG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts some stuff on a shelf,'." + }, + { + "from": "gpt", + "value": "From <11.5> to <20.8>." + } + ] + }, + { + "video_id": "TNWWG", + "question_id": "TNWWG", + "video_file": "Charades/videos/TNWWG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.7> to <19.9>." + } + ] + }, + { + "video_id": "TNWWG", + "question_id": "TNWWG", + "video_file": "Charades/videos/TNWWG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts some pillows on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.5> to <20.8>." + } + ] + }, + { + "video_id": "7TNGJ", + "question_id": "7TNGJ", + "video_file": "Charades/videos/7TNGJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes some medicine' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.7> to <14.7>." + } + ] + }, + { + "video_id": "7TNGJ", + "question_id": "7TNGJ", + "video_file": "Charades/videos/7TNGJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a medicine cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.4> to <5.9>." + } + ] + }, + { + "video_id": "7TNGJ", + "question_id": "7TNGJ", + "video_file": "Charades/videos/7TNGJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is seen opening the medicine cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <5.9>." + } + ] + }, + { + "video_id": "IDFTJ", + "question_id": "IDFTJ", + "video_file": "Charades/videos/IDFTJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts them into a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.3> to <31.0>." + } + ] + }, + { + "video_id": "IDFTJ", + "question_id": "IDFTJ", + "video_file": "Charades/videos/IDFTJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put them in a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.3> to <31.0>." + } + ] + }, + { + "video_id": "J3ZQS", + "question_id": "J3ZQS", + "video_file": "Charades/videos/J3ZQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'that person takes homework out of the bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "J3ZQS", + "question_id": "J3ZQS", + "video_file": "Charades/videos/J3ZQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <7.5>." + } + ] + }, + { + "video_id": "J3ZQS", + "question_id": "J3ZQS", + "video_file": "Charades/videos/J3ZQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person drinks from a cup of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.9> to <33.8>." + } + ] + }, + { + "video_id": "TLJRV", + "question_id": "TLJRV", + "video_file": "Charades/videos/TLJRV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person takes a box from under the sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <17.6>." + } + ] + }, + { + "video_id": "TLJRV", + "question_id": "TLJRV", + "video_file": "Charades/videos/TLJRV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.6> to <44.0>." + } + ] + }, + { + "video_id": "TLJRV", + "question_id": "TLJRV", + "video_file": "Charades/videos/TLJRV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is holding dishes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.7> to <20.2>." + } + ] + }, + { + "video_id": "GEAB1", + "question_id": "GEAB1", + "video_file": "Charades/videos/GEAB1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins drinking water from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.1> to <33.0>." + } + ] + }, + { + "video_id": "GEAB1", + "question_id": "GEAB1", + "video_file": "Charades/videos/GEAB1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins to eat a sandwich'." + }, + { + "from": "gpt", + "value": "From <14.7> to <29.9>." + } + ] + }, + { + "video_id": "GEAB1", + "question_id": "GEAB1", + "video_file": "Charades/videos/GEAB1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes a bite of their sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <18.4>." + } + ] + }, + { + "video_id": "YPTHX", + "question_id": "YPTHX", + "video_file": "Charades/videos/YPTHX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes a phone from a bag'." + }, + { + "from": "gpt", + "value": "From <6.9> to <17.7>." + } + ] + }, + { + "video_id": "YPTHX", + "question_id": "YPTHX", + "video_file": "Charades/videos/YPTHX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes out a phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.9> to <17.7>." + } + ] + }, + { + "video_id": "YPTHX", + "question_id": "YPTHX", + "video_file": "Charades/videos/YPTHX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person the put the phone in their pocket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.7> to <18.1>." + } + ] + }, + { + "video_id": "XQ4RB", + "question_id": "XQ4RB", + "video_file": "Charades/videos/XQ4RB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person turns on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.6> to <29.2>." + } + ] + }, + { + "video_id": "XQ4RB", + "question_id": "XQ4RB", + "video_file": "Charades/videos/XQ4RB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.6> to <29.2>." + } + ] + }, + { + "video_id": "IAK2E", + "question_id": "IAK2E", + "video_file": "Charades/videos/IAK2E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a glass'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "IAK2E", + "question_id": "IAK2E", + "video_file": "Charades/videos/IAK2E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <11.3>." + } + ] + }, + { + "video_id": "6BI70", + "question_id": "6BI70", + "video_file": "Charades/videos/6BI70.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.9> to <11.4>." + } + ] + }, + { + "video_id": "K9MPN", + "question_id": "K9MPN", + "video_file": "Charades/videos/K9MPN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person open the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <11.0>." + } + ] + }, + { + "video_id": "K9MPN", + "question_id": "K9MPN", + "video_file": "Charades/videos/K9MPN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <13.5>." + } + ] + }, + { + "video_id": "IRGXQ", + "question_id": "IRGXQ", + "video_file": "Charades/videos/IRGXQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs through the doorway laughing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <11.1>." + } + ] + }, + { + "video_id": "IRGXQ", + "question_id": "IRGXQ", + "video_file": "Charades/videos/IRGXQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs into a room'." + }, + { + "from": "gpt", + "value": "From <2.7> to <11.1>." + } + ] + }, + { + "video_id": "I6KKY", + "question_id": "I6KKY", + "video_file": "Charades/videos/I6KKY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <7.5>." + } + ] + }, + { + "video_id": "I6KKY", + "question_id": "I6KKY", + "video_file": "Charades/videos/I6KKY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <7.5>." + } + ] + }, + { + "video_id": "I6KKY", + "question_id": "I6KKY", + "video_file": "Charades/videos/I6KKY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes a towel out' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <8.7>." + } + ] + }, + { + "video_id": "DJR5H", + "question_id": "DJR5H", + "video_file": "Charades/videos/DJR5H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits on the sofa'." + }, + { + "from": "gpt", + "value": "From <20.2> to <31.0>." + } + ] + }, + { + "video_id": "CJ58B", + "question_id": "CJ58B", + "video_file": "Charades/videos/CJ58B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes some food from a bag of groceries' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <19.9>." + } + ] + }, + { + "video_id": "CJ58B", + "question_id": "CJ58B", + "video_file": "Charades/videos/CJ58B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts eating' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.1> to <32.0>." + } + ] + }, + { + "video_id": "CJ58B", + "question_id": "CJ58B", + "video_file": "Charades/videos/CJ58B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats it'." + }, + { + "from": "gpt", + "value": "From <23.1> to <32.0>." + } + ] + }, + { + "video_id": "CJ58B", + "question_id": "CJ58B", + "video_file": "Charades/videos/CJ58B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens bag of groceries' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.6> to <20.8>." + } + ] + }, + { + "video_id": "EV4HN", + "question_id": "EV4HN", + "video_file": "Charades/videos/EV4HN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person runs out of the room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.7> to <25.8>." + } + ] + }, + { + "video_id": "QXEWB", + "question_id": "QXEWB", + "video_file": "Charades/videos/QXEWB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person throws a book onto a couch' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "QXEWB", + "question_id": "QXEWB", + "video_file": "Charades/videos/QXEWB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is throwing a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "NCZOV", + "question_id": "NCZOV", + "video_file": "Charades/videos/NCZOV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a door on a doorknob' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "7FAB3", + "question_id": "7FAB3", + "video_file": "Charades/videos/7FAB3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is standing in the kitchen opening a refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.3> to <8.2>." + } + ] + }, + { + "video_id": "O9DOT", + "question_id": "O9DOT", + "video_file": "Charades/videos/O9DOT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person playing with their laptop on the stairs' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.8>." + } + ] + }, + { + "video_id": "HKS9J", + "question_id": "HKS9J", + "video_file": "Charades/videos/HKS9J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes out a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.7> to <14.4>." + } + ] + }, + { + "video_id": "RIV6I", + "question_id": "RIV6I", + "video_file": "Charades/videos/RIV6I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person standing in a doorway sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "RIV6I", + "question_id": "RIV6I", + "video_file": "Charades/videos/RIV6I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sneezing on a doorknob' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "BR8B3", + "question_id": "BR8B3", + "video_file": "Charades/videos/BR8B3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.0> to <32.0>." + } + ] + }, + { + "video_id": "BR8B3", + "question_id": "BR8B3", + "video_file": "Charades/videos/BR8B3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they eat some food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.9> to <32.0>." + } + ] + }, + { + "video_id": "BR8B3", + "question_id": "BR8B3", + "video_file": "Charades/videos/BR8B3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is lying on a couch eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.0> to <32.0>." + } + ] + }, + { + "video_id": "JJKNY", + "question_id": "JJKNY", + "video_file": "Charades/videos/JJKNY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <11.0>." + } + ] + }, + { + "video_id": "JJKNY", + "question_id": "JJKNY", + "video_file": "Charades/videos/JJKNY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a sneezing person is undressing in their bathroom'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "GDIW0", + "question_id": "GDIW0", + "video_file": "Charades/videos/GDIW0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they close the laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.6> to <24.0>." + } + ] + }, + { + "video_id": "GDIW0", + "question_id": "GDIW0", + "video_file": "Charades/videos/GDIW0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looks out the window for a moment' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <32.2>." + } + ] + }, + { + "video_id": "GDIW0", + "question_id": "GDIW0", + "video_file": "Charades/videos/GDIW0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person look out the window'." + }, + { + "from": "gpt", + "value": "From <18.7> to <32.2>." + } + ] + }, + { + "video_id": "TKOYQ", + "question_id": "TKOYQ", + "video_file": "Charades/videos/TKOYQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'this person takes some comic books off the shelf'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "BV5CC", + "question_id": "BV5CC", + "video_file": "Charades/videos/BV5CC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <21.3>." + } + ] + }, + { + "video_id": "BV5CC", + "question_id": "BV5CC", + "video_file": "Charades/videos/BV5CC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person start to eat' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.4> to <31.0>." + } + ] + }, + { + "video_id": "BV5CC", + "question_id": "BV5CC", + "video_file": "Charades/videos/BV5CC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they take a sandwich from the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <23.4>." + } + ] + }, + { + "video_id": "BV5CC", + "question_id": "BV5CC", + "video_file": "Charades/videos/BV5CC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person start eating it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.4> to <31.0>." + } + ] + }, + { + "video_id": "BV5CC", + "question_id": "BV5CC", + "video_file": "Charades/videos/BV5CC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they reach into an open cabinet'." + }, + { + "from": "gpt", + "value": "From <0.6> to <5.6>." + } + ] + }, + { + "video_id": "B0BDR", + "question_id": "B0BDR", + "video_file": "Charades/videos/B0BDR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they are slightly smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.2> to <16.7>." + } + ] + }, + { + "video_id": "46Z16", + "question_id": "46Z16", + "video_file": "Charades/videos/46Z16.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put on their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.6>." + } + ] + }, + { + "video_id": "46Z16", + "question_id": "46Z16", + "video_file": "Charades/videos/46Z16.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put their shoes on'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.6>." + } + ] + }, + { + "video_id": "Z0KN7", + "question_id": "Z0KN7", + "video_file": "Charades/videos/Z0KN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person sits on the floor watching a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.1>." + } + ] + }, + { + "video_id": "ME4D5", + "question_id": "ME4D5", + "video_file": "Charades/videos/ME4D5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they start sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.5> to <14.7>." + } + ] + }, + { + "video_id": "ME4D5", + "question_id": "ME4D5", + "video_file": "Charades/videos/ME4D5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person start sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.5> to <14.7>." + } + ] + }, + { + "video_id": "8ICEM", + "question_id": "8ICEM", + "video_file": "Charades/videos/8ICEM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting on a bed eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <14.2>." + } + ] + }, + { + "video_id": "8ICEM", + "question_id": "8ICEM", + "video_file": "Charades/videos/8ICEM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes out some clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.5> to <16.4>." + } + ] + }, + { + "video_id": "8ICEM", + "question_id": "8ICEM", + "video_file": "Charades/videos/8ICEM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.5> to <14.2>." + } + ] + }, + { + "video_id": "8ICEM", + "question_id": "8ICEM", + "video_file": "Charades/videos/8ICEM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person opens the closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.7> to <16.1>." + } + ] + }, + { + "video_id": "WJQPA", + "question_id": "WJQPA", + "video_file": "Charades/videos/WJQPA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.6> to <17.2>." + } + ] + }, + { + "video_id": "WJQPA", + "question_id": "WJQPA", + "video_file": "Charades/videos/WJQPA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the cabinet'." + }, + { + "from": "gpt", + "value": "From <25.6> to <34.0>." + } + ] + }, + { + "video_id": "538JV", + "question_id": "538JV", + "video_file": "Charades/videos/538JV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person take a drink out of a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.9> to <13.4>." + } + ] + }, + { + "video_id": "TIY7F", + "question_id": "TIY7F", + "video_file": "Charades/videos/TIY7F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinks water from a glass put the glass down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "O03Y4", + "question_id": "O03Y4", + "video_file": "Charades/videos/O03Y4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting it on the table'." + }, + { + "from": "gpt", + "value": "From <11.5> to <17.0>." + } + ] + }, + { + "video_id": "6B7HM", + "question_id": "6B7HM", + "video_file": "Charades/videos/6B7HM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person start smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.9> to <30.0>." + } + ] + }, + { + "video_id": "37GL4", + "question_id": "37GL4", + "video_file": "Charades/videos/37GL4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person a is fixing their hair in a mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.7> to <27.0>." + } + ] + }, + { + "video_id": "JQG43", + "question_id": "JQG43", + "video_file": "Charades/videos/JQG43.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens laptop'." + }, + { + "from": "gpt", + "value": "From <5.7> to <18.2>." + } + ] + }, + { + "video_id": "JQG43", + "question_id": "JQG43", + "video_file": "Charades/videos/JQG43.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person is opening the laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <18.2>." + } + ] + }, + { + "video_id": "1LSEQ", + "question_id": "1LSEQ", + "video_file": "Charades/videos/1LSEQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the glass on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.7> to <11.5>." + } + ] + }, + { + "video_id": "I5ZWF", + "question_id": "I5ZWF", + "video_file": "Charades/videos/I5ZWF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is walking into a garage holding a pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.4> to <23.5>." + } + ] + }, + { + "video_id": "IUXP8", + "question_id": "IUXP8", + "video_file": "Charades/videos/IUXP8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person start cooking on a stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.1> to <30.0>." + } + ] + }, + { + "video_id": "UGQD0", + "question_id": "UGQD0", + "video_file": "Charades/videos/UGQD0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting in a chair holding a phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.6>." + } + ] + }, + { + "video_id": "UGQD0", + "question_id": "UGQD0", + "video_file": "Charades/videos/UGQD0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is holding a phone'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.6>." + } + ] + }, + { + "video_id": "TWU2U", + "question_id": "TWU2U", + "video_file": "Charades/videos/TWU2U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person begins opening the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.1> to <17.9>." + } + ] + }, + { + "video_id": "TWU2U", + "question_id": "TWU2U", + "video_file": "Charades/videos/TWU2U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the refrigerator door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.1> to <17.9>." + } + ] + }, + { + "video_id": "DYYLD", + "question_id": "DYYLD", + "video_file": "Charades/videos/DYYLD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closing a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.5> to <15.2>." + } + ] + }, + { + "video_id": "DYYLD", + "question_id": "DYYLD", + "video_file": "Charades/videos/DYYLD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person the pull the door closed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <15.2>." + } + ] + }, + { + "video_id": "CF55N", + "question_id": "CF55N", + "video_file": "Charades/videos/CF55N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turning off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <30.4> to <34.6>." + } + ] + }, + { + "video_id": "CF55N", + "question_id": "CF55N", + "video_file": "Charades/videos/CF55N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person smiles into a mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.7> to <13.4>." + } + ] + }, + { + "video_id": "CF55N", + "question_id": "CF55N", + "video_file": "Charades/videos/CF55N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person smiling in front of a mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.7> to <13.4>." + } + ] + }, + { + "video_id": "CF55N", + "question_id": "CF55N", + "video_file": "Charades/videos/CF55N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person turns off the light'." + }, + { + "from": "gpt", + "value": "From <30.4> to <34.6>." + } + ] + }, + { + "video_id": "V29QF", + "question_id": "V29QF", + "video_file": "Charades/videos/V29QF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is holding a towel in the hallway'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "EFX3U", + "question_id": "EFX3U", + "video_file": "Charades/videos/EFX3U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person closes a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <16.4>." + } + ] + }, + { + "video_id": "EFX3U", + "question_id": "EFX3U", + "video_file": "Charades/videos/EFX3U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <32.0>." + } + ] + }, + { + "video_id": "EFX3U", + "question_id": "EFX3U", + "video_file": "Charades/videos/EFX3U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begin eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <17.9> to <32.0>." + } + ] + }, + { + "video_id": "X7GVM", + "question_id": "X7GVM", + "video_file": "Charades/videos/X7GVM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating food off a plate' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.1> to <33.0>." + } + ] + }, + { + "video_id": "Z5JZD", + "question_id": "Z5JZD", + "video_file": "Charades/videos/Z5JZD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the pantry door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.9> to <25.7>." + } + ] + }, + { + "video_id": "Z5JZD", + "question_id": "Z5JZD", + "video_file": "Charades/videos/Z5JZD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts them on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.3> to <19.8>." + } + ] + }, + { + "video_id": "Z5JZD", + "question_id": "Z5JZD", + "video_file": "Charades/videos/Z5JZD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door to the cabinet'." + }, + { + "from": "gpt", + "value": "From <19.9> to <25.7>." + } + ] + }, + { + "video_id": "Z5JZD", + "question_id": "Z5JZD", + "video_file": "Charades/videos/Z5JZD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is putting some groceries in a bag away' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <16.3>." + } + ] + }, + { + "video_id": "DMI87", + "question_id": "DMI87", + "video_file": "Charades/videos/DMI87.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'smiling person takes a towel'." + }, + { + "from": "gpt", + "value": "From <18.1> to <25.6>." + } + ] + }, + { + "video_id": "A42FN", + "question_id": "A42FN", + "video_file": "Charades/videos/A42FN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the broom away' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.2> to <13.6>." + } + ] + }, + { + "video_id": "A42FN", + "question_id": "A42FN", + "video_file": "Charades/videos/A42FN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is seen eating a chip'." + }, + { + "from": "gpt", + "value": "From <5.9> to <17.2>." + } + ] + }, + { + "video_id": "A42FN", + "question_id": "A42FN", + "video_file": "Charades/videos/A42FN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating something'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.9> to <17.2>." + } + ] + }, + { + "video_id": "A42FN", + "question_id": "A42FN", + "video_file": "Charades/videos/A42FN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the food on table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <9.0>." + } + ] + }, + { + "video_id": "XJE4Y", + "question_id": "XJE4Y", + "video_file": "Charades/videos/XJE4Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.4>." + } + ] + }, + { + "video_id": "XJE4Y", + "question_id": "XJE4Y", + "video_file": "Charades/videos/XJE4Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person undresses' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.4> to <14.7>." + } + ] + }, + { + "video_id": "XJE4Y", + "question_id": "XJE4Y", + "video_file": "Charades/videos/XJE4Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws their clothes into the washing machine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.6>." + } + ] + }, + { + "video_id": "XJE4Y", + "question_id": "XJE4Y", + "video_file": "Charades/videos/XJE4Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person gets dressed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <14.7>." + } + ] + }, + { + "video_id": "8AKGE", + "question_id": "8AKGE", + "video_file": "Charades/videos/8AKGE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a drink from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.6> to <25.6>." + } + ] + }, + { + "video_id": "8AKGE", + "question_id": "8AKGE", + "video_file": "Charades/videos/8AKGE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <16.7>." + } + ] + }, + { + "video_id": "8AKGE", + "question_id": "8AKGE", + "video_file": "Charades/videos/8AKGE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person turns on the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <16.7>." + } + ] + }, + { + "video_id": "8AKGE", + "question_id": "8AKGE", + "video_file": "Charades/videos/8AKGE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a drink of a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.6> to <25.6>." + } + ] + }, + { + "video_id": "KAQNT", + "question_id": "KAQNT", + "video_file": "Charades/videos/KAQNT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close a laptop that's on a counter'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <10.7>." + } + ] + }, + { + "video_id": "8WPPG", + "question_id": "8WPPG", + "video_file": "Charades/videos/8WPPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is seen throwing a towel down the stairs'." + }, + { + "from": "gpt", + "value": "From <17.9> to <25.7>." + } + ] + }, + { + "video_id": "8PL34", + "question_id": "8PL34", + "video_file": "Charades/videos/8PL34.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person closes a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <9.5>." + } + ] + }, + { + "video_id": "8PL34", + "question_id": "8PL34", + "video_file": "Charades/videos/8PL34.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.2> to <29.4>." + } + ] + }, + { + "video_id": "8PL34", + "question_id": "8PL34", + "video_file": "Charades/videos/8PL34.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens the door to their bedroom'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <29.4>." + } + ] + }, + { + "video_id": "8PL34", + "question_id": "8PL34", + "video_file": "Charades/videos/8PL34.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits on a bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <20.4>." + } + ] + }, + { + "video_id": "8PL34", + "question_id": "8PL34", + "video_file": "Charades/videos/8PL34.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits on the bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.7> to <20.4>." + } + ] + }, + { + "video_id": "8PL34", + "question_id": "8PL34", + "video_file": "Charades/videos/8PL34.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes their shoes off' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.0> to <20.7>." + } + ] + }, + { + "video_id": "8PL34", + "question_id": "8PL34", + "video_file": "Charades/videos/8PL34.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes off shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <20.7>." + } + ] + }, + { + "video_id": "8PL34", + "question_id": "8PL34", + "video_file": "Charades/videos/8PL34.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person take off their shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <20.7>." + } + ] + }, + { + "video_id": "8PL34", + "question_id": "8PL34", + "video_file": "Charades/videos/8PL34.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws the shoes on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.9> to <34.0>." + } + ] + }, + { + "video_id": "PNC99", + "question_id": "PNC99", + "video_file": "Charades/videos/PNC99.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking a glass of wine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <23.1>." + } + ] + }, + { + "video_id": "UTR56", + "question_id": "UTR56", + "video_file": "Charades/videos/UTR56.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person holds a doorknob' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.9> to <23.5>." + } + ] + }, + { + "video_id": "VLYM4", + "question_id": "VLYM4", + "video_file": "Charades/videos/VLYM4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.9> to <37.2>." + } + ] + }, + { + "video_id": "VLYM4", + "question_id": "VLYM4", + "video_file": "Charades/videos/VLYM4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sitting at a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.5> to <35.8>." + } + ] + }, + { + "video_id": "VLYM4", + "question_id": "VLYM4", + "video_file": "Charades/videos/VLYM4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating food in a kitchen' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.9> to <37.2>." + } + ] + }, + { + "video_id": "VLYM4", + "question_id": "VLYM4", + "video_file": "Charades/videos/VLYM4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.7> to <11.7>." + } + ] + }, + { + "video_id": "JC9SA", + "question_id": "JC9SA", + "video_file": "Charades/videos/JC9SA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.2> to <32.1>." + } + ] + }, + { + "video_id": "JC9SA", + "question_id": "JC9SA", + "video_file": "Charades/videos/JC9SA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens the door to their refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.2> to <32.1>." + } + ] + }, + { + "video_id": "X42YA", + "question_id": "X42YA", + "video_file": "Charades/videos/X42YA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person runs back in' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.0>." + } + ] + }, + { + "video_id": "X42YA", + "question_id": "X42YA", + "video_file": "Charades/videos/X42YA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person runs through'." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.0>." + } + ] + }, + { + "video_id": "R5CWJ", + "question_id": "R5CWJ", + "video_file": "Charades/videos/R5CWJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person another gets dressed in front of a mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.8> to <18.0>." + } + ] + }, + { + "video_id": "R5CWJ", + "question_id": "R5CWJ", + "video_file": "Charades/videos/R5CWJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person two laughing people hug' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.9> to <9.6>." + } + ] + }, + { + "video_id": "R5CWJ", + "question_id": "R5CWJ", + "video_file": "Charades/videos/R5CWJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person two people are laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.9> to <9.6>." + } + ] + }, + { + "video_id": "R5CWJ", + "question_id": "R5CWJ", + "video_file": "Charades/videos/R5CWJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person one starts dressing'." + }, + { + "from": "gpt", + "value": "From <13.8> to <18.0>." + } + ] + }, + { + "video_id": "38HNU", + "question_id": "38HNU", + "video_file": "Charades/videos/38HNU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person in the pantry is sneezing into the coffee' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <19.3>." + } + ] + }, + { + "video_id": "38HNU", + "question_id": "38HNU", + "video_file": "Charades/videos/38HNU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they laugh' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.0> to <22.4>." + } + ] + }, + { + "video_id": "JOWAU", + "question_id": "JOWAU", + "video_file": "Charades/videos/JOWAU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person awakens on a bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.4> to <12.8>." + } + ] + }, + { + "video_id": "JOWAU", + "question_id": "JOWAU", + "video_file": "Charades/videos/JOWAU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.8> to <19.8>." + } + ] + }, + { + "video_id": "INQNU", + "question_id": "INQNU", + "video_file": "Charades/videos/INQNU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person take a drink from a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.2> to <14.8>." + } + ] + }, + { + "video_id": "INQNU", + "question_id": "INQNU", + "video_file": "Charades/videos/INQNU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking a glass of milk' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.2> to <14.8>." + } + ] + }, + { + "video_id": "PZ2W1", + "question_id": "PZ2W1", + "video_file": "Charades/videos/PZ2W1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is drinking a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <9.2>." + } + ] + }, + { + "video_id": "PZ2W1", + "question_id": "PZ2W1", + "video_file": "Charades/videos/PZ2W1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking from a glass sits at the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <9.2>." + } + ] + }, + { + "video_id": "0Z36L", + "question_id": "0Z36L", + "video_file": "Charades/videos/0Z36L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person putting clothes in bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "0Z36L", + "question_id": "0Z36L", + "video_file": "Charades/videos/0Z36L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is putting clothing into a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.7> to <32.0>." + } + ] + }, + { + "video_id": "DIEEO", + "question_id": "DIEEO", + "video_file": "Charades/videos/DIEEO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is drinking from a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "6AHD0", + "question_id": "6AHD0", + "video_file": "Charades/videos/6AHD0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting on a sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <8.5>." + } + ] + }, + { + "video_id": "6AHD0", + "question_id": "6AHD0", + "video_file": "Charades/videos/6AHD0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is on the couch eating something'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <12.4>." + } + ] + }, + { + "video_id": "6AHD0", + "question_id": "6AHD0", + "video_file": "Charades/videos/6AHD0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person stand up' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.5> to <23.6>." + } + ] + }, + { + "video_id": "6AHD0", + "question_id": "6AHD0", + "video_file": "Charades/videos/6AHD0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is in a dining room eating food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <12.4>." + } + ] + }, + { + "video_id": "6AHD0", + "question_id": "6AHD0", + "video_file": "Charades/videos/6AHD0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating some food from a bowl'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <12.4>." + } + ] + }, + { + "video_id": "RVYXB", + "question_id": "RVYXB", + "video_file": "Charades/videos/RVYXB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person gets up to pour coffee in cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.4> to <26.2>." + } + ] + }, + { + "video_id": "RVYXB", + "question_id": "RVYXB", + "video_file": "Charades/videos/RVYXB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pour a cup of coffee' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.4> to <26.2>." + } + ] + }, + { + "video_id": "8SA46", + "question_id": "8SA46", + "video_file": "Charades/videos/8SA46.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is closing the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.9>." + } + ] + }, + { + "video_id": "0R6K4", + "question_id": "0R6K4", + "video_file": "Charades/videos/0R6K4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts the glass on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.4> to <23.9>." + } + ] + }, + { + "video_id": "F56YP", + "question_id": "F56YP", + "video_file": "Charades/videos/F56YP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'that same person closes the box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "61BEP", + "question_id": "61BEP", + "video_file": "Charades/videos/61BEP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person getting up to turn on the light'." + }, + { + "from": "gpt", + "value": "From <30.7> to <37.0>." + } + ] + }, + { + "video_id": "WOVOX", + "question_id": "WOVOX", + "video_file": "Charades/videos/WOVOX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person continues looking out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <34.8> to <58.0>." + } + ] + }, + { + "video_id": "WOVOX", + "question_id": "WOVOX", + "video_file": "Charades/videos/WOVOX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'as the person begins to undress' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.6> to <18.4>." + } + ] + }, + { + "video_id": "WOVOX", + "question_id": "WOVOX", + "video_file": "Charades/videos/WOVOX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person the laughing stops'." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.3>." + } + ] + }, + { + "video_id": "WOVOX", + "question_id": "WOVOX", + "video_file": "Charades/videos/WOVOX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person looks out a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <34.8> to <58.0>." + } + ] + }, + { + "video_id": "WOVOX", + "question_id": "WOVOX", + "video_file": "Charades/videos/WOVOX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks through a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "WOVOX", + "question_id": "WOVOX", + "video_file": "Charades/videos/WOVOX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person laughing heartily'." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.3>." + } + ] + }, + { + "video_id": "3HUXR", + "question_id": "3HUXR", + "video_file": "Charades/videos/3HUXR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turn off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <37.4> to <42.6>." + } + ] + }, + { + "video_id": "3HUXR", + "question_id": "3HUXR", + "video_file": "Charades/videos/3HUXR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <37.4> to <42.6>." + } + ] + }, + { + "video_id": "OIOIO", + "question_id": "OIOIO", + "video_file": "Charades/videos/OIOIO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person throwing clothes at a door'." + }, + { + "from": "gpt", + "value": "From <4.1> to <10.3>." + } + ] + }, + { + "video_id": "Q3FZ6", + "question_id": "Q3FZ6", + "video_file": "Charades/videos/Q3FZ6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting clothes in a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.9> to <18.9>." + } + ] + }, + { + "video_id": "5IVO1", + "question_id": "5IVO1", + "video_file": "Charades/videos/5IVO1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins to look through a book'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.6>." + } + ] + }, + { + "video_id": "5IVO1", + "question_id": "5IVO1", + "video_file": "Charades/videos/5IVO1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person looking at a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.6>." + } + ] + }, + { + "video_id": "5IVO1", + "question_id": "5IVO1", + "video_file": "Charades/videos/5IVO1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs from the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <27.6> to <32.0>." + } + ] + }, + { + "video_id": "0F1VF", + "question_id": "0F1VF", + "video_file": "Charades/videos/0F1VF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person walks to put a book on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "Y8M9P", + "question_id": "Y8M9P", + "video_file": "Charades/videos/Y8M9P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person is laughing on the phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <14.6>." + } + ] + }, + { + "video_id": "Y8M9P", + "question_id": "Y8M9P", + "video_file": "Charades/videos/Y8M9P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts laughing'." + }, + { + "from": "gpt", + "value": "From <7.0> to <14.6>." + } + ] + }, + { + "video_id": "H28KM", + "question_id": "H28KM", + "video_file": "Charades/videos/H28KM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person still holding the book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.6> to <26.2>." + } + ] + }, + { + "video_id": "H28KM", + "question_id": "H28KM", + "video_file": "Charades/videos/H28KM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a book in the entryway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.6> to <26.2>." + } + ] + }, + { + "video_id": "H28KM", + "question_id": "H28KM", + "video_file": "Charades/videos/H28KM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person holding a book picks up a mirror' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.6> to <26.2>." + } + ] + }, + { + "video_id": "H28KM", + "question_id": "H28KM", + "video_file": "Charades/videos/H28KM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person running in place'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <30.4>." + } + ] + }, + { + "video_id": "H28KM", + "question_id": "H28KM", + "video_file": "Charades/videos/H28KM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person begins to run in place'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <30.4>." + } + ] + }, + { + "video_id": "H28KM", + "question_id": "H28KM", + "video_file": "Charades/videos/H28KM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person run in place' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <30.4>." + } + ] + }, + { + "video_id": "H28KM", + "question_id": "H28KM", + "video_file": "Charades/videos/H28KM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.6> to <26.2>." + } + ] + }, + { + "video_id": "JTTAP", + "question_id": "JTTAP", + "video_file": "Charades/videos/JTTAP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person closes the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.4> to <23.1>." + } + ] + }, + { + "video_id": "JTTAP", + "question_id": "JTTAP", + "video_file": "Charades/videos/JTTAP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks into view drinking from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.6> to <11.0>." + } + ] + }, + { + "video_id": "JTTAP", + "question_id": "JTTAP", + "video_file": "Charades/videos/JTTAP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <20.0>." + } + ] + }, + { + "video_id": "JTTAP", + "question_id": "JTTAP", + "video_file": "Charades/videos/JTTAP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person the puts it away into the medicine shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <22.0>." + } + ] + }, + { + "video_id": "JTTAP", + "question_id": "JTTAP", + "video_file": "Charades/videos/JTTAP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person plays with their phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.9> to <18.0>." + } + ] + }, + { + "video_id": "JTTAP", + "question_id": "JTTAP", + "video_file": "Charades/videos/JTTAP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put it on the shelf'." + }, + { + "from": "gpt", + "value": "From <16.3> to <22.0>." + } + ] + }, + { + "video_id": "4021N", + "question_id": "4021N", + "video_file": "Charades/videos/4021N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put a camera into a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.0> to <28.7>." + } + ] + }, + { + "video_id": "MPFSX", + "question_id": "MPFSX", + "video_file": "Charades/videos/MPFSX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person walks through the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.8> to <31.3>." + } + ] + }, + { + "video_id": "WN7T3", + "question_id": "WN7T3", + "video_file": "Charades/videos/WN7T3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.5> to <20.4>." + } + ] + }, + { + "video_id": "X58SH", + "question_id": "X58SH", + "video_file": "Charades/videos/X58SH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting groceries away' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <26.6>." + } + ] + }, + { + "video_id": "X58SH", + "question_id": "X58SH", + "video_file": "Charades/videos/X58SH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they take a food dish' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.9> to <31.1>." + } + ] + }, + { + "video_id": "4021D", + "question_id": "4021D", + "video_file": "Charades/videos/4021D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.5> to <31.1>." + } + ] + }, + { + "video_id": "4021D", + "question_id": "4021D", + "video_file": "Charades/videos/4021D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.5> to <31.1>." + } + ] + }, + { + "video_id": "9C5PT", + "question_id": "9C5PT", + "video_file": "Charades/videos/9C5PT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.3> to <19.7>." + } + ] + }, + { + "video_id": "9C5PT", + "question_id": "9C5PT", + "video_file": "Charades/videos/9C5PT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes off their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.3> to <19.7>." + } + ] + }, + { + "video_id": "CVX3K", + "question_id": "CVX3K", + "video_file": "Charades/videos/CVX3K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person plays on their phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.4> to <34.0>." + } + ] + }, + { + "video_id": "CVX3K", + "question_id": "CVX3K", + "video_file": "Charades/videos/CVX3K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinks water from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <22.4>." + } + ] + }, + { + "video_id": "SNECW", + "question_id": "SNECW", + "video_file": "Charades/videos/SNECW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.0> to <27.7>." + } + ] + }, + { + "video_id": "SNECW", + "question_id": "SNECW", + "video_file": "Charades/videos/SNECW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakens in their bedroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.1> to <22.5>." + } + ] + }, + { + "video_id": "SNECW", + "question_id": "SNECW", + "video_file": "Charades/videos/SNECW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turn on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.0> to <27.7>." + } + ] + }, + { + "video_id": "SNECW", + "question_id": "SNECW", + "video_file": "Charades/videos/SNECW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opens a window'." + }, + { + "from": "gpt", + "value": "From <25.6> to <31.5>." + } + ] + }, + { + "video_id": "I3N3J", + "question_id": "I3N3J", + "video_file": "Charades/videos/I3N3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looking through the book'." + }, + { + "from": "gpt", + "value": "From <16.3> to <27.2>." + } + ] + }, + { + "video_id": "VNQJ0", + "question_id": "VNQJ0", + "video_file": "Charades/videos/VNQJ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person another is tidying the area with a broom'." + }, + { + "from": "gpt", + "value": "From <11.2> to <26.8>." + } + ] + }, + { + "video_id": "0RROG", + "question_id": "0RROG", + "video_file": "Charades/videos/0RROG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is fixing a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.6>." + } + ] + }, + { + "video_id": "MUO4G", + "question_id": "MUO4G", + "video_file": "Charades/videos/MUO4G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.7> to <12.8>." + } + ] + }, + { + "video_id": "MUO4G", + "question_id": "MUO4G", + "video_file": "Charades/videos/MUO4G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person is laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <19.0>." + } + ] + }, + { + "video_id": "G8SWD", + "question_id": "G8SWD", + "video_file": "Charades/videos/G8SWD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person runs out of the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.0> to <28.0>." + } + ] + }, + { + "video_id": "G8SWD", + "question_id": "G8SWD", + "video_file": "Charades/videos/G8SWD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person closes the refrigerator'." + }, + { + "from": "gpt", + "value": "From <9.6> to <14.4>." + } + ] + }, + { + "video_id": "7W1ZI", + "question_id": "7W1ZI", + "video_file": "Charades/videos/7W1ZI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts them back in the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.4> to <29.0>." + } + ] + }, + { + "video_id": "7W1ZI", + "question_id": "7W1ZI", + "video_file": "Charades/videos/7W1ZI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person puts the cards in a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.4> to <29.0>." + } + ] + }, + { + "video_id": "23QFL", + "question_id": "23QFL", + "video_file": "Charades/videos/23QFL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <12.7>." + } + ] + }, + { + "video_id": "EG0IU", + "question_id": "EG0IU", + "video_file": "Charades/videos/EG0IU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turning off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.8> to <16.1>." + } + ] + }, + { + "video_id": "YVOVB", + "question_id": "YVOVB", + "video_file": "Charades/videos/YVOVB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is smiling' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.1> to <17.8>." + } + ] + }, + { + "video_id": "DRGKV", + "question_id": "DRGKV", + "video_file": "Charades/videos/DRGKV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person hold a camera up towards the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.4> to <30.6>." + } + ] + }, + { + "video_id": "DRGKV", + "question_id": "DRGKV", + "video_file": "Charades/videos/DRGKV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person undressing by taking their jacket off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <14.3>." + } + ] + }, + { + "video_id": "PWR6D", + "question_id": "PWR6D", + "video_file": "Charades/videos/PWR6D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person is sneezing'." + }, + { + "from": "gpt", + "value": "From <21.1> to <32.0>." + } + ] + }, + { + "video_id": "UBYBQ", + "question_id": "UBYBQ", + "video_file": "Charades/videos/UBYBQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is holding a blanket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.9>." + } + ] + }, + { + "video_id": "UBYBQ", + "question_id": "UBYBQ", + "video_file": "Charades/videos/UBYBQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts the blanket in the washer' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.4> to <9.0>." + } + ] + }, + { + "video_id": "ETT15", + "question_id": "ETT15", + "video_file": "Charades/videos/ETT15.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they stand up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <12.4>." + } + ] + }, + { + "video_id": "ETT15", + "question_id": "ETT15", + "video_file": "Charades/videos/ETT15.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person who is seated closes a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <8.8>." + } + ] + }, + { + "video_id": "ETT15", + "question_id": "ETT15", + "video_file": "Charades/videos/ETT15.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person sits in a chair at a desk' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.3>." + } + ] + }, + { + "video_id": "3P38N", + "question_id": "3P38N", + "video_file": "Charades/videos/3P38N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.0> to <29.0>." + } + ] + }, + { + "video_id": "3P38N", + "question_id": "3P38N", + "video_file": "Charades/videos/3P38N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person laughs as they read a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <18.4>." + } + ] + }, + { + "video_id": "57QQ7", + "question_id": "57QQ7", + "video_file": "Charades/videos/57QQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pouring a cup of coffee in the dining room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.3> to <22.5>." + } + ] + }, + { + "video_id": "57QQ7", + "question_id": "57QQ7", + "video_file": "Charades/videos/57QQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they pour coffee into a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.3> to <22.5>." + } + ] + }, + { + "video_id": "IYCOH", + "question_id": "IYCOH", + "video_file": "Charades/videos/IYCOH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is watching television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <22.2>." + } + ] + }, + { + "video_id": "IYCOH", + "question_id": "IYCOH", + "video_file": "Charades/videos/IYCOH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person working on a laptop in a bedroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.1> to <32.0>." + } + ] + }, + { + "video_id": "20WZU", + "question_id": "20WZU", + "video_file": "Charades/videos/20WZU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <21.5>." + } + ] + }, + { + "video_id": "20WZU", + "question_id": "20WZU", + "video_file": "Charades/videos/20WZU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sneezing as they grasp a sandwich'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "20WZU", + "question_id": "20WZU", + "video_file": "Charades/videos/20WZU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they open the book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.2> to <21.5>." + } + ] + }, + { + "video_id": "20WZU", + "question_id": "20WZU", + "video_file": "Charades/videos/20WZU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they eat the sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "20WZU", + "question_id": "20WZU", + "video_file": "Charades/videos/20WZU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.2> to <21.5>." + } + ] + }, + { + "video_id": "IRF8X", + "question_id": "IRF8X", + "video_file": "Charades/videos/IRF8X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is typing on a laptop repeatedly sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.9> to <12.5>." + } + ] + }, + { + "video_id": "7WJUW", + "question_id": "7WJUW", + "video_file": "Charades/videos/7WJUW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes out a blanket' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.0> to <12.0>." + } + ] + }, + { + "video_id": "7WJUW", + "question_id": "7WJUW", + "video_file": "Charades/videos/7WJUW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a phone from a table'." + }, + { + "from": "gpt", + "value": "From <29.1> to <36.0>." + } + ] + }, + { + "video_id": "RJCNV", + "question_id": "RJCNV", + "video_file": "Charades/videos/RJCNV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a book on the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <17.9>." + } + ] + }, + { + "video_id": "KAHM2", + "question_id": "KAHM2", + "video_file": "Charades/videos/KAHM2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person washes their hands in the sink' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.3> to <18.7>." + } + ] + }, + { + "video_id": "9OP4S", + "question_id": "9OP4S", + "video_file": "Charades/videos/9OP4S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting on stairs looking at a picture' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <12.1>." + } + ] + }, + { + "video_id": "9OP4S", + "question_id": "9OP4S", + "video_file": "Charades/videos/9OP4S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looking at a picture' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <12.1>." + } + ] + }, + { + "video_id": "9OP4S", + "question_id": "9OP4S", + "video_file": "Charades/videos/9OP4S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put the picture down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.0> to <15.5>." + } + ] + }, + { + "video_id": "FPFFN", + "question_id": "FPFFN", + "video_file": "Charades/videos/FPFFN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns lights off' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.7> to <19.0>." + } + ] + }, + { + "video_id": "FPFFN", + "question_id": "FPFFN", + "video_file": "Charades/videos/FPFFN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person is walking across room holding a phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <9.6>." + } + ] + }, + { + "video_id": "FPFFN", + "question_id": "FPFFN", + "video_file": "Charades/videos/FPFFN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walks through doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.7> to <19.0>." + } + ] + }, + { + "video_id": "HL4V4", + "question_id": "HL4V4", + "video_file": "Charades/videos/HL4V4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'that same person puts a laptop on it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.6> to <21.9>." + } + ] + }, + { + "video_id": "HL4V4", + "question_id": "HL4V4", + "video_file": "Charades/videos/HL4V4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts a laptop on it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.6> to <21.9>." + } + ] + }, + { + "video_id": "JA3MN", + "question_id": "JA3MN", + "video_file": "Charades/videos/JA3MN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person runs into the bathroom holding a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.2> to <8.9>." + } + ] + }, + { + "video_id": "JA3MN", + "question_id": "JA3MN", + "video_file": "Charades/videos/JA3MN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person runs in the room'." + }, + { + "from": "gpt", + "value": "From <2.2> to <8.9>." + } + ] + }, + { + "video_id": "JA3MN", + "question_id": "JA3MN", + "video_file": "Charades/videos/JA3MN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person open the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.6> to <6.2>." + } + ] + }, + { + "video_id": "JA3MN", + "question_id": "JA3MN", + "video_file": "Charades/videos/JA3MN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.6> to <6.2>." + } + ] + }, + { + "video_id": "BKNO9", + "question_id": "BKNO9", + "video_file": "Charades/videos/BKNO9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs into the room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.5> to <7.1>." + } + ] + }, + { + "video_id": "BKNO9", + "question_id": "BKNO9", + "video_file": "Charades/videos/BKNO9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <7.1>." + } + ] + }, + { + "video_id": "BKNO9", + "question_id": "BKNO9", + "video_file": "Charades/videos/BKNO9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws a towel onto a pile of clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <8.1>." + } + ] + }, + { + "video_id": "DHR83", + "question_id": "DHR83", + "video_file": "Charades/videos/DHR83.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person walks in with homework sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.2> to <10.0>." + } + ] + }, + { + "video_id": "953I1", + "question_id": "953I1", + "video_file": "Charades/videos/953I1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pours pop in to a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.3> to <13.5>." + } + ] + }, + { + "video_id": "953I1", + "question_id": "953I1", + "video_file": "Charades/videos/953I1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes a glass off of a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.3> to <11.0>." + } + ] + }, + { + "video_id": "953I1", + "question_id": "953I1", + "video_file": "Charades/videos/953I1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the glass on the shelf'." + }, + { + "from": "gpt", + "value": "From <14.0> to <21.1>." + } + ] + }, + { + "video_id": "Z1GUW", + "question_id": "Z1GUW", + "video_file": "Charades/videos/Z1GUW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is putting clothes away' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.0> to <25.0>." + } + ] + }, + { + "video_id": "PB7PQ", + "question_id": "PB7PQ", + "video_file": "Charades/videos/PB7PQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is fixing their hair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.8> to <10.3>." + } + ] + }, + { + "video_id": "PB7PQ", + "question_id": "PB7PQ", + "video_file": "Charades/videos/PB7PQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the front door'." + }, + { + "from": "gpt", + "value": "From <14.7> to <21.3>." + } + ] + }, + { + "video_id": "PB7PQ", + "question_id": "PB7PQ", + "video_file": "Charades/videos/PB7PQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the door to let a dog in'." + }, + { + "from": "gpt", + "value": "From <14.7> to <21.3>." + } + ] + }, + { + "video_id": "WM336", + "question_id": "WM336", + "video_file": "Charades/videos/WM336.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts tidying up the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.1> to <25.7>." + } + ] + }, + { + "video_id": "WM336", + "question_id": "WM336", + "video_file": "Charades/videos/WM336.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begin to tidy up the table'." + }, + { + "from": "gpt", + "value": "From <16.1> to <25.7>." + } + ] + }, + { + "video_id": "LC8T4", + "question_id": "LC8T4", + "video_file": "Charades/videos/LC8T4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closed the door of the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.5> to <21.2>." + } + ] + }, + { + "video_id": "LC8T4", + "question_id": "LC8T4", + "video_file": "Charades/videos/LC8T4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person cooked something' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "LC8T4", + "question_id": "LC8T4", + "video_file": "Charades/videos/LC8T4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person cooking some thing on the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "1Y5H7", + "question_id": "1Y5H7", + "video_file": "Charades/videos/1Y5H7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting clothes in a laundry basket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <22.2>." + } + ] + }, + { + "video_id": "VYMZC", + "question_id": "VYMZC", + "video_file": "Charades/videos/VYMZC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person turns the light off as the person leaving' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <11.9>." + } + ] + }, + { + "video_id": "OVYAC", + "question_id": "OVYAC", + "video_file": "Charades/videos/OVYAC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is throwing clothes all over the room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <8.5>." + } + ] + }, + { + "video_id": "OVYAC", + "question_id": "OVYAC", + "video_file": "Charades/videos/OVYAC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person throw a bunch of clothes on the floor'." + }, + { + "from": "gpt", + "value": "From <1.3> to <8.5>." + } + ] + }, + { + "video_id": "OVYAC", + "question_id": "OVYAC", + "video_file": "Charades/videos/OVYAC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes some clothes out of a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <23.6>." + } + ] + }, + { + "video_id": "QJ11V", + "question_id": "QJ11V", + "video_file": "Charades/videos/QJ11V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the phone down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <18.9>." + } + ] + }, + { + "video_id": "QJ11V", + "question_id": "QJ11V", + "video_file": "Charades/videos/QJ11V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <21.2>." + } + ] + }, + { + "video_id": "AHLVF", + "question_id": "AHLVF", + "video_file": "Charades/videos/AHLVF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is putting on their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.8> to <22.3>." + } + ] + }, + { + "video_id": "AHLVF", + "question_id": "AHLVF", + "video_file": "Charades/videos/AHLVF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting on shoes in a bathroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.8> to <22.3>." + } + ] + }, + { + "video_id": "5KZLW", + "question_id": "5KZLW", + "video_file": "Charades/videos/5KZLW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting in a chair laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.6> to <15.4>." + } + ] + }, + { + "video_id": "5KZLW", + "question_id": "5KZLW", + "video_file": "Charades/videos/5KZLW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person one is sitting in a chair laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.6> to <15.4>." + } + ] + }, + { + "video_id": "NA5F9", + "question_id": "NA5F9", + "video_file": "Charades/videos/NA5F9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a drink from a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <24.0>." + } + ] + }, + { + "video_id": "MUI01", + "question_id": "MUI01", + "video_file": "Charades/videos/MUI01.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins dressing in front of it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <28.9>." + } + ] + }, + { + "video_id": "KONWE", + "question_id": "KONWE", + "video_file": "Charades/videos/KONWE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws the pillow on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <22.5>." + } + ] + }, + { + "video_id": "KONWE", + "question_id": "KONWE", + "video_file": "Charades/videos/KONWE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws the pillow onto the side table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.6> to <22.5>." + } + ] + }, + { + "video_id": "M0ENV", + "question_id": "M0ENV", + "video_file": "Charades/videos/M0ENV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws an item onto the floor'." + }, + { + "from": "gpt", + "value": "From <22.1> to <27.4>." + } + ] + }, + { + "video_id": "M0ENV", + "question_id": "M0ENV", + "video_file": "Charades/videos/M0ENV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person decides to throw coffee beans all over the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.1> to <27.4>." + } + ] + }, + { + "video_id": "1JH1W", + "question_id": "1JH1W", + "video_file": "Charades/videos/1JH1W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person looks at picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <16.6>." + } + ] + }, + { + "video_id": "1JH1W", + "question_id": "1JH1W", + "video_file": "Charades/videos/1JH1W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looked at a picture hanging on the wall'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <16.6>." + } + ] + }, + { + "video_id": "IKBYS", + "question_id": "IKBYS", + "video_file": "Charades/videos/IKBYS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person holds a medicine bottle'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <14.9>." + } + ] + }, + { + "video_id": "WMR4G", + "question_id": "WMR4G", + "video_file": "Charades/videos/WMR4G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is holding a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.5> to <16.1>." + } + ] + }, + { + "video_id": "WMR4G", + "question_id": "WMR4G", + "video_file": "Charades/videos/WMR4G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is putting a glass next to some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.0> to <17.1>." + } + ] + }, + { + "video_id": "WMR4G", + "question_id": "WMR4G", + "video_file": "Charades/videos/WMR4G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats some food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.4> to <34.0>." + } + ] + }, + { + "video_id": "WMR4G", + "question_id": "WMR4G", + "video_file": "Charades/videos/WMR4G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats food from a bowl'." + }, + { + "from": "gpt", + "value": "From <21.4> to <34.0>." + } + ] + }, + { + "video_id": "MXW4N", + "question_id": "MXW4N", + "video_file": "Charades/videos/MXW4N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person running into a close' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "MXW4N", + "question_id": "MXW4N", + "video_file": "Charades/videos/MXW4N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the other person is holding up a phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.0> to <18.6>." + } + ] + }, + { + "video_id": "MXW4N", + "question_id": "MXW4N", + "video_file": "Charades/videos/MXW4N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is running with a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "MXW4N", + "question_id": "MXW4N", + "video_file": "Charades/videos/MXW4N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs into another room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "JCNHL", + "question_id": "JCNHL", + "video_file": "Charades/videos/JCNHL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they start undressing by taking their jacket off' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.4> to <19.1>." + } + ] + }, + { + "video_id": "JCNHL", + "question_id": "JCNHL", + "video_file": "Charades/videos/JCNHL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person their smiling turns into laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.4> to <8.8>." + } + ] + }, + { + "video_id": "HWHLG", + "question_id": "HWHLG", + "video_file": "Charades/videos/HWHLG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens window curtains' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.7> to <12.8>." + } + ] + }, + { + "video_id": "HWHLG", + "question_id": "HWHLG", + "video_file": "Charades/videos/HWHLG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looked out the window'." + }, + { + "from": "gpt", + "value": "From <7.9> to <18.0>." + } + ] + }, + { + "video_id": "HWHLG", + "question_id": "HWHLG", + "video_file": "Charades/videos/HWHLG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person runs over' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.9> to <24.3>." + } + ] + }, + { + "video_id": "HWHLG", + "question_id": "HWHLG", + "video_file": "Charades/videos/HWHLG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person makes tidy clothes on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <11.6>." + } + ] + }, + { + "video_id": "S1LQF", + "question_id": "S1LQF", + "video_file": "Charades/videos/S1LQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a man pouring some water out of a bottle' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <14.6>." + } + ] + }, + { + "video_id": "S1LQF", + "question_id": "S1LQF", + "video_file": "Charades/videos/S1LQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pours water in glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.2> to <14.6>." + } + ] + }, + { + "video_id": "S1LQF", + "question_id": "S1LQF", + "video_file": "Charades/videos/S1LQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting into a glass'." + }, + { + "from": "gpt", + "value": "From <31.6> to <38.0>." + } + ] + }, + { + "video_id": "5KGCY", + "question_id": "5KGCY", + "video_file": "Charades/videos/5KGCY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take a glass off the table'." + }, + { + "from": "gpt", + "value": "From <9.8> to <15.6>." + } + ] + }, + { + "video_id": "BM5B8", + "question_id": "BM5B8", + "video_file": "Charades/videos/BM5B8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opened the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.4> to <23.8>." + } + ] + }, + { + "video_id": "BM5B8", + "question_id": "BM5B8", + "video_file": "Charades/videos/BM5B8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person walks away opens a door removes hat' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.4> to <23.8>." + } + ] + }, + { + "video_id": "282ZZ", + "question_id": "282ZZ", + "video_file": "Charades/videos/282ZZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.5> to <26.2>." + } + ] + }, + { + "video_id": "282ZZ", + "question_id": "282ZZ", + "video_file": "Charades/videos/282ZZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating a sandwich in the living room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.5> to <26.2>." + } + ] + }, + { + "video_id": "HAIVQ", + "question_id": "HAIVQ", + "video_file": "Charades/videos/HAIVQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin eating something' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.4> to <29.1>." + } + ] + }, + { + "video_id": "HAIVQ", + "question_id": "HAIVQ", + "video_file": "Charades/videos/HAIVQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is throwing clothes down the stairs'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <14.7>." + } + ] + }, + { + "video_id": "HAIVQ", + "question_id": "HAIVQ", + "video_file": "Charades/videos/HAIVQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is throwing clothes down some stairs' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.4> to <14.7>." + } + ] + }, + { + "video_id": "HAIVQ", + "question_id": "HAIVQ", + "video_file": "Charades/videos/HAIVQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating food on the stairs' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.4> to <29.1>." + } + ] + }, + { + "video_id": "HG8G1", + "question_id": "HG8G1", + "video_file": "Charades/videos/HG8G1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person starts smiling randomly'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.0> to <38.0>." + } + ] + }, + { + "video_id": "BLIW5", + "question_id": "BLIW5", + "video_file": "Charades/videos/BLIW5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is seen putting some clothes away'." + }, + { + "from": "gpt", + "value": "From <7.7> to <12.2>." + } + ] + }, + { + "video_id": "POXTH", + "question_id": "POXTH", + "video_file": "Charades/videos/POXTH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they stand up'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "LLLXJ", + "question_id": "LLLXJ", + "video_file": "Charades/videos/LLLXJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person is running with a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <10.3>." + } + ] + }, + { + "video_id": "G9IUG", + "question_id": "G9IUG", + "video_file": "Charades/videos/G9IUG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person runs out' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.6> to <32.0>." + } + ] + }, + { + "video_id": "BUBT0", + "question_id": "BUBT0", + "video_file": "Charades/videos/BUBT0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sits in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.8> to <20.8>." + } + ] + }, + { + "video_id": "BUBT0", + "question_id": "BUBT0", + "video_file": "Charades/videos/BUBT0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person sits in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.8> to <20.8>." + } + ] + }, + { + "video_id": "BUBT0", + "question_id": "BUBT0", + "video_file": "Charades/videos/BUBT0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person laughs at the closed box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <31.6>." + } + ] + }, + { + "video_id": "FDIJX", + "question_id": "FDIJX", + "video_file": "Charades/videos/FDIJX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person snuggling with a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <18.0>." + } + ] + }, + { + "video_id": "ZVV1Y", + "question_id": "ZVV1Y", + "video_file": "Charades/videos/ZVV1Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <22.3>." + } + ] + }, + { + "video_id": "ZVV1Y", + "question_id": "ZVV1Y", + "video_file": "Charades/videos/ZVV1Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the door'." + }, + { + "from": "gpt", + "value": "From <20.0> to <25.2>." + } + ] + }, + { + "video_id": "IULMX", + "question_id": "IULMX", + "video_file": "Charades/videos/IULMX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is lying down snuggling a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.5>." + } + ] + }, + { + "video_id": "IULMX", + "question_id": "IULMX", + "video_file": "Charades/videos/IULMX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person awakens' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.6> to <11.0>." + } + ] + }, + { + "video_id": "2GRGY", + "question_id": "2GRGY", + "video_file": "Charades/videos/2GRGY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opens a refrigerator'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "2GRGY", + "question_id": "2GRGY", + "video_file": "Charades/videos/2GRGY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pours it into a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.8> to <28.3>." + } + ] + }, + { + "video_id": "2GRGY", + "question_id": "2GRGY", + "video_file": "Charades/videos/2GRGY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens refrigerator grabs sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "2GRGY", + "question_id": "2GRGY", + "video_file": "Charades/videos/2GRGY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts pouring juice into a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.8> to <28.3>." + } + ] + }, + { + "video_id": "0M6IY", + "question_id": "0M6IY", + "video_file": "Charades/videos/0M6IY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person the puts the broom away'." + }, + { + "from": "gpt", + "value": "From <8.6> to <14.0>." + } + ] + }, + { + "video_id": "TQ9GQ", + "question_id": "TQ9GQ", + "video_file": "Charades/videos/TQ9GQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts some cups into the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.1> to <32.6>." + } + ] + }, + { + "video_id": "CWM1T", + "question_id": "CWM1T", + "video_file": "Charades/videos/CWM1T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking on the stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.3>." + } + ] + }, + { + "video_id": "CWM1T", + "question_id": "CWM1T", + "video_file": "Charades/videos/CWM1T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person grasping a knife is cooking in their kitchen' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.3>." + } + ] + }, + { + "video_id": "QT4W0", + "question_id": "QT4W0", + "video_file": "Charades/videos/QT4W0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts dressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.4> to <24.0>." + } + ] + }, + { + "video_id": "QT4W0", + "question_id": "QT4W0", + "video_file": "Charades/videos/QT4W0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs into their recreation room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.0> to <11.1>." + } + ] + }, + { + "video_id": "0W4F6", + "question_id": "0W4F6", + "video_file": "Charades/videos/0W4F6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens his laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <21.7>." + } + ] + }, + { + "video_id": "0W4F6", + "question_id": "0W4F6", + "video_file": "Charades/videos/0W4F6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens their laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <21.7>." + } + ] + }, + { + "video_id": "WM8QQ", + "question_id": "WM8QQ", + "video_file": "Charades/videos/WM8QQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes a bag of groceries into a kitchen' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.7> to <29.9>." + } + ] + }, + { + "video_id": "WM8QQ", + "question_id": "WM8QQ", + "video_file": "Charades/videos/WM8QQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting the bag on the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.4> to <13.8>." + } + ] + }, + { + "video_id": "WM8QQ", + "question_id": "WM8QQ", + "video_file": "Charades/videos/WM8QQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes some of the food out' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "WM8QQ", + "question_id": "WM8QQ", + "video_file": "Charades/videos/WM8QQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <13.8>." + } + ] + }, + { + "video_id": "OOEME", + "question_id": "OOEME", + "video_file": "Charades/videos/OOEME.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person is dressing by the wardrobe'." + }, + { + "from": "gpt", + "value": "From <3.3> to <16.7>." + } + ] + }, + { + "video_id": "OOEME", + "question_id": "OOEME", + "video_file": "Charades/videos/OOEME.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person gets dressed'." + }, + { + "from": "gpt", + "value": "From <3.3> to <16.7>." + } + ] + }, + { + "video_id": "1DGSW", + "question_id": "1DGSW", + "video_file": "Charades/videos/1DGSW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person gets dressed'." + }, + { + "from": "gpt", + "value": "From <16.7> to <31.8>." + } + ] + }, + { + "video_id": "CS01T", + "question_id": "CS01T", + "video_file": "Charades/videos/CS01T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits on their bed'." + }, + { + "from": "gpt", + "value": "From <22.7> to <35.0>." + } + ] + }, + { + "video_id": "CS01T", + "question_id": "CS01T", + "video_file": "Charades/videos/CS01T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they sit down on the bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.7> to <35.0>." + } + ] + }, + { + "video_id": "TPH12", + "question_id": "TPH12", + "video_file": "Charades/videos/TPH12.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a book a little' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.7> to <6.4>." + } + ] + }, + { + "video_id": "TPH12", + "question_id": "TPH12", + "video_file": "Charades/videos/TPH12.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a smiling person is in their kitchen reading a book'." + }, + { + "from": "gpt", + "value": "From <11.2> to <20.5>." + } + ] + }, + { + "video_id": "TPH12", + "question_id": "TPH12", + "video_file": "Charades/videos/TPH12.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person reads a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <15.0>." + } + ] + }, + { + "video_id": "485I1", + "question_id": "485I1", + "video_file": "Charades/videos/485I1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person sits in a chair'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "485I1", + "question_id": "485I1", + "video_file": "Charades/videos/485I1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is watching television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.3>." + } + ] + }, + { + "video_id": "485I1", + "question_id": "485I1", + "video_file": "Charades/videos/485I1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking from a glass on the stairs' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.4> to <31.0>." + } + ] + }, + { + "video_id": "BDWIX", + "question_id": "BDWIX", + "video_file": "Charades/videos/BDWIX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes off their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "BDWIX", + "question_id": "BDWIX", + "video_file": "Charades/videos/BDWIX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins washing clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.8> to <23.6>." + } + ] + }, + { + "video_id": "BDWIX", + "question_id": "BDWIX", + "video_file": "Charades/videos/BDWIX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they suddenly start violently sneezing'." + }, + { + "from": "gpt", + "value": "From <19.6> to <31.0>." + } + ] + }, + { + "video_id": "UFWKL", + "question_id": "UFWKL", + "video_file": "Charades/videos/UFWKL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sneezing on a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <7.2>." + } + ] + }, + { + "video_id": "MC25U", + "question_id": "MC25U", + "video_file": "Charades/videos/MC25U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person runs away' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.8> to <25.0>." + } + ] + }, + { + "video_id": "MC25U", + "question_id": "MC25U", + "video_file": "Charades/videos/MC25U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person open the box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.9> to <22.8>." + } + ] + }, + { + "video_id": "MC25U", + "question_id": "MC25U", + "video_file": "Charades/videos/MC25U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person run out of the garage' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.8> to <25.0>." + } + ] + }, + { + "video_id": "RP64Z", + "question_id": "RP64Z", + "video_file": "Charades/videos/RP64Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person is laughing as they clean the stairs'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "RP64Z", + "question_id": "RP64Z", + "video_file": "Charades/videos/RP64Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walks away through a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <20.2>." + } + ] + }, + { + "video_id": "2MR5K", + "question_id": "2MR5K", + "video_file": "Charades/videos/2MR5K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is working on their laptop they'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.7>." + } + ] + }, + { + "video_id": "2MR5K", + "question_id": "2MR5K", + "video_file": "Charades/videos/2MR5K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is working on a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.7>." + } + ] + }, + { + "video_id": "QCAUL", + "question_id": "QCAUL", + "video_file": "Charades/videos/QCAUL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts their shoes on a chair'." + }, + { + "from": "gpt", + "value": "From <8.6> to <19.2>." + } + ] + }, + { + "video_id": "QCAUL", + "question_id": "QCAUL", + "video_file": "Charades/videos/QCAUL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person puts food on the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.6> to <7.8>." + } + ] + }, + { + "video_id": "QCAUL", + "question_id": "QCAUL", + "video_file": "Charades/videos/QCAUL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes off their shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.4> to <23.0>." + } + ] + }, + { + "video_id": "QCAUL", + "question_id": "QCAUL", + "video_file": "Charades/videos/QCAUL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person a plate than takes their shoes off' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.4> to <23.0>." + } + ] + }, + { + "video_id": "RF00I", + "question_id": "RF00I", + "video_file": "Charades/videos/RF00I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person talking on the phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.4> to <42.8>." + } + ] + }, + { + "video_id": "5JSGK", + "question_id": "5JSGK", + "video_file": "Charades/videos/5JSGK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.8> to <9.7>." + } + ] + }, + { + "video_id": "5JSGK", + "question_id": "5JSGK", + "video_file": "Charades/videos/5JSGK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turning off the light behind them' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.9> to <28.2>." + } + ] + }, + { + "video_id": "Z9XLJ", + "question_id": "Z9XLJ", + "video_file": "Charades/videos/Z9XLJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is walking in a hallway undressing their clothes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <18.4>." + } + ] + }, + { + "video_id": "Z9XLJ", + "question_id": "Z9XLJ", + "video_file": "Charades/videos/Z9XLJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'closes the door.the person turns' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.7> to <21.5>." + } + ] + }, + { + "video_id": "Z9XLJ", + "question_id": "Z9XLJ", + "video_file": "Charades/videos/Z9XLJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'takes off their jacket .the person opens a closet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.9> to <18.8>." + } + ] + }, + { + "video_id": "QTRIN", + "question_id": "QTRIN", + "video_file": "Charades/videos/QTRIN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding a pillow opens the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.1> to <32.0>." + } + ] + }, + { + "video_id": "QTRIN", + "question_id": "QTRIN", + "video_file": "Charades/videos/QTRIN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opened a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.1> to <32.0>." + } + ] + }, + { + "video_id": "9LHP3", + "question_id": "9LHP3", + "video_file": "Charades/videos/9LHP3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits on a sofa'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <12.3>." + } + ] + }, + { + "video_id": "9LHP3", + "question_id": "9LHP3", + "video_file": "Charades/videos/9LHP3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'shoes other person talking on phone'." + }, + { + "from": "gpt", + "value": "From <24.7> to <32.2>." + } + ] + }, + { + "video_id": "2MJA1", + "question_id": "2MJA1", + "video_file": "Charades/videos/2MJA1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they decide to start undressing by removing their sweater' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.0> to <24.2>." + } + ] + }, + { + "video_id": "6MH42", + "question_id": "6MH42", + "video_file": "Charades/videos/6MH42.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person begins laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.7> to <21.9>." + } + ] + }, + { + "video_id": "6MH42", + "question_id": "6MH42", + "video_file": "Charades/videos/6MH42.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.3>." + } + ] + }, + { + "video_id": "6MH42", + "question_id": "6MH42", + "video_file": "Charades/videos/6MH42.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is snuggling a blanket'." + }, + { + "from": "gpt", + "value": "From <13.2> to <21.9>." + } + ] + }, + { + "video_id": "6MH42", + "question_id": "6MH42", + "video_file": "Charades/videos/6MH42.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinks a glass of orange juice' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.3>." + } + ] + }, + { + "video_id": "ZS9XR", + "question_id": "ZS9XR", + "video_file": "Charades/videos/ZS9XR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats a snack' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <17.5>." + } + ] + }, + { + "video_id": "ZS9XR", + "question_id": "ZS9XR", + "video_file": "Charades/videos/ZS9XR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating food out of a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <17.5>." + } + ] + }, + { + "video_id": "ZS9XR", + "question_id": "ZS9XR", + "video_file": "Charades/videos/ZS9XR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts to eat food from a bag'." + }, + { + "from": "gpt", + "value": "From <2.4> to <17.5>." + } + ] + }, + { + "video_id": "ZS9XR", + "question_id": "ZS9XR", + "video_file": "Charades/videos/ZS9XR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits on a sofa'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <18.8>." + } + ] + }, + { + "video_id": "ZS9XR", + "question_id": "ZS9XR", + "video_file": "Charades/videos/ZS9XR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person sits on a couch' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <18.8>." + } + ] + }, + { + "video_id": "UFQK1", + "question_id": "UFQK1", + "video_file": "Charades/videos/UFQK1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they close the cabinet door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.7> to <28.0>." + } + ] + }, + { + "video_id": "UFQK1", + "question_id": "UFQK1", + "video_file": "Charades/videos/UFQK1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person cooks some food in a pot' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.5>." + } + ] + }, + { + "video_id": "UFQK1", + "question_id": "UFQK1", + "video_file": "Charades/videos/UFQK1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person stands at a stove cooking' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.5>." + } + ] + }, + { + "video_id": "UFQK1", + "question_id": "UFQK1", + "video_file": "Charades/videos/UFQK1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking at a stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.5>." + } + ] + }, + { + "video_id": "UFQK1", + "question_id": "UFQK1", + "video_file": "Charades/videos/UFQK1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the dish on a shelf in a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <26.9>." + } + ] + }, + { + "video_id": "MF09L", + "question_id": "MF09L", + "video_file": "Charades/videos/MF09L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "MF09L", + "question_id": "MF09L", + "video_file": "Charades/videos/MF09L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pours water in a cup'." + }, + { + "from": "gpt", + "value": "From <27.8> to <33.9>." + } + ] + }, + { + "video_id": "MF09L", + "question_id": "MF09L", + "video_file": "Charades/videos/MF09L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pour water in a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.8> to <33.9>." + } + ] + }, + { + "video_id": "MF09L", + "question_id": "MF09L", + "video_file": "Charades/videos/MF09L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take a picture of it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <39.4> to <52.0>." + } + ] + }, + { + "video_id": "MF09L", + "question_id": "MF09L", + "video_file": "Charades/videos/MF09L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a picture of the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <39.4> to <52.0>." + } + ] + }, + { + "video_id": "MF09L", + "question_id": "MF09L", + "video_file": "Charades/videos/MF09L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "0D5I8", + "question_id": "0D5I8", + "video_file": "Charades/videos/0D5I8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <32.0>." + } + ] + }, + { + "video_id": "QS0AZ", + "question_id": "QS0AZ", + "video_file": "Charades/videos/QS0AZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they open the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.7> to <14.3>." + } + ] + }, + { + "video_id": "CXKFB", + "question_id": "CXKFB", + "video_file": "Charades/videos/CXKFB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they start sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.1> to <19.1>." + } + ] + }, + { + "video_id": "CXKFB", + "question_id": "CXKFB", + "video_file": "Charades/videos/CXKFB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'causes the person to start sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.1> to <19.1>." + } + ] + }, + { + "video_id": "41GXC", + "question_id": "41GXC", + "video_file": "Charades/videos/41GXC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'that person puts the glass down on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <12.3>." + } + ] + }, + { + "video_id": "41GXC", + "question_id": "41GXC", + "video_file": "Charades/videos/41GXC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person sits at the table drinking from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.2>." + } + ] + }, + { + "video_id": "41GXC", + "question_id": "41GXC", + "video_file": "Charades/videos/41GXC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person at the refrigerator they open the door'." + }, + { + "from": "gpt", + "value": "From <19.7> to <27.7>." + } + ] + }, + { + "video_id": "7DU77", + "question_id": "7DU77", + "video_file": "Charades/videos/7DU77.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes a glass of water from the table'." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.6>." + } + ] + }, + { + "video_id": "7DU77", + "question_id": "7DU77", + "video_file": "Charades/videos/7DU77.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person closes a door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "7DU77", + "question_id": "7DU77", + "video_file": "Charades/videos/7DU77.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a bottle off a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.6>." + } + ] + }, + { + "video_id": "7DU77", + "question_id": "7DU77", + "video_file": "Charades/videos/7DU77.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinks from the bottle' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.3> to <21.3>." + } + ] + }, + { + "video_id": "4ZSAR", + "question_id": "4ZSAR", + "video_file": "Charades/videos/4ZSAR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sit on a chair with a coffee cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <35.2> to <39.0>." + } + ] + }, + { + "video_id": "4ZSAR", + "question_id": "4ZSAR", + "video_file": "Charades/videos/4ZSAR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sit in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <35.2> to <39.0>." + } + ] + }, + { + "video_id": "CF9QP", + "question_id": "CF9QP", + "video_file": "Charades/videos/CF9QP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins to eat from the bag of groceries' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.4> to <25.6>." + } + ] + }, + { + "video_id": "CF9QP", + "question_id": "CF9QP", + "video_file": "Charades/videos/CF9QP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person open up a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.6> to <20.9>." + } + ] + }, + { + "video_id": "CF9QP", + "question_id": "CF9QP", + "video_file": "Charades/videos/CF9QP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.4> to <25.6>." + } + ] + }, + { + "video_id": "CF9QP", + "question_id": "CF9QP", + "video_file": "Charades/videos/CF9QP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person start eating' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.4> to <25.6>." + } + ] + }, + { + "video_id": "CF9QP", + "question_id": "CF9QP", + "video_file": "Charades/videos/CF9QP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take some food out' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <25.7>." + } + ] + }, + { + "video_id": "XE4IA", + "question_id": "XE4IA", + "video_file": "Charades/videos/XE4IA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <29.7>." + } + ] + }, + { + "video_id": "XE4IA", + "question_id": "XE4IA", + "video_file": "Charades/videos/XE4IA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person close their book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <29.7>." + } + ] + }, + { + "video_id": "OCSJP", + "question_id": "OCSJP", + "video_file": "Charades/videos/OCSJP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person standing at the stove cooking' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.7> to <20.9>." + } + ] + }, + { + "video_id": "OCSJP", + "question_id": "OCSJP", + "video_file": "Charades/videos/OCSJP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is cooking on the stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <20.9>." + } + ] + }, + { + "video_id": "JHTO5", + "question_id": "JHTO5", + "video_file": "Charades/videos/JHTO5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person proceeds to open a medicine cabinet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.9>." + } + ] + }, + { + "video_id": "JHTO5", + "question_id": "JHTO5", + "video_file": "Charades/videos/JHTO5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they open a cabinet'." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.9>." + } + ] + }, + { + "video_id": "JHTO5", + "question_id": "JHTO5", + "video_file": "Charades/videos/JHTO5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the medicine cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.9>." + } + ] + }, + { + "video_id": "YSABV", + "question_id": "YSABV", + "video_file": "Charades/videos/YSABV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person the turned the light off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.2> to <18.0>." + } + ] + }, + { + "video_id": "YSABV", + "question_id": "YSABV", + "video_file": "Charades/videos/YSABV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person puts a pile of clothes into the corner' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.2> to <10.1>." + } + ] + }, + { + "video_id": "YSABV", + "question_id": "YSABV", + "video_file": "Charades/videos/YSABV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person laughs as they leave' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.4> to <20.3>." + } + ] + }, + { + "video_id": "YSABV", + "question_id": "YSABV", + "video_file": "Charades/videos/YSABV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turning off the light behind them' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.2> to <18.0>." + } + ] + }, + { + "video_id": "X3ASU", + "question_id": "X3ASU", + "video_file": "Charades/videos/X3ASU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes a selfie with a phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.1> to <19.9>." + } + ] + }, + { + "video_id": "X3ASU", + "question_id": "X3ASU", + "video_file": "Charades/videos/X3ASU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <5.9>." + } + ] + }, + { + "video_id": "X3ASU", + "question_id": "X3ASU", + "video_file": "Charades/videos/X3ASU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting in a chair reading a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <5.9>." + } + ] + }, + { + "video_id": "X3ASU", + "question_id": "X3ASU", + "video_file": "Charades/videos/X3ASU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the book down'." + }, + { + "from": "gpt", + "value": "From <28.6> to <33.3>." + } + ] + }, + { + "video_id": "X3ASU", + "question_id": "X3ASU", + "video_file": "Charades/videos/X3ASU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person close the book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <17.9>." + } + ] + }, + { + "video_id": "GE4SD", + "question_id": "GE4SD", + "video_file": "Charades/videos/GE4SD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a sip,before putting the glass on the counter'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.8> to <37.8>." + } + ] + }, + { + "video_id": "GE4SD", + "question_id": "GE4SD", + "video_file": "Charades/videos/GE4SD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is in the kitchen washing dishes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <27.1>." + } + ] + }, + { + "video_id": "GE4SD", + "question_id": "GE4SD", + "video_file": "Charades/videos/GE4SD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts washing dishes in the sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <27.1>." + } + ] + }, + { + "video_id": "GE4SD", + "question_id": "GE4SD", + "video_file": "Charades/videos/GE4SD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <37.9> to <50.0>." + } + ] + }, + { + "video_id": "3WD4E", + "question_id": "3WD4E", + "video_file": "Charades/videos/3WD4E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts a laptop in a bag'." + }, + { + "from": "gpt", + "value": "From <30.5> to <36.0>." + } + ] + }, + { + "video_id": "8JR94", + "question_id": "8JR94", + "video_file": "Charades/videos/8JR94.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eat the sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.4> to <33.0>." + } + ] + }, + { + "video_id": "LVLVR", + "question_id": "LVLVR", + "video_file": "Charades/videos/LVLVR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.9> to <25.2>." + } + ] + }, + { + "video_id": "LVLVR", + "question_id": "LVLVR", + "video_file": "Charades/videos/LVLVR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they open the bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <10.5>." + } + ] + }, + { + "video_id": "LVLVR", + "question_id": "LVLVR", + "video_file": "Charades/videos/LVLVR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begin eating' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.9> to <25.2>." + } + ] + }, + { + "video_id": "FKFU3", + "question_id": "FKFU3", + "video_file": "Charades/videos/FKFU3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.0>." + } + ] + }, + { + "video_id": "FKFU3", + "question_id": "FKFU3", + "video_file": "Charades/videos/FKFU3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they are smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "FKFU3", + "question_id": "FKFU3", + "video_file": "Charades/videos/FKFU3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person smiling out the window'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "7WIKW", + "question_id": "7WIKW", + "video_file": "Charades/videos/7WIKW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pours a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.7> to <11.1>." + } + ] + }, + { + "video_id": "7WIKW", + "question_id": "7WIKW", + "video_file": "Charades/videos/7WIKW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pours coffee into a glass'." + }, + { + "from": "gpt", + "value": "From <2.7> to <11.1>." + } + ] + }, + { + "video_id": "R1RDT", + "question_id": "R1RDT", + "video_file": "Charades/videos/R1RDT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person close the door behind them' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.7> to <26.1>." + } + ] + }, + { + "video_id": "R1RDT", + "question_id": "R1RDT", + "video_file": "Charades/videos/R1RDT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person awakens in bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "R1RDT", + "question_id": "R1RDT", + "video_file": "Charades/videos/R1RDT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person leave the bedroom closing the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.7> to <26.1>." + } + ] + }, + { + "video_id": "WTMMC", + "question_id": "WTMMC", + "video_file": "Charades/videos/WTMMC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they take a drink from a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <39.7> to <50.0>." + } + ] + }, + { + "video_id": "WTMMC", + "question_id": "WTMMC", + "video_file": "Charades/videos/WTMMC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throwing clothes in washer' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.1> to <42.5>." + } + ] + }, + { + "video_id": "ZA7ST", + "question_id": "ZA7ST", + "video_file": "Charades/videos/ZA7ST.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <16.7>." + } + ] + }, + { + "video_id": "XWBF1", + "question_id": "XWBF1", + "video_file": "Charades/videos/XWBF1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person walks into the room talking on their phone'." + }, + { + "from": "gpt", + "value": "From <3.6> to <14.3>." + } + ] + }, + { + "video_id": "XWBF1", + "question_id": "XWBF1", + "video_file": "Charades/videos/XWBF1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person talking on the phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.6> to <14.3>." + } + ] + }, + { + "video_id": "M0RUS", + "question_id": "M0RUS", + "video_file": "Charades/videos/M0RUS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person walks through the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <25.5>." + } + ] + }, + { + "video_id": "R9382", + "question_id": "R9382", + "video_file": "Charades/videos/R9382.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is cooking on the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "R9382", + "question_id": "R9382", + "video_file": "Charades/videos/R9382.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pulls something out of the fridge that he eats'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.8> to <29.0>." + } + ] + }, + { + "video_id": "R9382", + "question_id": "R9382", + "video_file": "Charades/videos/R9382.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person close the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <21.1>." + } + ] + }, + { + "video_id": "R9382", + "question_id": "R9382", + "video_file": "Charades/videos/R9382.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eat it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.8> to <29.0>." + } + ] + }, + { + "video_id": "R9382", + "question_id": "R9382", + "video_file": "Charades/videos/R9382.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take out a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <21.9>." + } + ] + }, + { + "video_id": "R9382", + "question_id": "R9382", + "video_file": "Charades/videos/R9382.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking at the stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "5M9JX", + "question_id": "5M9JX", + "video_file": "Charades/videos/5M9JX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person isdrinking a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "5M9JX", + "question_id": "5M9JX", + "video_file": "Charades/videos/5M9JX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person is drinking water from a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "5M9JX", + "question_id": "5M9JX", + "video_file": "Charades/videos/5M9JX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking a soda from a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "5M9JX", + "question_id": "5M9JX", + "video_file": "Charades/videos/5M9JX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is standing on their stairs eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <8.8>." + } + ] + }, + { + "video_id": "5M9JX", + "question_id": "5M9JX", + "video_file": "Charades/videos/5M9JX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they are eating the food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <9.0>." + } + ] + }, + { + "video_id": "I30LP", + "question_id": "I30LP", + "video_file": "Charades/videos/I30LP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person laughing at the television'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.9>." + } + ] + }, + { + "video_id": "1US0H", + "question_id": "1US0H", + "video_file": "Charades/videos/1US0H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sneezing into a wardrobe' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.8> to <15.4>." + } + ] + }, + { + "video_id": "1US0H", + "question_id": "1US0H", + "video_file": "Charades/videos/1US0H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person fixes hair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.5> to <28.1>." + } + ] + }, + { + "video_id": "1US0H", + "question_id": "1US0H", + "video_file": "Charades/videos/1US0H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sneezing into a towel' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.8> to <15.4>." + } + ] + }, + { + "video_id": "8PRDP", + "question_id": "8PRDP", + "video_file": "Charades/videos/8PRDP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinks a beverage from a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <28.7>." + } + ] + }, + { + "video_id": "8PRDP", + "question_id": "8PRDP", + "video_file": "Charades/videos/8PRDP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking some water from a glass'." + }, + { + "from": "gpt", + "value": "From <20.1> to <28.7>." + } + ] + }, + { + "video_id": "2GF6R", + "question_id": "2GF6R", + "video_file": "Charades/videos/2GF6R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating a sandwich in a living room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.3> to <16.2>." + } + ] + }, + { + "video_id": "2GF6R", + "question_id": "2GF6R", + "video_file": "Charades/videos/2GF6R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take a bite of a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "2GF6R", + "question_id": "2GF6R", + "video_file": "Charades/videos/2GF6R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is awakening'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <13.2>." + } + ] + }, + { + "video_id": "2GQDN", + "question_id": "2GQDN", + "video_file": "Charades/videos/2GQDN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats a sandwich'." + }, + { + "from": "gpt", + "value": "From <21.2> to <31.0>." + } + ] + }, + { + "video_id": "2GQDN", + "question_id": "2GQDN", + "video_file": "Charades/videos/2GQDN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.1> to <15.4>." + } + ] + }, + { + "video_id": "2GQDN", + "question_id": "2GQDN", + "video_file": "Charades/videos/2GQDN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts eating them' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.2> to <31.0>." + } + ] + }, + { + "video_id": "2GQDN", + "question_id": "2GQDN", + "video_file": "Charades/videos/2GQDN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens the closet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "2GQDN", + "question_id": "2GQDN", + "video_file": "Charades/videos/2GQDN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins eating the food'." + }, + { + "from": "gpt", + "value": "From <21.2> to <31.0>." + } + ] + }, + { + "video_id": "5UFCJ", + "question_id": "5UFCJ", + "video_file": "Charades/videos/5UFCJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person looking out a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.4> to <30.0>." + } + ] + }, + { + "video_id": "5UFCJ", + "question_id": "5UFCJ", + "video_file": "Charades/videos/5UFCJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours medicine into a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <16.0>." + } + ] + }, + { + "video_id": "IUOC0", + "question_id": "IUOC0", + "video_file": "Charades/videos/IUOC0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs into the room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.1> to <15.0>." + } + ] + }, + { + "video_id": "IUOC0", + "question_id": "IUOC0", + "video_file": "Charades/videos/IUOC0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <11.1> to <16.7>." + } + ] + }, + { + "video_id": "IUOC0", + "question_id": "IUOC0", + "video_file": "Charades/videos/IUOC0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <29.0>." + } + ] + }, + { + "video_id": "IUOC0", + "question_id": "IUOC0", + "video_file": "Charades/videos/IUOC0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the door to the office'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <16.7>." + } + ] + }, + { + "video_id": "IUOC0", + "question_id": "IUOC0", + "video_file": "Charades/videos/IUOC0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits on the chair with a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <29.0>." + } + ] + }, + { + "video_id": "IUOC0", + "question_id": "IUOC0", + "video_file": "Charades/videos/IUOC0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sitting in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.0> to <29.0>." + } + ] + }, + { + "video_id": "IUOC0", + "question_id": "IUOC0", + "video_file": "Charades/videos/IUOC0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs into the middle of an office' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.1> to <15.0>." + } + ] + }, + { + "video_id": "8BSG7", + "question_id": "8BSG7", + "video_file": "Charades/videos/8BSG7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person working on a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.6>." + } + ] + }, + { + "video_id": "8BSG7", + "question_id": "8BSG7", + "video_file": "Charades/videos/8BSG7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open a door'." + }, + { + "from": "gpt", + "value": "From <16.7> to <21.6>." + } + ] + }, + { + "video_id": "Q3XBQ", + "question_id": "Q3XBQ", + "video_file": "Charades/videos/Q3XBQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting at a table doing their homework'." + }, + { + "from": "gpt", + "value": "From <1.2> to <12.6>." + } + ] + }, + { + "video_id": "Q3XBQ", + "question_id": "Q3XBQ", + "video_file": "Charades/videos/Q3XBQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pours themselves a glass of water to drink' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.1> to <11.7>." + } + ] + }, + { + "video_id": "Q3XBQ", + "question_id": "Q3XBQ", + "video_file": "Charades/videos/Q3XBQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person throws a blanket on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.9> to <30.0>." + } + ] + }, + { + "video_id": "Q3XBQ", + "question_id": "Q3XBQ", + "video_file": "Charades/videos/Q3XBQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <16.9>." + } + ] + }, + { + "video_id": "L5ZAB", + "question_id": "L5ZAB", + "video_file": "Charades/videos/L5ZAB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sneezing into a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <6.9>." + } + ] + }, + { + "video_id": "WA3CG", + "question_id": "WA3CG", + "video_file": "Charades/videos/WA3CG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.4> to <13.4>." + } + ] + }, + { + "video_id": "1SCZE", + "question_id": "1SCZE", + "video_file": "Charades/videos/1SCZE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open the other door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <17.0>." + } + ] + }, + { + "video_id": "1SCZE", + "question_id": "1SCZE", + "video_file": "Charades/videos/1SCZE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they turn on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <11.0>." + } + ] + }, + { + "video_id": "1SCZE", + "question_id": "1SCZE", + "video_file": "Charades/videos/1SCZE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is putting a bag in the wardrobe' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.0> to <12.0>." + } + ] + }, + { + "video_id": "UYJVE", + "question_id": "UYJVE", + "video_file": "Charades/videos/UYJVE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins sneezing'." + }, + { + "from": "gpt", + "value": "From <9.6> to <16.7>." + } + ] + }, + { + "video_id": "HZTB1", + "question_id": "HZTB1", + "video_file": "Charades/videos/HZTB1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person is eating food out of the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <27.2>." + } + ] + }, + { + "video_id": "HZTB1", + "question_id": "HZTB1", + "video_file": "Charades/videos/HZTB1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes out a baggy' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <9.1>." + } + ] + }, + { + "video_id": "HZTB1", + "question_id": "HZTB1", + "video_file": "Charades/videos/HZTB1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats something'." + }, + { + "from": "gpt", + "value": "From <18.7> to <27.2>." + } + ] + }, + { + "video_id": "1Z5FK", + "question_id": "1Z5FK", + "video_file": "Charades/videos/1Z5FK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they start sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.6> to <13.6>." + } + ] + }, + { + "video_id": "H0O39", + "question_id": "H0O39", + "video_file": "Charades/videos/H0O39.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person running down stairs' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.2> to <40.0>." + } + ] + }, + { + "video_id": "H0O39", + "question_id": "H0O39", + "video_file": "Charades/videos/H0O39.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is awakening from beneath a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.0>." + } + ] + }, + { + "video_id": "UZIQS", + "question_id": "UZIQS", + "video_file": "Charades/videos/UZIQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts it on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.9> to <31.0>." + } + ] + }, + { + "video_id": "UZIQS", + "question_id": "UZIQS", + "video_file": "Charades/videos/UZIQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes a towel from a rack'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.4> to <21.2>." + } + ] + }, + { + "video_id": "UZIQS", + "question_id": "UZIQS", + "video_file": "Charades/videos/UZIQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes a towel from a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.4> to <21.2>." + } + ] + }, + { + "video_id": "RDUOV", + "question_id": "RDUOV", + "video_file": "Charades/videos/RDUOV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is putting a broom in a closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.3> to <15.9>." + } + ] + }, + { + "video_id": "OBXY5", + "question_id": "OBXY5", + "video_file": "Charades/videos/OBXY5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person washes their hands' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <14.3>." + } + ] + }, + { + "video_id": "OBXY5", + "question_id": "OBXY5", + "video_file": "Charades/videos/OBXY5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person washes their hands in the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.4> to <14.3>." + } + ] + }, + { + "video_id": "9CCDY", + "question_id": "9CCDY", + "video_file": "Charades/videos/9CCDY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person laughs at the television' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.8> to <16.4>." + } + ] + }, + { + "video_id": "9CCDY", + "question_id": "9CCDY", + "video_file": "Charades/videos/9CCDY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking from a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <11.1>." + } + ] + }, + { + "video_id": "V0XW3", + "question_id": "V0XW3", + "video_file": "Charades/videos/V0XW3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they grab a drinking glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.5> to <25.9>." + } + ] + }, + { + "video_id": "RKGXU", + "question_id": "RKGXU", + "video_file": "Charades/videos/RKGXU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person also drinking a glass of coffee' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <23.4>." + } + ] + }, + { + "video_id": "F6A4W", + "question_id": "F6A4W", + "video_file": "Charades/videos/F6A4W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking from a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <9.4>." + } + ] + }, + { + "video_id": "W7WJR", + "question_id": "W7WJR", + "video_file": "Charades/videos/W7WJR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns on the light'." + }, + { + "from": "gpt", + "value": "From <14.7> to <21.6>." + } + ] + }, + { + "video_id": "W7WJR", + "question_id": "W7WJR", + "video_file": "Charades/videos/W7WJR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they turn on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <21.6>." + } + ] + }, + { + "video_id": "P6S89", + "question_id": "P6S89", + "video_file": "Charades/videos/P6S89.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person washing a mirror'." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.8>." + } + ] + }, + { + "video_id": "P6S89", + "question_id": "P6S89", + "video_file": "Charades/videos/P6S89.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walking out the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.1> to <32.0>." + } + ] + }, + { + "video_id": "P5ANU", + "question_id": "P5ANU", + "video_file": "Charades/videos/P5ANU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pouring a cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.9> to <27.1>." + } + ] + }, + { + "video_id": "MB1MR", + "question_id": "MB1MR", + "video_file": "Charades/videos/MB1MR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting on their shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.4> to <30.0>." + } + ] + }, + { + "video_id": "MB1MR", + "question_id": "MB1MR", + "video_file": "Charades/videos/MB1MR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put on shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.4> to <30.0>." + } + ] + }, + { + "video_id": "ARCUY", + "question_id": "ARCUY", + "video_file": "Charades/videos/ARCUY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.6> to <34.9>." + } + ] + }, + { + "video_id": "BI86J", + "question_id": "BI86J", + "video_file": "Charades/videos/BI86J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.0> to <14.2>." + } + ] + }, + { + "video_id": "BI86J", + "question_id": "BI86J", + "video_file": "Charades/videos/BI86J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens cabinet'." + }, + { + "from": "gpt", + "value": "From <16.0> to <22.9>." + } + ] + }, + { + "video_id": "BI86J", + "question_id": "BI86J", + "video_file": "Charades/videos/BI86J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put food on counter' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.4> to <29.5>." + } + ] + }, + { + "video_id": "3Z1YC", + "question_id": "3Z1YC", + "video_file": "Charades/videos/3Z1YC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person walks in holding a broom'." + }, + { + "from": "gpt", + "value": "From <14.8> to <29.2>." + } + ] + }, + { + "video_id": "3Z1YC", + "question_id": "3Z1YC", + "video_file": "Charades/videos/3Z1YC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person runs in with a broom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.8> to <27.4>." + } + ] + }, + { + "video_id": "GDS6Q", + "question_id": "GDS6Q", + "video_file": "Charades/videos/GDS6Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they start running around' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.0> to <14.2>." + } + ] + }, + { + "video_id": "FB7MZ", + "question_id": "FB7MZ", + "video_file": "Charades/videos/FB7MZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the shoes on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.8> to <30.0>." + } + ] + }, + { + "video_id": "FB7MZ", + "question_id": "FB7MZ", + "video_file": "Charades/videos/FB7MZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opened the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.1> to <32.0>." + } + ] + }, + { + "video_id": "FB7MZ", + "question_id": "FB7MZ", + "video_file": "Charades/videos/FB7MZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens window blinds' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.1> to <32.0>." + } + ] + }, + { + "video_id": "8XFAV", + "question_id": "8XFAV", + "video_file": "Charades/videos/8XFAV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "8XFAV", + "question_id": "8XFAV", + "video_file": "Charades/videos/8XFAV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they pour a glass of juice' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <13.6>." + } + ] + }, + { + "video_id": "I6H4R", + "question_id": "I6H4R", + "video_file": "Charades/videos/I6H4R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes the bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <11.6>." + } + ] + }, + { + "video_id": "I6H4R", + "question_id": "I6H4R", + "video_file": "Charades/videos/I6H4R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they turn off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.9> to <20.3>." + } + ] + }, + { + "video_id": "KQKNE", + "question_id": "KQKNE", + "video_file": "Charades/videos/KQKNE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "KQKNE", + "question_id": "KQKNE", + "video_file": "Charades/videos/KQKNE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts undressing than dressing again' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "ZP4X6", + "question_id": "ZP4X6", + "video_file": "Charades/videos/ZP4X6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes a phone out from underneath them'." + }, + { + "from": "gpt", + "value": "From <5.3> to <12.5>." + } + ] + }, + { + "video_id": "MWZJY", + "question_id": "MWZJY", + "video_file": "Charades/videos/MWZJY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs into the kitchen' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <8.0>." + } + ] + }, + { + "video_id": "MWZJY", + "question_id": "MWZJY", + "video_file": "Charades/videos/MWZJY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person goes running from one room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <8.0>." + } + ] + }, + { + "video_id": "MWZJY", + "question_id": "MWZJY", + "video_file": "Charades/videos/MWZJY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs into the pantry' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.0> to <8.0>." + } + ] + }, + { + "video_id": "MWZJY", + "question_id": "MWZJY", + "video_file": "Charades/videos/MWZJY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person runs outside'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <8.0>." + } + ] + }, + { + "video_id": "NT2HT", + "question_id": "NT2HT", + "video_file": "Charades/videos/NT2HT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.2> to <12.5>." + } + ] + }, + { + "video_id": "NT2HT", + "question_id": "NT2HT", + "video_file": "Charades/videos/NT2HT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put the book on the boxes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.9> to <26.0>." + } + ] + }, + { + "video_id": "NT2HT", + "question_id": "NT2HT", + "video_file": "Charades/videos/NT2HT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts it back on the table'." + }, + { + "from": "gpt", + "value": "From <21.6> to <26.5>." + } + ] + }, + { + "video_id": "8IUTG", + "question_id": "8IUTG", + "video_file": "Charades/videos/8IUTG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a bag on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.4> to <11.3>." + } + ] + }, + { + "video_id": "8IUTG", + "question_id": "8IUTG", + "video_file": "Charades/videos/8IUTG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they open the bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <11.3>." + } + ] + }, + { + "video_id": "8IUTG", + "question_id": "8IUTG", + "video_file": "Charades/videos/8IUTG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is opening a bag on a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <11.3>." + } + ] + }, + { + "video_id": "M0QER", + "question_id": "M0QER", + "video_file": "Charades/videos/M0QER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the book down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <42.1> to <47.3>." + } + ] + }, + { + "video_id": "DS5J7", + "question_id": "DS5J7", + "video_file": "Charades/videos/DS5J7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'in the dim light a smiling person is watching television' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.3> to <26.0>." + } + ] + }, + { + "video_id": "K2LY1", + "question_id": "K2LY1", + "video_file": "Charades/videos/K2LY1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks through a doorway to the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.8> to <20.1>." + } + ] + }, + { + "video_id": "K2LY1", + "question_id": "K2LY1", + "video_file": "Charades/videos/K2LY1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is awakening from a nap' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.8> to <14.4>." + } + ] + }, + { + "video_id": "HRFGY", + "question_id": "HRFGY", + "video_file": "Charades/videos/HRFGY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts it on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <16.1>." + } + ] + }, + { + "video_id": "HRFGY", + "question_id": "HRFGY", + "video_file": "Charades/videos/HRFGY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting a book on the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <16.6>." + } + ] + }, + { + "video_id": "Q4UO4", + "question_id": "Q4UO4", + "video_file": "Charades/videos/Q4UO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is cooking on a stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "Q4UO4", + "question_id": "Q4UO4", + "video_file": "Charades/videos/Q4UO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person washing their hands'." + }, + { + "from": "gpt", + "value": "From <7.2> to <13.6>." + } + ] + }, + { + "video_id": "Q4UO4", + "question_id": "Q4UO4", + "video_file": "Charades/videos/Q4UO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they open the door to the refrigerator'." + }, + { + "from": "gpt", + "value": "From <14.7> to <23.1>." + } + ] + }, + { + "video_id": "SNUVF", + "question_id": "SNUVF", + "video_file": "Charades/videos/SNUVF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person comes in turns on light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "SNUVF", + "question_id": "SNUVF", + "video_file": "Charades/videos/SNUVF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts eating'." + }, + { + "from": "gpt", + "value": "From <33.0> to <38.0>." + } + ] + }, + { + "video_id": "SNUVF", + "question_id": "SNUVF", + "video_file": "Charades/videos/SNUVF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they eat a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <32.7> to <38.0>." + } + ] + }, + { + "video_id": "SNUVF", + "question_id": "SNUVF", + "video_file": "Charades/videos/SNUVF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person turns on a light in their kitchen' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "LKHKF", + "question_id": "LKHKF", + "video_file": "Charades/videos/LKHKF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person begins laughing hysterically' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.4> to <29.2>." + } + ] + }, + { + "video_id": "2KAWJ", + "question_id": "2KAWJ", + "video_file": "Charades/videos/2KAWJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they stand up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.0> to <20.0>." + } + ] + }, + { + "video_id": "2KAWJ", + "question_id": "2KAWJ", + "video_file": "Charades/videos/2KAWJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person run out of the room through the doorway'." + }, + { + "from": "gpt", + "value": "From <16.0> to <21.9>." + } + ] + }, + { + "video_id": "2KAWJ", + "question_id": "2KAWJ", + "video_file": "Charades/videos/2KAWJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person is sitting on a chair at table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.0> to <33.6>." + } + ] + }, + { + "video_id": "V7KBN", + "question_id": "V7KBN", + "video_file": "Charades/videos/V7KBN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts it on table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.9> to <10.5>." + } + ] + }, + { + "video_id": "VIDIE", + "question_id": "VIDIE", + "video_file": "Charades/videos/VIDIE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws the pair of shoes into the corner' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.0> to <25.9>." + } + ] + }, + { + "video_id": "J5CAN", + "question_id": "J5CAN", + "video_file": "Charades/videos/J5CAN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting on the sofa holding a book'." + }, + { + "from": "gpt", + "value": "From <5.3> to <10.7>." + } + ] + }, + { + "video_id": "J5CAN", + "question_id": "J5CAN", + "video_file": "Charades/videos/J5CAN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the book down on the couch' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <10.1>." + } + ] + }, + { + "video_id": "J5CAN", + "question_id": "J5CAN", + "video_file": "Charades/videos/J5CAN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person snuggling a book awakens from bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <6.8>." + } + ] + }, + { + "video_id": "UEXO4", + "question_id": "UEXO4", + "video_file": "Charades/videos/UEXO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens on a sofa'." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "FP5MM", + "question_id": "FP5MM", + "video_file": "Charades/videos/FP5MM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the same person washes the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <14.5>." + } + ] + }, + { + "video_id": "4ISIX", + "question_id": "4ISIX", + "video_file": "Charades/videos/4ISIX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they open a door'." + }, + { + "from": "gpt", + "value": "From <18.6> to <25.9>." + } + ] + }, + { + "video_id": "4ISIX", + "question_id": "4ISIX", + "video_file": "Charades/videos/4ISIX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walks away though doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.4> to <25.9>." + } + ] + }, + { + "video_id": "4ISIX", + "question_id": "4ISIX", + "video_file": "Charades/videos/4ISIX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throwing clothes on chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <15.9>." + } + ] + }, + { + "video_id": "XYTFC", + "question_id": "XYTFC", + "video_file": "Charades/videos/XYTFC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.0> to <10.4>." + } + ] + }, + { + "video_id": "XYTFC", + "question_id": "XYTFC", + "video_file": "Charades/videos/XYTFC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person takes a fresh towel from the cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <12.6>." + } + ] + }, + { + "video_id": "XYTFC", + "question_id": "XYTFC", + "video_file": "Charades/videos/XYTFC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <10.4>." + } + ] + }, + { + "video_id": "B7DNX", + "question_id": "B7DNX", + "video_file": "Charades/videos/B7DNX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drink a glass of soda' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <29.3>." + } + ] + }, + { + "video_id": "B7DNX", + "question_id": "B7DNX", + "video_file": "Charades/videos/B7DNX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person awakens in a living room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "9MOH2", + "question_id": "9MOH2", + "video_file": "Charades/videos/9MOH2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is walking through a doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.2>." + } + ] + }, + { + "video_id": "9MOH2", + "question_id": "9MOH2", + "video_file": "Charades/videos/9MOH2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening a medicine cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.7> to <16.6>." + } + ] + }, + { + "video_id": "9MOH2", + "question_id": "9MOH2", + "video_file": "Charades/videos/9MOH2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <16.6>." + } + ] + }, + { + "video_id": "XWQ1D", + "question_id": "XWQ1D", + "video_file": "Charades/videos/XWQ1D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person washes their hands' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.1> to <28.0>." + } + ] + }, + { + "video_id": "XWQ1D", + "question_id": "XWQ1D", + "video_file": "Charades/videos/XWQ1D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person wash their hands' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.1> to <28.0>." + } + ] + }, + { + "video_id": "YBP5F", + "question_id": "YBP5F", + "video_file": "Charades/videos/YBP5F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a phone of the fridge' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.4> to <15.5>." + } + ] + }, + { + "video_id": "815KV", + "question_id": "815KV", + "video_file": "Charades/videos/815KV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they walk to a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.2> to <21.7>." + } + ] + }, + { + "video_id": "815KV", + "question_id": "815KV", + "video_file": "Charades/videos/815KV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person puts something onto a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <12.0>." + } + ] + }, + { + "video_id": "815KV", + "question_id": "815KV", + "video_file": "Charades/videos/815KV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts a camera on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.0> to <12.0>." + } + ] + }, + { + "video_id": "815KV", + "question_id": "815KV", + "video_file": "Charades/videos/815KV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the camera on the shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.0> to <12.0>." + } + ] + }, + { + "video_id": "61XR6", + "question_id": "61XR6", + "video_file": "Charades/videos/61XR6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns on a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.2> to <32.3>." + } + ] + }, + { + "video_id": "KU3WI", + "question_id": "KU3WI", + "video_file": "Charades/videos/KU3WI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <8.8>." + } + ] + }, + { + "video_id": "KU3WI", + "question_id": "KU3WI", + "video_file": "Charades/videos/KU3WI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "WTBB2", + "question_id": "WTBB2", + "video_file": "Charades/videos/WTBB2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person was standing up reading a book,then bent down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.6> to <30.3>." + } + ] + }, + { + "video_id": "FKGY7", + "question_id": "FKGY7", + "video_file": "Charades/videos/FKGY7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is smiling as they undress' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.2> to <9.0>." + } + ] + }, + { + "video_id": "FKGY7", + "question_id": "FKGY7", + "video_file": "Charades/videos/FKGY7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the clothes on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.4> to <10.5>." + } + ] + }, + { + "video_id": "FKGY7", + "question_id": "FKGY7", + "video_file": "Charades/videos/FKGY7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a smiling person undresses'." + }, + { + "from": "gpt", + "value": "From <3.2> to <9.0>." + } + ] + }, + { + "video_id": "W0A0Z", + "question_id": "W0A0Z", + "video_file": "Charades/videos/W0A0Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person was holding a cup of coffee' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.8> to <13.0>." + } + ] + }, + { + "video_id": "W0A0Z", + "question_id": "W0A0Z", + "video_file": "Charades/videos/W0A0Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person holds a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.8> to <13.0>." + } + ] + }, + { + "video_id": "KP8VI", + "question_id": "KP8VI", + "video_file": "Charades/videos/KP8VI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a man sitting on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.0>." + } + ] + }, + { + "video_id": "46GP8", + "question_id": "46GP8", + "video_file": "Charades/videos/46GP8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looking out of a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.9> to <21.2>." + } + ] + }, + { + "video_id": "46GP8", + "question_id": "46GP8", + "video_file": "Charades/videos/46GP8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person watching something out a window'." + }, + { + "from": "gpt", + "value": "From <11.9> to <21.2>." + } + ] + }, + { + "video_id": "M2XIS", + "question_id": "M2XIS", + "video_file": "Charades/videos/M2XIS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts their laptop down on a desk'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "M2XIS", + "question_id": "M2XIS", + "video_file": "Charades/videos/M2XIS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person sits in the chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.1> to <6.4>." + } + ] + }, + { + "video_id": "M2XIS", + "question_id": "M2XIS", + "video_file": "Charades/videos/M2XIS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person proceeds to put the lap top the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "BYMO6", + "question_id": "BYMO6", + "video_file": "Charades/videos/BYMO6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person watching a video on their laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "BYMO6", + "question_id": "BYMO6", + "video_file": "Charades/videos/BYMO6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person watches a video on the laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "KSAU8", + "question_id": "KSAU8", + "video_file": "Charades/videos/KSAU8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.8> to <13.9>." + } + ] + }, + { + "video_id": "KOAQD", + "question_id": "KOAQD", + "video_file": "Charades/videos/KOAQD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person undresses' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.8> to <36.1>." + } + ] + }, + { + "video_id": "KOAQD", + "question_id": "KOAQD", + "video_file": "Charades/videos/KOAQD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person remembers the medicine they need to take' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.7> to <26.1>." + } + ] + }, + { + "video_id": "KOAQD", + "question_id": "KOAQD", + "video_file": "Charades/videos/KOAQD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes some medicine'." + }, + { + "from": "gpt", + "value": "From <11.7> to <26.1>." + } + ] + }, + { + "video_id": "55687", + "question_id": "55687", + "video_file": "Charades/videos/55687.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walk out through the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.5> to <32.0>." + } + ] + }, + { + "video_id": "55687", + "question_id": "55687", + "video_file": "Charades/videos/55687.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person stands in the entryway looking at a picture'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.8>." + } + ] + }, + { + "video_id": "55687", + "question_id": "55687", + "video_file": "Charades/videos/55687.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'this person opens the doorknob to leave' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.0> to <32.0>." + } + ] + }, + { + "video_id": "55687", + "question_id": "55687", + "video_file": "Charades/videos/55687.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.0> to <32.0>." + } + ] + }, + { + "video_id": "DGPR1", + "question_id": "DGPR1", + "video_file": "Charades/videos/DGPR1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is throwing a bag down the stairs' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.1> to <7.7>." + } + ] + }, + { + "video_id": "8LPZ0", + "question_id": "8LPZ0", + "video_file": "Charades/videos/8LPZ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a box of groceries from them' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.7> to <34.5>." + } + ] + }, + { + "video_id": "8LPZ0", + "question_id": "8LPZ0", + "video_file": "Charades/videos/8LPZ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person b lies down on a nearby sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <30.3> to <39.0>." + } + ] + }, + { + "video_id": "V1JIC", + "question_id": "V1JIC", + "video_file": "Charades/videos/V1JIC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person lie on a couch'." + }, + { + "from": "gpt", + "value": "From <16.1> to <28.0>." + } + ] + }, + { + "video_id": "V1JIC", + "question_id": "V1JIC", + "video_file": "Charades/videos/V1JIC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes shoes off' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "V1JIC", + "question_id": "V1JIC", + "video_file": "Charades/videos/V1JIC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person grasp a pillow to put under a head' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.2> to <16.4>." + } + ] + }, + { + "video_id": "BZ189", + "question_id": "BZ189", + "video_file": "Charades/videos/BZ189.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person washes their hands in a sink' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <12.8>." + } + ] + }, + { + "video_id": "BZ189", + "question_id": "BZ189", + "video_file": "Charades/videos/BZ189.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they start undressing by taking their jacket off' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.6> to <25.5>." + } + ] + }, + { + "video_id": "37SE6", + "question_id": "37SE6", + "video_file": "Charades/videos/37SE6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begins pouring more soda from a bottle' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <35.8> to <39.0>." + } + ] + }, + { + "video_id": "7HWS7", + "question_id": "7HWS7", + "video_file": "Charades/videos/7HWS7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "2PRCP", + "question_id": "2PRCP", + "video_file": "Charades/videos/2PRCP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a drink from a cup of coffee' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <19.1>." + } + ] + }, + { + "video_id": "D1VWG", + "question_id": "D1VWG", + "video_file": "Charades/videos/D1VWG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they open a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.3> to <10.0>." + } + ] + }, + { + "video_id": "4LU0G", + "question_id": "4LU0G", + "video_file": "Charades/videos/4LU0G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is seen putting dishes on their desk' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <8.3>." + } + ] + }, + { + "video_id": "119W9", + "question_id": "119W9", + "video_file": "Charades/videos/119W9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person closes the refrigerator door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <9.7>." + } + ] + }, + { + "video_id": "119W9", + "question_id": "119W9", + "video_file": "Charades/videos/119W9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begins cooking food on the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <26.0>." + } + ] + }, + { + "video_id": "CBXVQ", + "question_id": "CBXVQ", + "video_file": "Charades/videos/CBXVQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <22.1>." + } + ] + }, + { + "video_id": "CBXVQ", + "question_id": "CBXVQ", + "video_file": "Charades/videos/CBXVQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens up a cabinet'." + }, + { + "from": "gpt", + "value": "From <7.4> to <22.1>." + } + ] + }, + { + "video_id": "CBXVQ", + "question_id": "CBXVQ", + "video_file": "Charades/videos/CBXVQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person wash their hands'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.8> to <35.0>." + } + ] + }, + { + "video_id": "SM1QV", + "question_id": "SM1QV", + "video_file": "Charades/videos/SM1QV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sneezing in the entryway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <14.0>." + } + ] + }, + { + "video_id": "94ICP", + "question_id": "94ICP", + "video_file": "Charades/videos/94ICP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person he is cooking'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "94ICP", + "question_id": "94ICP", + "video_file": "Charades/videos/94ICP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person cooks on a stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "D07V2", + "question_id": "D07V2", + "video_file": "Charades/videos/D07V2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.6> to <10.3>." + } + ] + }, + { + "video_id": "D07V2", + "question_id": "D07V2", + "video_file": "Charades/videos/D07V2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person go to the stove to start cooking' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.7> to <28.0>." + } + ] + }, + { + "video_id": "D07V2", + "question_id": "D07V2", + "video_file": "Charades/videos/D07V2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is drinking a cup of coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "D07V2", + "question_id": "D07V2", + "video_file": "Charades/videos/D07V2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they open the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.6> to <10.3>." + } + ] + }, + { + "video_id": "9SKPI", + "question_id": "9SKPI", + "video_file": "Charades/videos/9SKPI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is tidying up a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <17.6>." + } + ] + }, + { + "video_id": "9SKPI", + "question_id": "9SKPI", + "video_file": "Charades/videos/9SKPI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person putting dishes on a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <26.0>." + } + ] + }, + { + "video_id": "9SKPI", + "question_id": "9SKPI", + "video_file": "Charades/videos/9SKPI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins to tidy up a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <18.0>." + } + ] + }, + { + "video_id": "TPSUY", + "question_id": "TPSUY", + "video_file": "Charades/videos/TPSUY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person holding a broom throws it'." + }, + { + "from": "gpt", + "value": "From <6.7> to <11.9>." + } + ] + }, + { + "video_id": "7KEON", + "question_id": "7KEON", + "video_file": "Charades/videos/7KEON.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts a book on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.8> to <25.6>." + } + ] + }, + { + "video_id": "7KEON", + "question_id": "7KEON", + "video_file": "Charades/videos/7KEON.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put a book on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.8> to <25.6>." + } + ] + }, + { + "video_id": "V00AL", + "question_id": "V00AL", + "video_file": "Charades/videos/V00AL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person talks on the phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "V00AL", + "question_id": "V00AL", + "video_file": "Charades/videos/V00AL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person laying down smiling on phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.7> to <19.4>." + } + ] + }, + { + "video_id": "3AU5K", + "question_id": "3AU5K", + "video_file": "Charades/videos/3AU5K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a sip from a glass'." + }, + { + "from": "gpt", + "value": "From <6.1> to <11.8>." + } + ] + }, + { + "video_id": "GTK8W", + "question_id": "GTK8W", + "video_file": "Charades/videos/GTK8W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a bag out of the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <30.8> to <36.6>." + } + ] + }, + { + "video_id": "GTK8W", + "question_id": "GTK8W", + "video_file": "Charades/videos/GTK8W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits back down in the chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.6> to <38.0>." + } + ] + }, + { + "video_id": "GTK8W", + "question_id": "GTK8W", + "video_file": "Charades/videos/GTK8W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating food'." + }, + { + "from": "gpt", + "value": "From <26.8> to <32.7>." + } + ] + }, + { + "video_id": "E4C1C", + "question_id": "E4C1C", + "video_file": "Charades/videos/E4C1C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <20.6> to <25.3>." + } + ] + }, + { + "video_id": "6WZ1K", + "question_id": "6WZ1K", + "video_file": "Charades/videos/6WZ1K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person working on their laptop'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "FWUV8", + "question_id": "FWUV8", + "video_file": "Charades/videos/FWUV8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is drinking a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.9> to <19.0>." + } + ] + }, + { + "video_id": "FWUV8", + "question_id": "FWUV8", + "video_file": "Charades/videos/FWUV8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person watches holding a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.3> to <38.0>." + } + ] + }, + { + "video_id": "FWUV8", + "question_id": "FWUV8", + "video_file": "Charades/videos/FWUV8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person drinks from a cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.9> to <19.0>." + } + ] + }, + { + "video_id": "FWUV8", + "question_id": "FWUV8", + "video_file": "Charades/videos/FWUV8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a pillow'." + }, + { + "from": "gpt", + "value": "From <23.3> to <38.0>." + } + ] + }, + { + "video_id": "Y063A", + "question_id": "Y063A", + "video_file": "Charades/videos/Y063A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they start sneezing'." + }, + { + "from": "gpt", + "value": "From <19.5> to <26.0>." + } + ] + }, + { + "video_id": "UY08H", + "question_id": "UY08H", + "video_file": "Charades/videos/UY08H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "UY08H", + "question_id": "UY08H", + "video_file": "Charades/videos/UY08H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts it into the box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.7>." + } + ] + }, + { + "video_id": "QIQXR", + "question_id": "QIQXR", + "video_file": "Charades/videos/QIQXR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting on pink tennis shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.0> to <15.9>." + } + ] + }, + { + "video_id": "QIQXR", + "question_id": "QIQXR", + "video_file": "Charades/videos/QIQXR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person puts on shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.0> to <15.9>." + } + ] + }, + { + "video_id": "WJ33D", + "question_id": "WJ33D", + "video_file": "Charades/videos/WJ33D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes some clothes from a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.8> to <19.1>." + } + ] + }, + { + "video_id": "WJ33D", + "question_id": "WJ33D", + "video_file": "Charades/videos/WJ33D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts the towel on top of the wardrobe' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.2> to <24.0>." + } + ] + }, + { + "video_id": "C4KZ4", + "question_id": "C4KZ4", + "video_file": "Charades/videos/C4KZ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person closes a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <5.8>." + } + ] + }, + { + "video_id": "V115C", + "question_id": "V115C", + "video_file": "Charades/videos/V115C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws the clothes onto a desk' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <23.0>." + } + ] + }, + { + "video_id": "SIUU5", + "question_id": "SIUU5", + "video_file": "Charades/videos/SIUU5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turn on the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <12.2>." + } + ] + }, + { + "video_id": "SIUU5", + "question_id": "SIUU5", + "video_file": "Charades/videos/SIUU5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <7.2>." + } + ] + }, + { + "video_id": "SIUU5", + "question_id": "SIUU5", + "video_file": "Charades/videos/SIUU5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns on the light'." + }, + { + "from": "gpt", + "value": "From <6.6> to <12.2>." + } + ] + }, + { + "video_id": "1ZEJJ", + "question_id": "1ZEJJ", + "video_file": "Charades/videos/1ZEJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person start laughing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.7> to <18.9>." + } + ] + }, + { + "video_id": "LN3WU", + "question_id": "LN3WU", + "video_file": "Charades/videos/LN3WU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person laughing on the phone'." + }, + { + "from": "gpt", + "value": "From <2.0> to <8.9>." + } + ] + }, + { + "video_id": "LN3WU", + "question_id": "LN3WU", + "video_file": "Charades/videos/LN3WU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person bends down takes off shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <17.6>." + } + ] + }, + { + "video_id": "SILHS", + "question_id": "SILHS", + "video_file": "Charades/videos/SILHS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is holding a blanket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.8>." + } + ] + }, + { + "video_id": "SILHS", + "question_id": "SILHS", + "video_file": "Charades/videos/SILHS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is putting a blanket in a wardrobe'." + }, + { + "from": "gpt", + "value": "From <3.6> to <9.8>." + } + ] + }, + { + "video_id": "3SAO5", + "question_id": "3SAO5", + "video_file": "Charades/videos/3SAO5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the cabinet'." + }, + { + "from": "gpt", + "value": "From <22.4> to <30.7>." + } + ] + }, + { + "video_id": "3SAO5", + "question_id": "3SAO5", + "video_file": "Charades/videos/3SAO5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes out some food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.3> to <32.9>." + } + ] + }, + { + "video_id": "3SAO5", + "question_id": "3SAO5", + "video_file": "Charades/videos/3SAO5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person reading a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <17.7>." + } + ] + }, + { + "video_id": "3SAO5", + "question_id": "3SAO5", + "video_file": "Charades/videos/3SAO5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sitting down in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.6>." + } + ] + }, + { + "video_id": "3SAO5", + "question_id": "3SAO5", + "video_file": "Charades/videos/3SAO5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <16.5>." + } + ] + }, + { + "video_id": "3SAO5", + "question_id": "3SAO5", + "video_file": "Charades/videos/3SAO5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <24.7>." + } + ] + }, + { + "video_id": "3X52E", + "question_id": "3X52E", + "video_file": "Charades/videos/3X52E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is undressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.9>." + } + ] + }, + { + "video_id": "3X52E", + "question_id": "3X52E", + "video_file": "Charades/videos/3X52E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throw it on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "3X52E", + "question_id": "3X52E", + "video_file": "Charades/videos/3X52E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they start throwing a blanket on the ground'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <19.8>." + } + ] + }, + { + "video_id": "ZWVNF", + "question_id": "ZWVNF", + "video_file": "Charades/videos/ZWVNF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is seen opening a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <9.1>." + } + ] + }, + { + "video_id": "ZWVNF", + "question_id": "ZWVNF", + "video_file": "Charades/videos/ZWVNF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a bag on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <9.1>." + } + ] + }, + { + "video_id": "0NCWX", + "question_id": "0NCWX", + "video_file": "Charades/videos/0NCWX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is opening a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "70S6Y", + "question_id": "70S6Y", + "video_file": "Charades/videos/70S6Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs into a bedroom holding a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <15.4>." + } + ] + }, + { + "video_id": "70S6Y", + "question_id": "70S6Y", + "video_file": "Charades/videos/70S6Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <13.3>." + } + ] + }, + { + "video_id": "70S6Y", + "question_id": "70S6Y", + "video_file": "Charades/videos/70S6Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.1> to <19.8>." + } + ] + }, + { + "video_id": "KZTDH", + "question_id": "KZTDH", + "video_file": "Charades/videos/KZTDH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person washes their hands in a sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.8> to <12.1>." + } + ] + }, + { + "video_id": "KZTDH", + "question_id": "KZTDH", + "video_file": "Charades/videos/KZTDH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is washing their hands at the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.8> to <12.1>." + } + ] + }, + { + "video_id": "KZTDH", + "question_id": "KZTDH", + "video_file": "Charades/videos/KZTDH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person washes their hands' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.8> to <12.1>." + } + ] + }, + { + "video_id": "FZIW9", + "question_id": "FZIW9", + "video_file": "Charades/videos/FZIW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a sandwich'." + }, + { + "from": "gpt", + "value": "From <26.8> to <34.0>." + } + ] + }, + { + "video_id": "HC62A", + "question_id": "HC62A", + "video_file": "Charades/videos/HC62A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person sits in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "HC62A", + "question_id": "HC62A", + "video_file": "Charades/videos/HC62A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes a bag of food from the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <12.2>." + } + ] + }, + { + "video_id": "HC62A", + "question_id": "HC62A", + "video_file": "Charades/videos/HC62A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a piece of food out' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <12.2>." + } + ] + }, + { + "video_id": "HC62A", + "question_id": "HC62A", + "video_file": "Charades/videos/HC62A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person sits down in a chair to watch television' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "HC62A", + "question_id": "HC62A", + "video_file": "Charades/videos/HC62A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens a bag of groceries' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.6>." + } + ] + }, + { + "video_id": "5AM4C", + "question_id": "5AM4C", + "video_file": "Charades/videos/5AM4C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person throws a bag on the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "IUVT3", + "question_id": "IUVT3", + "video_file": "Charades/videos/IUVT3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is opening the bathroom door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.7> to <8.9>." + } + ] + }, + { + "video_id": "IUVT3", + "question_id": "IUVT3", + "video_file": "Charades/videos/IUVT3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a bathroom door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.7> to <8.9>." + } + ] + }, + { + "video_id": "T4NAU", + "question_id": "T4NAU", + "video_file": "Charades/videos/T4NAU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "T4NAU", + "question_id": "T4NAU", + "video_file": "Charades/videos/T4NAU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they start undressing by taking a jacket off' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.0> to <31.3>." + } + ] + }, + { + "video_id": "5B2K2", + "question_id": "5B2K2", + "video_file": "Charades/videos/5B2K2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person walks outside to take a picture of shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.6> to <31.0>." + } + ] + }, + { + "video_id": "5B2K2", + "question_id": "5B2K2", + "video_file": "Charades/videos/5B2K2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person the people in the video are laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <9.9>." + } + ] + }, + { + "video_id": "SHKHX", + "question_id": "SHKHX", + "video_file": "Charades/videos/SHKHX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person drinks from a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.1> to <6.7>." + } + ] + }, + { + "video_id": "JZ46A", + "question_id": "JZ46A", + "video_file": "Charades/videos/JZ46A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes a pillow from nearby' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.6> to <17.4>." + } + ] + }, + { + "video_id": "JCSTR", + "question_id": "JCSTR", + "video_file": "Charades/videos/JCSTR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person watching something on their laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "JCSTR", + "question_id": "JCSTR", + "video_file": "Charades/videos/JCSTR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person drinks from a coffee cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <9.3>." + } + ] + }, + { + "video_id": "HG00H", + "question_id": "HG00H", + "video_file": "Charades/videos/HG00H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take a picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.2> to <36.0>." + } + ] + }, + { + "video_id": "HG00H", + "question_id": "HG00H", + "video_file": "Charades/videos/HG00H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take a camera out'." + }, + { + "from": "gpt", + "value": "From <23.9> to <30.5>." + } + ] + }, + { + "video_id": "HG00H", + "question_id": "HG00H", + "video_file": "Charades/videos/HG00H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes out a camera' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.9> to <30.5>." + } + ] + }, + { + "video_id": "HG00H", + "question_id": "HG00H", + "video_file": "Charades/videos/HG00H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.2> to <36.0>." + } + ] + }, + { + "video_id": "JOBFJ", + "question_id": "JOBFJ", + "video_file": "Charades/videos/JOBFJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person holds shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <8.8>." + } + ] + }, + { + "video_id": "JOBFJ", + "question_id": "JOBFJ", + "video_file": "Charades/videos/JOBFJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person puts their shoes on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.4> to <9.9>." + } + ] + }, + { + "video_id": "KPNJ5", + "question_id": "KPNJ5", + "video_file": "Charades/videos/KPNJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person walks in playing on a phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.5> to <21.5>." + } + ] + }, + { + "video_id": "KPNJ5", + "question_id": "KPNJ5", + "video_file": "Charades/videos/KPNJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person playing a game on their phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.5> to <21.5>." + } + ] + }, + { + "video_id": "KPNJ5", + "question_id": "KPNJ5", + "video_file": "Charades/videos/KPNJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person walks through the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.5> to <17.6>." + } + ] + }, + { + "video_id": "T0LOH", + "question_id": "T0LOH", + "video_file": "Charades/videos/T0LOH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person sits on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.6> to <16.0>." + } + ] + }, + { + "video_id": "T0LOH", + "question_id": "T0LOH", + "video_file": "Charades/videos/T0LOH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <16.0>." + } + ] + }, + { + "video_id": "ZSREG", + "question_id": "ZSREG", + "video_file": "Charades/videos/ZSREG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a cabinet door that is at eye level'." + }, + { + "from": "gpt", + "value": "From <0.7> to <7.8>." + } + ] + }, + { + "video_id": "M3S4D", + "question_id": "M3S4D", + "video_file": "Charades/videos/M3S4D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person runs into the room'." + }, + { + "from": "gpt", + "value": "From <2.9> to <8.7>." + } + ] + }, + { + "video_id": "M3S4D", + "question_id": "M3S4D", + "video_file": "Charades/videos/M3S4D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts a box on the ground'." + }, + { + "from": "gpt", + "value": "From <1.2> to <11.0>." + } + ] + }, + { + "video_id": "M3S4D", + "question_id": "M3S4D", + "video_file": "Charades/videos/M3S4D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person runs into the room holding a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.9> to <8.7>." + } + ] + }, + { + "video_id": "M3S4D", + "question_id": "M3S4D", + "video_file": "Charades/videos/M3S4D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts down the box on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.2> to <11.0>." + } + ] + }, + { + "video_id": "M3S4D", + "question_id": "M3S4D", + "video_file": "Charades/videos/M3S4D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinks from the glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.7> to <17.5>." + } + ] + }, + { + "video_id": "1SCAZ", + "question_id": "1SCAZ", + "video_file": "Charades/videos/1SCAZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person was tidying the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.5>." + } + ] + }, + { + "video_id": "19MUM", + "question_id": "19MUM", + "video_file": "Charades/videos/19MUM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the door behind them' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <27.8> to <32.0>." + } + ] + }, + { + "video_id": "19MUM", + "question_id": "19MUM", + "video_file": "Charades/videos/19MUM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.0> to <32.0>." + } + ] + }, + { + "video_id": "19MUM", + "question_id": "19MUM", + "video_file": "Charades/videos/19MUM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is washing dishes in a sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "19MUM", + "question_id": "19MUM", + "video_file": "Charades/videos/19MUM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drink from a water bottle'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <23.0>." + } + ] + }, + { + "video_id": "7XBPE", + "question_id": "7XBPE", + "video_file": "Charades/videos/7XBPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens the door of the fridge' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "7XBPE", + "question_id": "7XBPE", + "video_file": "Charades/videos/7XBPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes out a bag'." + }, + { + "from": "gpt", + "value": "From <2.3> to <8.7>." + } + ] + }, + { + "video_id": "2BFZG", + "question_id": "2BFZG", + "video_file": "Charades/videos/2BFZG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is dressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.4> to <31.0>." + } + ] + }, + { + "video_id": "X3952", + "question_id": "X3952", + "video_file": "Charades/videos/X3952.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person is undressing in the bathroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "DTUJD", + "question_id": "DTUJD", + "video_file": "Charades/videos/DTUJD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a closet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <17.0>." + } + ] + }, + { + "video_id": "DTUJD", + "question_id": "DTUJD", + "video_file": "Charades/videos/DTUJD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts some clothes on the top shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <17.0>." + } + ] + }, + { + "video_id": "MYGJA", + "question_id": "MYGJA", + "video_file": "Charades/videos/MYGJA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turn the doorknob to open the pantry door'." + }, + { + "from": "gpt", + "value": "From <5.7> to <11.6>." + } + ] + }, + { + "video_id": "MYGJA", + "question_id": "MYGJA", + "video_file": "Charades/videos/MYGJA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts a blanket in the closet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <15.0>." + } + ] + }, + { + "video_id": "M02OO", + "question_id": "M02OO", + "video_file": "Charades/videos/M02OO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person is laughing in a doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "SHTQ9", + "question_id": "SHTQ9", + "video_file": "Charades/videos/SHTQ9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person runs through'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "R1S1Z", + "question_id": "R1S1Z", + "video_file": "Charades/videos/R1S1Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws a bag at them' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <15.0>." + } + ] + }, + { + "video_id": "SZJ09", + "question_id": "SZJ09", + "video_file": "Charades/videos/SZJ09.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting in a chair reading a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <17.9>." + } + ] + }, + { + "video_id": "0CQL8", + "question_id": "0CQL8", + "video_file": "Charades/videos/0CQL8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person holding a box opens a closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <9.7>." + } + ] + }, + { + "video_id": "81NAT", + "question_id": "81NAT", + "video_file": "Charades/videos/81NAT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.8> to <13.6>." + } + ] + }, + { + "video_id": "81NAT", + "question_id": "81NAT", + "video_file": "Charades/videos/81NAT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is in a dining room undressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.8> to <13.6>." + } + ] + }, + { + "video_id": "E2EXR", + "question_id": "E2EXR", + "video_file": "Charades/videos/E2EXR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person was drinking from a bottle' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.1> to <22.6>." + } + ] + }, + { + "video_id": "E2EXR", + "question_id": "E2EXR", + "video_file": "Charades/videos/E2EXR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking something out of a bottle' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.1> to <22.6>." + } + ] + }, + { + "video_id": "HDMN4", + "question_id": "HDMN4", + "video_file": "Charades/videos/HDMN4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person lies down on a bed next to a laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.9> to <31.0>." + } + ] + }, + { + "video_id": "HDMN4", + "question_id": "HDMN4", + "video_file": "Charades/videos/HDMN4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts the glass down'." + }, + { + "from": "gpt", + "value": "From <4.8> to <9.6>." + } + ] + }, + { + "video_id": "E2EHR", + "question_id": "E2EHR", + "video_file": "Charades/videos/E2EHR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.9>." + } + ] + }, + { + "video_id": "8IPWO", + "question_id": "8IPWO", + "video_file": "Charades/videos/8IPWO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person reading a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.5> to <32.0>." + } + ] + }, + { + "video_id": "8IPWO", + "question_id": "8IPWO", + "video_file": "Charades/videos/8IPWO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opened a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.7> to <14.4>." + } + ] + }, + { + "video_id": "63XVB", + "question_id": "63XVB", + "video_file": "Charades/videos/63XVB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person smiling to themselves thoughtfully'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <26.0>." + } + ] + }, + { + "video_id": "TJ3GZ", + "question_id": "TJ3GZ", + "video_file": "Charades/videos/TJ3GZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is putting on shoes in the bedroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.5> to <13.8>." + } + ] + }, + { + "video_id": "TJ3GZ", + "question_id": "TJ3GZ", + "video_file": "Charades/videos/TJ3GZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person watching the tv puts on shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <13.8>." + } + ] + }, + { + "video_id": "XFRYR", + "question_id": "XFRYR", + "video_file": "Charades/videos/XFRYR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes some boxes out of a cabinet'." + }, + { + "from": "gpt", + "value": "From <5.4> to <15.2>." + } + ] + }, + { + "video_id": "LAJ9K", + "question_id": "LAJ9K", + "video_file": "Charades/videos/LAJ9K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is walking near the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <9.0>." + } + ] + }, + { + "video_id": "A2DDA", + "question_id": "A2DDA", + "video_file": "Charades/videos/A2DDA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person take some medicine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <10.8>." + } + ] + }, + { + "video_id": "WCCY7", + "question_id": "WCCY7", + "video_file": "Charades/videos/WCCY7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person undressing by a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.8> to <11.3>." + } + ] + }, + { + "video_id": "KW0K8", + "question_id": "KW0K8", + "video_file": "Charades/videos/KW0K8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person talking on phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "KW0K8", + "question_id": "KW0K8", + "video_file": "Charades/videos/KW0K8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins to sit on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <9.4>." + } + ] + }, + { + "video_id": "KW0K8", + "question_id": "KW0K8", + "video_file": "Charades/videos/KW0K8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person talking on the phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "KW0K8", + "question_id": "KW0K8", + "video_file": "Charades/videos/KW0K8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sits down on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.5> to <9.4>." + } + ] + }, + { + "video_id": "KW0K8", + "question_id": "KW0K8", + "video_file": "Charades/videos/KW0K8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person sits down in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <9.4>." + } + ] + }, + { + "video_id": "KW0K8", + "question_id": "KW0K8", + "video_file": "Charades/videos/KW0K8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is talking on the phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "TTQ3V", + "question_id": "TTQ3V", + "video_file": "Charades/videos/TTQ3V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a blanket in one hand'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.4> to <12.2>." + } + ] + }, + { + "video_id": "TTQ3V", + "question_id": "TTQ3V", + "video_file": "Charades/videos/TTQ3V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they start working on something on their laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.1>." + } + ] + }, + { + "video_id": "TTQ3V", + "question_id": "TTQ3V", + "video_file": "Charades/videos/TTQ3V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is running down their stairs' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.2> to <26.4>." + } + ] + }, + { + "video_id": "D0E8H", + "question_id": "D0E8H", + "video_file": "Charades/videos/D0E8H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they begin eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.5> to <51.5>." + } + ] + }, + { + "video_id": "H40VV", + "question_id": "H40VV", + "video_file": "Charades/videos/H40VV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.9> to <24.7>." + } + ] + }, + { + "video_id": "H40VV", + "question_id": "H40VV", + "video_file": "Charades/videos/H40VV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.4> to <11.7>." + } + ] + }, + { + "video_id": "VYW49", + "question_id": "VYW49", + "video_file": "Charades/videos/VYW49.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits down in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.0> to <32.0>." + } + ] + }, + { + "video_id": "VYW49", + "question_id": "VYW49", + "video_file": "Charades/videos/VYW49.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.0> to <32.0>." + } + ] + }, + { + "video_id": "PBV4T", + "question_id": "PBV4T", + "video_file": "Charades/videos/PBV4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting in a chair eating' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "PBV4T", + "question_id": "PBV4T", + "video_file": "Charades/videos/PBV4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person sitting in a chair eats a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "Y3NN8", + "question_id": "Y3NN8", + "video_file": "Charades/videos/Y3NN8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they stand up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.5> to <23.4>." + } + ] + }, + { + "video_id": "15FVJ", + "question_id": "15FVJ", + "video_file": "Charades/videos/15FVJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws a pillow on a bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.9> to <8.2>." + } + ] + }, + { + "video_id": "15FVJ", + "question_id": "15FVJ", + "video_file": "Charades/videos/15FVJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person throws a pillow onto a bed'." + }, + { + "from": "gpt", + "value": "From <2.9> to <8.2>." + } + ] + }, + { + "video_id": "15FVJ", + "question_id": "15FVJ", + "video_file": "Charades/videos/15FVJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person holding a pillow lies in bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <7.9>." + } + ] + }, + { + "video_id": "DGPUE", + "question_id": "DGPUE", + "video_file": "Charades/videos/DGPUE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person drinks from a cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.5> to <21.9>." + } + ] + }, + { + "video_id": "GDTYU", + "question_id": "GDTYU", + "video_file": "Charades/videos/GDTYU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person decides to undress their clothes holding a broom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.0> to <29.0>." + } + ] + }, + { + "video_id": "GDTYU", + "question_id": "GDTYU", + "video_file": "Charades/videos/GDTYU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is holding a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.5>." + } + ] + }, + { + "video_id": "H8S4L", + "question_id": "H8S4L", + "video_file": "Charades/videos/H8S4L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes a bottle of medicine from the cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.8> to <19.9>." + } + ] + }, + { + "video_id": "H8S4L", + "question_id": "H8S4L", + "video_file": "Charades/videos/H8S4L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes some medicine' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.8> to <19.9>." + } + ] + }, + { + "video_id": "H8S4L", + "question_id": "H8S4L", + "video_file": "Charades/videos/H8S4L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is working on their laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "A1PEO", + "question_id": "A1PEO", + "video_file": "Charades/videos/A1PEO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <1.9> to <8.3>." + } + ] + }, + { + "video_id": "A1PEO", + "question_id": "A1PEO", + "video_file": "Charades/videos/A1PEO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the door to leave' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "OX5TB", + "question_id": "OX5TB", + "video_file": "Charades/videos/OX5TB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is dressing in clothes'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "NFAA5", + "question_id": "NFAA5", + "video_file": "Charades/videos/NFAA5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person is in the house talking on the phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.3> to <32.0>." + } + ] + }, + { + "video_id": "NFAA5", + "question_id": "NFAA5", + "video_file": "Charades/videos/NFAA5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person smiling down the hall' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.9> to <32.0>." + } + ] + }, + { + "video_id": "ZAOWM", + "question_id": "ZAOWM", + "video_file": "Charades/videos/ZAOWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person is eating on the phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "EIZFC", + "question_id": "EIZFC", + "video_file": "Charades/videos/EIZFC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person is laughing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.5> to <32.0>." + } + ] + }, + { + "video_id": "MGDO2", + "question_id": "MGDO2", + "video_file": "Charades/videos/MGDO2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the door behind them'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.8> to <29.0>." + } + ] + }, + { + "video_id": "SQUX3", + "question_id": "SQUX3", + "video_file": "Charades/videos/SQUX3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put on some shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.4> to <38.3>." + } + ] + }, + { + "video_id": "SQUX3", + "question_id": "SQUX3", + "video_file": "Charades/videos/SQUX3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person lied down on a bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <21.7>." + } + ] + }, + { + "video_id": "SQUX3", + "question_id": "SQUX3", + "video_file": "Charades/videos/SQUX3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting on shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.4> to <38.3>." + } + ] + }, + { + "video_id": "ZKNHG", + "question_id": "ZKNHG", + "video_file": "Charades/videos/ZKNHG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take a picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.2> to <39.2>." + } + ] + }, + { + "video_id": "ZKNHG", + "question_id": "ZKNHG", + "video_file": "Charades/videos/ZKNHG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person play with the phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.2>." + } + ] + }, + { + "video_id": "ZKNHG", + "question_id": "ZKNHG", + "video_file": "Charades/videos/ZKNHG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is undressing by their bedroom door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.5> to <15.4>." + } + ] + }, + { + "video_id": "DB52Q", + "question_id": "DB52Q", + "video_file": "Charades/videos/DB52Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person pours water into a glass'." + }, + { + "from": "gpt", + "value": "From <17.0> to <24.6>." + } + ] + }, + { + "video_id": "DB52Q", + "question_id": "DB52Q", + "video_file": "Charades/videos/DB52Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a cabinet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "DB52Q", + "question_id": "DB52Q", + "video_file": "Charades/videos/DB52Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pours some water into a cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.0> to <24.6>." + } + ] + }, + { + "video_id": "9YNN1", + "question_id": "9YNN1", + "video_file": "Charades/videos/9YNN1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person tidying a shelf'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "DFJ3N", + "question_id": "DFJ3N", + "video_file": "Charades/videos/DFJ3N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person drinks from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.7> to <12.1>." + } + ] + }, + { + "video_id": "DFJ3N", + "question_id": "DFJ3N", + "video_file": "Charades/videos/DFJ3N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is drinking water from a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <12.1>." + } + ] + }, + { + "video_id": "DFJ3N", + "question_id": "DFJ3N", + "video_file": "Charades/videos/DFJ3N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts a bag of groceries on the table'." + }, + { + "from": "gpt", + "value": "From <11.9> to <19.6>." + } + ] + }, + { + "video_id": "D7MHV", + "question_id": "D7MHV", + "video_file": "Charades/videos/D7MHV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a cabinet'." + }, + { + "from": "gpt", + "value": "From <19.6> to <30.0>." + } + ] + }, + { + "video_id": "BKXN4", + "question_id": "BKXN4", + "video_file": "Charades/videos/BKXN4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person start dressing by putting their shoes on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <26.4>." + } + ] + }, + { + "video_id": "BKXN4", + "question_id": "BKXN4", + "video_file": "Charades/videos/BKXN4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person was working on laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.7> to <9.2>." + } + ] + }, + { + "video_id": "BKXN4", + "question_id": "BKXN4", + "video_file": "Charades/videos/BKXN4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person proceeded to put shoes on' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.8> to <26.4>." + } + ] + }, + { + "video_id": "BKXN4", + "question_id": "BKXN4", + "video_file": "Charades/videos/BKXN4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is in a garage working on a laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.7> to <9.2>." + } + ] + }, + { + "video_id": "2WH4F", + "question_id": "2WH4F", + "video_file": "Charades/videos/2WH4F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person starts undressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.6> to <31.0>." + } + ] + }, + { + "video_id": "2WH4F", + "question_id": "2WH4F", + "video_file": "Charades/videos/2WH4F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the clothes on a chair in the bathroom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.2> to <31.0>." + } + ] + }, + { + "video_id": "2WH4F", + "question_id": "2WH4F", + "video_file": "Charades/videos/2WH4F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open a door under the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.3> to <25.8>." + } + ] + }, + { + "video_id": "40TYU", + "question_id": "40TYU", + "video_file": "Charades/videos/40TYU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws the box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.7> to <14.6>." + } + ] + }, + { + "video_id": "40TYU", + "question_id": "40TYU", + "video_file": "Charades/videos/40TYU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws the box onto the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <14.6>." + } + ] + }, + { + "video_id": "O2294", + "question_id": "O2294", + "video_file": "Charades/videos/O2294.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person undressed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.7> to <18.9>." + } + ] + }, + { + "video_id": "O2294", + "question_id": "O2294", + "video_file": "Charades/videos/O2294.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opened a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.6> to <32.0>." + } + ] + }, + { + "video_id": "O2294", + "question_id": "O2294", + "video_file": "Charades/videos/O2294.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening a bag in a garage' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.6> to <32.0>." + } + ] + }, + { + "video_id": "O2294", + "question_id": "O2294", + "video_file": "Charades/videos/O2294.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person watching television' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.6> to <29.7>." + } + ] + }, + { + "video_id": "O2294", + "question_id": "O2294", + "video_file": "Charades/videos/O2294.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walked through the doorway into another room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.3> to <12.5>." + } + ] + }, + { + "video_id": "O2294", + "question_id": "O2294", + "video_file": "Charades/videos/O2294.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person watches television' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.6> to <29.7>." + } + ] + }, + { + "video_id": "8MLCU", + "question_id": "8MLCU", + "video_file": "Charades/videos/8MLCU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts away a box of cereal' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.4> to <10.2>." + } + ] + }, + { + "video_id": "8MLCU", + "question_id": "8MLCU", + "video_file": "Charades/videos/8MLCU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put a cereal box into the pantry'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <10.2>." + } + ] + }, + { + "video_id": "A7BUX", + "question_id": "A7BUX", + "video_file": "Charades/videos/A7BUX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person is sitting on teh kitchen table eating some food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.2>." + } + ] + }, + { + "video_id": "A7BUX", + "question_id": "A7BUX", + "video_file": "Charades/videos/A7BUX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.7>." + } + ] + }, + { + "video_id": "MDOSW", + "question_id": "MDOSW", + "video_file": "Charades/videos/MDOSW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating on a small sofa'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "MDOSW", + "question_id": "MDOSW", + "video_file": "Charades/videos/MDOSW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is eating out of a bag'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "GKSL2", + "question_id": "GKSL2", + "video_file": "Charades/videos/GKSL2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking a cup of milk'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.2> to <17.6>." + } + ] + }, + { + "video_id": "GKSL2", + "question_id": "GKSL2", + "video_file": "Charades/videos/GKSL2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person talk on a phone'." + }, + { + "from": "gpt", + "value": "From <24.4> to <39.0>." + } + ] + }, + { + "video_id": "GKSL2", + "question_id": "GKSL2", + "video_file": "Charades/videos/GKSL2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person talking on the phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.4> to <39.0>." + } + ] + }, + { + "video_id": "GKSL2", + "question_id": "GKSL2", + "video_file": "Charades/videos/GKSL2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is in their living room eating a cookie' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.1> to <13.2>." + } + ] + }, + { + "video_id": "GKSL2", + "question_id": "GKSL2", + "video_file": "Charades/videos/GKSL2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person sits in a chair eating a snack' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.1> to <13.2>." + } + ] + }, + { + "video_id": "GKSL2", + "question_id": "GKSL2", + "video_file": "Charades/videos/GKSL2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.1> to <13.2>." + } + ] + }, + { + "video_id": "5I54H", + "question_id": "5I54H", + "video_file": "Charades/videos/5I54H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person start sneezing into it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.9> to <21.2>." + } + ] + }, + { + "video_id": "5I54H", + "question_id": "5I54H", + "video_file": "Charades/videos/5I54H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts their watch on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.3> to <12.9>." + } + ] + }, + { + "video_id": "O8VYU", + "question_id": "O8VYU", + "video_file": "Charades/videos/O8VYU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.5> to <33.6>." + } + ] + }, + { + "video_id": "MZJ9D", + "question_id": "MZJ9D", + "video_file": "Charades/videos/MZJ9D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'that person opens a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "MZJ9D", + "question_id": "MZJ9D", + "video_file": "Charades/videos/MZJ9D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person awhile they partially close the laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <12.8>." + } + ] + }, + { + "video_id": "MZJ9D", + "question_id": "MZJ9D", + "video_file": "Charades/videos/MZJ9D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'there is a person sitting on a couch' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "MZJ9D", + "question_id": "MZJ9D", + "video_file": "Charades/videos/MZJ9D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person closes their laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <12.8>." + } + ] + }, + { + "video_id": "4GCPM", + "question_id": "4GCPM", + "video_file": "Charades/videos/4GCPM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes food from the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.0> to <11.8>." + } + ] + }, + { + "video_id": "ZL61Y", + "question_id": "ZL61Y", + "video_file": "Charades/videos/ZL61Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person the puts the homework on the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.8> to <6.9>." + } + ] + }, + { + "video_id": "1C6P3", + "question_id": "1C6P3", + "video_file": "Charades/videos/1C6P3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'an older person who is eating a sandwich watches him'." + }, + { + "from": "gpt", + "value": "From <0.7> to <8.6>." + } + ] + }, + { + "video_id": "23Y3I", + "question_id": "23Y3I", + "video_file": "Charades/videos/23Y3I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person putting a towel over a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <14.3>." + } + ] + }, + { + "video_id": "23Y3I", + "question_id": "23Y3I", + "video_file": "Charades/videos/23Y3I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.4> to <25.5>." + } + ] + }, + { + "video_id": "8Z02S", + "question_id": "8Z02S", + "video_file": "Charades/videos/8Z02S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is reading a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "8Z02S", + "question_id": "8Z02S", + "video_file": "Charades/videos/8Z02S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person opens the door to take out the clothes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <23.7>." + } + ] + }, + { + "video_id": "7VHXG", + "question_id": "7VHXG", + "video_file": "Charades/videos/7VHXG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is cooking something on the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <15.5>." + } + ] + }, + { + "video_id": "FU5NY", + "question_id": "FU5NY", + "video_file": "Charades/videos/FU5NY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throw the box on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.5> to <12.1>." + } + ] + }, + { + "video_id": "FU5NY", + "question_id": "FU5NY", + "video_file": "Charades/videos/FU5NY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws a box onto the floor'." + }, + { + "from": "gpt", + "value": "From <3.5> to <12.1>." + } + ] + }, + { + "video_id": "XQWLY", + "question_id": "XQWLY", + "video_file": "Charades/videos/XQWLY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person is putting it on a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <12.1>." + } + ] + }, + { + "video_id": "XQWLY", + "question_id": "XQWLY", + "video_file": "Charades/videos/XQWLY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is putting a book on a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <12.1>." + } + ] + }, + { + "video_id": "XQWLY", + "question_id": "XQWLY", + "video_file": "Charades/videos/XQWLY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the other book up'." + }, + { + "from": "gpt", + "value": "From <7.3> to <12.1>." + } + ] + }, + { + "video_id": "QJXNG", + "question_id": "QJXNG", + "video_file": "Charades/videos/QJXNG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts it into a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <17.5>." + } + ] + }, + { + "video_id": "QJXNG", + "question_id": "QJXNG", + "video_file": "Charades/videos/QJXNG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put their phone in it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <17.2>." + } + ] + }, + { + "video_id": "QJXNG", + "question_id": "QJXNG", + "video_file": "Charades/videos/QJXNG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person sits on their bed'." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.6>." + } + ] + }, + { + "video_id": "QJXNG", + "question_id": "QJXNG", + "video_file": "Charades/videos/QJXNG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person plays a game on their phone'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.5>." + } + ] + }, + { + "video_id": "3TK88", + "question_id": "3TK88", + "video_file": "Charades/videos/3TK88.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.1> to <15.3>." + } + ] + }, + { + "video_id": "3TK88", + "question_id": "3TK88", + "video_file": "Charades/videos/3TK88.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns on a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.1> to <15.3>." + } + ] + }, + { + "video_id": "3TK88", + "question_id": "3TK88", + "video_file": "Charades/videos/3TK88.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <37.5> to <45.4>." + } + ] + }, + { + "video_id": "ZGTP8", + "question_id": "ZGTP8", + "video_file": "Charades/videos/ZGTP8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begin sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.9> to <10.0>." + } + ] + }, + { + "video_id": "ZGTP8", + "question_id": "ZGTP8", + "video_file": "Charades/videos/ZGTP8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sitting in a stairwell opens a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "VVCK0", + "question_id": "VVCK0", + "video_file": "Charades/videos/VVCK0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is eating beside the stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <13.7>." + } + ] + }, + { + "video_id": "K34BE", + "question_id": "K34BE", + "video_file": "Charades/videos/K34BE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a broom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.8> to <18.0>." + } + ] + }, + { + "video_id": "U1LFT", + "question_id": "U1LFT", + "video_file": "Charades/videos/U1LFT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts it in a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <23.2>." + } + ] + }, + { + "video_id": "U1LFT", + "question_id": "U1LFT", + "video_file": "Charades/videos/U1LFT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts it in the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <23.2>." + } + ] + }, + { + "video_id": "70SEI", + "question_id": "70SEI", + "video_file": "Charades/videos/70SEI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting at the table drinking soda' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.8>." + } + ] + }, + { + "video_id": "JCWVV", + "question_id": "JCWVV", + "video_file": "Charades/videos/JCWVV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits on a couch' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.0> to <13.4>." + } + ] + }, + { + "video_id": "JCWVV", + "question_id": "JCWVV", + "video_file": "Charades/videos/JCWVV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the book on the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "JCWVV", + "question_id": "JCWVV", + "video_file": "Charades/videos/JCWVV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks into a room holding a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "W98IU", + "question_id": "W98IU", + "video_file": "Charades/videos/W98IU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person looking out a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.4>." + } + ] + }, + { + "video_id": "FJRGJ", + "question_id": "FJRGJ", + "video_file": "Charades/videos/FJRGJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'in a rush a person quickly opens the pantry door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <10.5>." + } + ] + }, + { + "video_id": "FJRGJ", + "question_id": "FJRGJ", + "video_file": "Charades/videos/FJRGJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put them inside the pantry on the shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.8> to <16.9>." + } + ] + }, + { + "video_id": "1NVWD", + "question_id": "1NVWD", + "video_file": "Charades/videos/1NVWD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the phone in a purse' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.7> to <15.7>." + } + ] + }, + { + "video_id": "1NVWD", + "question_id": "1NVWD", + "video_file": "Charades/videos/1NVWD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put the towel on a table'." + }, + { + "from": "gpt", + "value": "From <9.4> to <14.8>." + } + ] + }, + { + "video_id": "1NVWD", + "question_id": "1NVWD", + "video_file": "Charades/videos/1NVWD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is running with the phone in the bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.3> to <22.5>." + } + ] + }, + { + "video_id": "KVGRN", + "question_id": "KVGRN", + "video_file": "Charades/videos/KVGRN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looks out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "Q0BNP", + "question_id": "Q0BNP", + "video_file": "Charades/videos/Q0BNP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts to get undressed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.9> to <26.2>." + } + ] + }, + { + "video_id": "Q0BNP", + "question_id": "Q0BNP", + "video_file": "Charades/videos/Q0BNP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinks from a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.8> to <32.0>." + } + ] + }, + { + "video_id": "EMV0A", + "question_id": "EMV0A", + "video_file": "Charades/videos/EMV0A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <18.7>." + } + ] + }, + { + "video_id": "EMV0A", + "question_id": "EMV0A", + "video_file": "Charades/videos/EMV0A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they cook'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "EMV0A", + "question_id": "EMV0A", + "video_file": "Charades/videos/EMV0A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "EMV0A", + "question_id": "EMV0A", + "video_file": "Charades/videos/EMV0A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sneezing by the stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "RJQL6", + "question_id": "RJQL6", + "video_file": "Charades/videos/RJQL6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes a towel off the shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.9> to <12.5>." + } + ] + }, + { + "video_id": "OI3BC", + "question_id": "OI3BC", + "video_file": "Charades/videos/OI3BC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person close the door of a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.4> to <19.8>." + } + ] + }, + { + "video_id": "4VCHD", + "question_id": "4VCHD", + "video_file": "Charades/videos/4VCHD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person is eating' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.3> to <14.2>." + } + ] + }, + { + "video_id": "4VCHD", + "question_id": "4VCHD", + "video_file": "Charades/videos/4VCHD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <7.7> to <15.1>." + } + ] + }, + { + "video_id": "4VCHD", + "question_id": "4VCHD", + "video_file": "Charades/videos/4VCHD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens the door to the recreation room'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.2>." + } + ] + }, + { + "video_id": "4VCHD", + "question_id": "4VCHD", + "video_file": "Charades/videos/4VCHD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person sits in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.8> to <11.3>." + } + ] + }, + { + "video_id": "D9EDE", + "question_id": "D9EDE", + "video_file": "Charades/videos/D9EDE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is holding a bag of groceries'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "NMTRI", + "question_id": "NMTRI", + "video_file": "Charades/videos/NMTRI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person eats a cookie'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <22.5>." + } + ] + }, + { + "video_id": "NMTRI", + "question_id": "NMTRI", + "video_file": "Charades/videos/NMTRI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a drink from a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.0> to <31.4>." + } + ] + }, + { + "video_id": "NMTRI", + "question_id": "NMTRI", + "video_file": "Charades/videos/NMTRI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person drinks from a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.0> to <31.4>." + } + ] + }, + { + "video_id": "NMTRI", + "question_id": "NMTRI", + "video_file": "Charades/videos/NMTRI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats it'." + }, + { + "from": "gpt", + "value": "From <17.7> to <22.5>." + } + ] + }, + { + "video_id": "NMTRI", + "question_id": "NMTRI", + "video_file": "Charades/videos/NMTRI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakens in bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "NMTRI", + "question_id": "NMTRI", + "video_file": "Charades/videos/NMTRI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eat something'." + }, + { + "from": "gpt", + "value": "From <17.7> to <22.5>." + } + ] + }, + { + "video_id": "6OROW", + "question_id": "6OROW", + "video_file": "Charades/videos/6OROW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sitting in a chair writing in a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.0> to <21.3>." + } + ] + }, + { + "video_id": "6OROW", + "question_id": "6OROW", + "video_file": "Charades/videos/6OROW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting on a chair opening a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <21.3>." + } + ] + }, + { + "video_id": "6OROW", + "question_id": "6OROW", + "video_file": "Charades/videos/6OROW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take a glass of something'." + }, + { + "from": "gpt", + "value": "From <18.8> to <24.1>." + } + ] + }, + { + "video_id": "6OROW", + "question_id": "6OROW", + "video_file": "Charades/videos/6OROW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking coffee out of glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.8> to <33.0>." + } + ] + }, + { + "video_id": "6YGBD", + "question_id": "6YGBD", + "video_file": "Charades/videos/6YGBD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking a glass of water with the other hand'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.2>." + } + ] + }, + { + "video_id": "6YGBD", + "question_id": "6YGBD", + "video_file": "Charades/videos/6YGBD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is holding a sandwich in one hand'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.4>." + } + ] + }, + { + "video_id": "6YGBD", + "question_id": "6YGBD", + "video_file": "Charades/videos/6YGBD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking out of glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.2>." + } + ] + }, + { + "video_id": "6YGBD", + "question_id": "6YGBD", + "video_file": "Charades/videos/6YGBD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they are holding a sandwich in their hand' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.4>." + } + ] + }, + { + "video_id": "BSDSA", + "question_id": "BSDSA", + "video_file": "Charades/videos/BSDSA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.0> to <29.0>." + } + ] + }, + { + "video_id": "6MJGP", + "question_id": "6MJGP", + "video_file": "Charades/videos/6MJGP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "5MRPU", + "question_id": "5MRPU", + "video_file": "Charades/videos/5MRPU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take a picture of themselves' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.0> to <34.4>." + } + ] + }, + { + "video_id": "U43FI", + "question_id": "U43FI", + "video_file": "Charades/videos/U43FI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes the glass from the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.6> to <19.0>." + } + ] + }, + { + "video_id": "3X6SI", + "question_id": "3X6SI", + "video_file": "Charades/videos/3X6SI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits on a chair next to a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <16.0>." + } + ] + }, + { + "video_id": "6XT4W", + "question_id": "6XT4W", + "video_file": "Charades/videos/6XT4W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person looks at a picture nearby' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.1> to <24.9>." + } + ] + }, + { + "video_id": "WJOR3", + "question_id": "WJOR3", + "video_file": "Charades/videos/WJOR3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sitting down watching a laptop sitting on the desk' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "JVXMI", + "question_id": "JVXMI", + "video_file": "Charades/videos/JVXMI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person takes out their phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <4.7>." + } + ] + }, + { + "video_id": "JVXMI", + "question_id": "JVXMI", + "video_file": "Charades/videos/JVXMI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a picture'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <19.0>." + } + ] + }, + { + "video_id": "JVXMI", + "question_id": "JVXMI", + "video_file": "Charades/videos/JVXMI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person in bedroom takes out phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.1> to <4.7>." + } + ] + }, + { + "video_id": "RO3A2", + "question_id": "RO3A2", + "video_file": "Charades/videos/RO3A2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs into the kitchen' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "RO3A2", + "question_id": "RO3A2", + "video_file": "Charades/videos/RO3A2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins washing dishes in the sink'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <12.5>." + } + ] + }, + { + "video_id": "65IQ0", + "question_id": "65IQ0", + "video_file": "Charades/videos/65IQ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throwing things on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.6> to <20.9>." + } + ] + }, + { + "video_id": "J4QIT", + "question_id": "J4QIT", + "video_file": "Charades/videos/J4QIT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open the book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "8EYHC", + "question_id": "8EYHC", + "video_file": "Charades/videos/8EYHC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.0> to <22.4>." + } + ] + }, + { + "video_id": "8EYHC", + "question_id": "8EYHC", + "video_file": "Charades/videos/8EYHC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.0> to <22.4>." + } + ] + }, + { + "video_id": "NC75G", + "question_id": "NC75G", + "video_file": "Charades/videos/NC75G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes a towel from the shelf'." + }, + { + "from": "gpt", + "value": "From <3.7> to <10.1>." + } + ] + }, + { + "video_id": "XDP02", + "question_id": "XDP02", + "video_file": "Charades/videos/XDP02.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'awakening another person with a pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <13.7>." + } + ] + }, + { + "video_id": "W1SRN", + "question_id": "W1SRN", + "video_file": "Charades/videos/W1SRN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws it on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <19.7>." + } + ] + }, + { + "video_id": "W1SRN", + "question_id": "W1SRN", + "video_file": "Charades/videos/W1SRN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws a blanket on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.0> to <19.7>." + } + ] + }, + { + "video_id": "W1SRN", + "question_id": "W1SRN", + "video_file": "Charades/videos/W1SRN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begin snuggling with the blanket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.5> to <33.0>." + } + ] + }, + { + "video_id": "W1SRN", + "question_id": "W1SRN", + "video_file": "Charades/videos/W1SRN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the cabinet'." + }, + { + "from": "gpt", + "value": "From <6.4> to <12.2>." + } + ] + }, + { + "video_id": "W1SRN", + "question_id": "W1SRN", + "video_file": "Charades/videos/W1SRN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person throws the blanket on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <19.7>." + } + ] + }, + { + "video_id": "W1SRN", + "question_id": "W1SRN", + "video_file": "Charades/videos/W1SRN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person turns on a light in the hallway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <6.9>." + } + ] + }, + { + "video_id": "W1SRN", + "question_id": "W1SRN", + "video_file": "Charades/videos/W1SRN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins snuggling the blanket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.5> to <33.0>." + } + ] + }, + { + "video_id": "64SL1", + "question_id": "64SL1", + "video_file": "Charades/videos/64SL1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person laughing grab a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.9> to <13.7>." + } + ] + }, + { + "video_id": "3ONCH", + "question_id": "3ONCH", + "video_file": "Charades/videos/3ONCH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts them in a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.5> to <22.2>." + } + ] + }, + { + "video_id": "3ONCH", + "question_id": "3ONCH", + "video_file": "Charades/videos/3ONCH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they are seen putting their shoes into a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.5> to <22.2>." + } + ] + }, + { + "video_id": "IJT65", + "question_id": "IJT65", + "video_file": "Charades/videos/IJT65.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sneezing on food in the laundry room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <16.6>." + } + ] + }, + { + "video_id": "ZP53M", + "question_id": "ZP53M", + "video_file": "Charades/videos/ZP53M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts a boo on table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.4> to <19.4>." + } + ] + }, + { + "video_id": "1ZAN8", + "question_id": "1ZAN8", + "video_file": "Charades/videos/1ZAN8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.0> to <32.0>." + } + ] + }, + { + "video_id": "1ZAN8", + "question_id": "1ZAN8", + "video_file": "Charades/videos/1ZAN8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closed a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.0> to <32.0>." + } + ] + }, + { + "video_id": "1ZAN8", + "question_id": "1ZAN8", + "video_file": "Charades/videos/1ZAN8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person left the room closing the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.0> to <32.0>." + } + ] + }, + { + "video_id": "XE4FF", + "question_id": "XE4FF", + "video_file": "Charades/videos/XE4FF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the shoes they were holding on a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.5> to <14.8>." + } + ] + }, + { + "video_id": "FP7N3", + "question_id": "FP7N3", + "video_file": "Charades/videos/FP7N3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <5.1> to <10.4>." + } + ] + }, + { + "video_id": "FP7N3", + "question_id": "FP7N3", + "video_file": "Charades/videos/FP7N3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put the cup on the counter' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.9> to <27.8>." + } + ] + }, + { + "video_id": "FP7N3", + "question_id": "FP7N3", + "video_file": "Charades/videos/FP7N3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pours some water into the cup'." + }, + { + "from": "gpt", + "value": "From <10.6> to <22.5>." + } + ] + }, + { + "video_id": "FP7N3", + "question_id": "FP7N3", + "video_file": "Charades/videos/FP7N3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the door of a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.6> to <10.5>." + } + ] + }, + { + "video_id": "78LXZ", + "question_id": "78LXZ", + "video_file": "Charades/videos/78LXZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person takes a stack of clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.2> to <17.0>." + } + ] + }, + { + "video_id": "TCQ97", + "question_id": "TCQ97", + "video_file": "Charades/videos/TCQ97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a second person takes clothes off a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <13.8>." + } + ] + }, + { + "video_id": "TCQ97", + "question_id": "TCQ97", + "video_file": "Charades/videos/TCQ97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person puts clothes into a closet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.5> to <19.7>." + } + ] + }, + { + "video_id": "TCQ97", + "question_id": "TCQ97", + "video_file": "Charades/videos/TCQ97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits on a couch'." + }, + { + "from": "gpt", + "value": "From <5.8> to <14.4>." + } + ] + }, + { + "video_id": "TCQ97", + "question_id": "TCQ97", + "video_file": "Charades/videos/TCQ97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person throws a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <7.4>." + } + ] + }, + { + "video_id": "TCQ97", + "question_id": "TCQ97", + "video_file": "Charades/videos/TCQ97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits on a sofa'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.8> to <14.4>." + } + ] + }, + { + "video_id": "XXU6H", + "question_id": "XXU6H", + "video_file": "Charades/videos/XXU6H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <6.7>." + } + ] + }, + { + "video_id": "XXU6H", + "question_id": "XXU6H", + "video_file": "Charades/videos/XXU6H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a doorway to get into the house' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "VN8SW", + "question_id": "VN8SW", + "video_file": "Charades/videos/VN8SW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put on some shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.9> to <33.0>." + } + ] + }, + { + "video_id": "VN8SW", + "question_id": "VN8SW", + "video_file": "Charades/videos/VN8SW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting on blue tennis shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.9> to <33.0>." + } + ] + }, + { + "video_id": "CO4AU", + "question_id": "CO4AU", + "video_file": "Charades/videos/CO4AU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person is holding a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.9>." + } + ] + }, + { + "video_id": "SMS5G", + "question_id": "SMS5G", + "video_file": "Charades/videos/SMS5G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throwing clothes under the doorway in the entryway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.8> to <27.1>." + } + ] + }, + { + "video_id": "M4FW0", + "question_id": "M4FW0", + "video_file": "Charades/videos/M4FW0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person walks to the desk holding a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "M4FW0", + "question_id": "M4FW0", + "video_file": "Charades/videos/M4FW0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding a bag of groceries' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "3ZC4Y", + "question_id": "3ZC4Y", + "video_file": "Charades/videos/3ZC4Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'that same person opens the refrigerator door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.1> to <21.7>." + } + ] + }, + { + "video_id": "3ZC4Y", + "question_id": "3ZC4Y", + "video_file": "Charades/videos/3ZC4Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.1> to <21.7>." + } + ] + }, + { + "video_id": "MOWLI", + "question_id": "MOWLI", + "video_file": "Charades/videos/MOWLI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening a door they go in the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.9> to <11.3>." + } + ] + }, + { + "video_id": "XXX1O", + "question_id": "XXX1O", + "video_file": "Charades/videos/XXX1O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person sits at a table'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.7>." + } + ] + }, + { + "video_id": "XXX1O", + "question_id": "XXX1O", + "video_file": "Charades/videos/XXX1O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sitting on the sofa'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <31.0>." + } + ] + }, + { + "video_id": "XXX1O", + "question_id": "XXX1O", + "video_file": "Charades/videos/XXX1O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take some food out of a bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "XXX1O", + "question_id": "XXX1O", + "video_file": "Charades/videos/XXX1O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eat it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <13.1>." + } + ] + }, + { + "video_id": "XXX1O", + "question_id": "XXX1O", + "video_file": "Charades/videos/XXX1O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pouring a drink into a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.9> to <20.0>." + } + ] + }, + { + "video_id": "XXX1O", + "question_id": "XXX1O", + "video_file": "Charades/videos/XXX1O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits on the couch using his laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.2> to <31.0>." + } + ] + }, + { + "video_id": "XXX1O", + "question_id": "XXX1O", + "video_file": "Charades/videos/XXX1O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pour some water into a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <20.0>." + } + ] + }, + { + "video_id": "XXX1O", + "question_id": "XXX1O", + "video_file": "Charades/videos/XXX1O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is eating out of a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <13.1>." + } + ] + }, + { + "video_id": "XXX1O", + "question_id": "XXX1O", + "video_file": "Charades/videos/XXX1O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <15.7>." + } + ] + }, + { + "video_id": "38T9C", + "question_id": "38T9C", + "video_file": "Charades/videos/38T9C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'there is a person eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.2> to <13.8>." + } + ] + }, + { + "video_id": "38T9C", + "question_id": "38T9C", + "video_file": "Charades/videos/38T9C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'that same person puts the sandwich on a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <14.2>." + } + ] + }, + { + "video_id": "38T9C", + "question_id": "38T9C", + "video_file": "Charades/videos/38T9C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person eats the sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.2> to <13.8>." + } + ] + }, + { + "video_id": "SCCZG", + "question_id": "SCCZG", + "video_file": "Charades/videos/SCCZG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they start watching themselves in the mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.3> to <31.9>." + } + ] + }, + { + "video_id": "YBVUE", + "question_id": "YBVUE", + "video_file": "Charades/videos/YBVUE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is undressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.9>." + } + ] + }, + { + "video_id": "YBVUE", + "question_id": "YBVUE", + "video_file": "Charades/videos/YBVUE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting their clothes on the table'." + }, + { + "from": "gpt", + "value": "From <17.2> to <31.0>." + } + ] + }, + { + "video_id": "LD8PU", + "question_id": "LD8PU", + "video_file": "Charades/videos/LD8PU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs through the kitchen doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <7.9>." + } + ] + }, + { + "video_id": "JVJZX", + "question_id": "JVJZX", + "video_file": "Charades/videos/JVJZX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person closes the laptop'." + }, + { + "from": "gpt", + "value": "From <12.9> to <18.4>." + } + ] + }, + { + "video_id": "IGQJP", + "question_id": "IGQJP", + "video_file": "Charades/videos/IGQJP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <17.9>." + } + ] + }, + { + "video_id": "7GPSY", + "question_id": "7GPSY", + "video_file": "Charades/videos/7GPSY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.2> to <32.7>." + } + ] + }, + { + "video_id": "2E8GU", + "question_id": "2E8GU", + "video_file": "Charades/videos/2E8GU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a broom by the desk' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "9PXC0", + "question_id": "9PXC0", + "video_file": "Charades/videos/9PXC0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they open up the box'." + }, + { + "from": "gpt", + "value": "From <11.2> to <19.1>." + } + ] + }, + { + "video_id": "0GYRH", + "question_id": "0GYRH", + "video_file": "Charades/videos/0GYRH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a bag of groceries' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.6> to <29.4>." + } + ] + }, + { + "video_id": "JQ2SN", + "question_id": "JQ2SN", + "video_file": "Charades/videos/JQ2SN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting on a bed eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <13.1>." + } + ] + }, + { + "video_id": "JQ2SN", + "question_id": "JQ2SN", + "video_file": "Charades/videos/JQ2SN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.4> to <19.5>." + } + ] + }, + { + "video_id": "JQ2SN", + "question_id": "JQ2SN", + "video_file": "Charades/videos/JQ2SN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <13.1>." + } + ] + }, + { + "video_id": "JQ2SN", + "question_id": "JQ2SN", + "video_file": "Charades/videos/JQ2SN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put the sandwich down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <17.2>." + } + ] + }, + { + "video_id": "E7TLO", + "question_id": "E7TLO", + "video_file": "Charades/videos/E7TLO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person inside turns around a takes a picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.9> to <32.0>." + } + ] + }, + { + "video_id": "E7TLO", + "question_id": "E7TLO", + "video_file": "Charades/videos/E7TLO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a picture'." + }, + { + "from": "gpt", + "value": "From <24.9> to <32.0>." + } + ] + }, + { + "video_id": "X5YL3", + "question_id": "X5YL3", + "video_file": "Charades/videos/X5YL3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person holding a bag opens the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.8>." + } + ] + }, + { + "video_id": "X5YL3", + "question_id": "X5YL3", + "video_file": "Charades/videos/X5YL3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person holding a bag throws food into a refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.8>." + } + ] + }, + { + "video_id": "X5YL3", + "question_id": "X5YL3", + "video_file": "Charades/videos/X5YL3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person with bags in their hands opens the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.7>." + } + ] + }, + { + "video_id": "KW4Y1", + "question_id": "KW4Y1", + "video_file": "Charades/videos/KW4Y1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.1>." + } + ] + }, + { + "video_id": "1WI77", + "question_id": "1WI77", + "video_file": "Charades/videos/1WI77.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <9.8>." + } + ] + }, + { + "video_id": "1WI77", + "question_id": "1WI77", + "video_file": "Charades/videos/1WI77.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <9.8>." + } + ] + }, + { + "video_id": "OHS4T", + "question_id": "OHS4T", + "video_file": "Charades/videos/OHS4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits on a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <31.0>." + } + ] + }, + { + "video_id": "OHS4T", + "question_id": "OHS4T", + "video_file": "Charades/videos/OHS4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits down in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <31.0>." + } + ] + }, + { + "video_id": "OHS4T", + "question_id": "OHS4T", + "video_file": "Charades/videos/OHS4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.4> to <7.2>." + } + ] + }, + { + "video_id": "OHS4T", + "question_id": "OHS4T", + "video_file": "Charades/videos/OHS4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.4> to <7.2>." + } + ] + }, + { + "video_id": "OHS4T", + "question_id": "OHS4T", + "video_file": "Charades/videos/OHS4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs into a bathroom'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "OHS4T", + "question_id": "OHS4T", + "video_file": "Charades/videos/OHS4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs into the bathroom'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "Z1ZUI", + "question_id": "Z1ZUI", + "video_file": "Charades/videos/Z1ZUI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they stand up to grab a blanket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.5> to <12.0>." + } + ] + }, + { + "video_id": "7WIRG", + "question_id": "7WIRG", + "video_file": "Charades/videos/7WIRG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.1> to <12.0>." + } + ] + }, + { + "video_id": "Z6I0R", + "question_id": "Z6I0R", + "video_file": "Charades/videos/Z6I0R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person walks through a doorwaywhile using their phone'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "Z6I0R", + "question_id": "Z6I0R", + "video_file": "Charades/videos/Z6I0R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person the cup was put down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.4> to <20.6>." + } + ] + }, + { + "video_id": "Z6I0R", + "question_id": "Z6I0R", + "video_file": "Charades/videos/Z6I0R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes a drink from a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.8> to <16.7>." + } + ] + }, + { + "video_id": "Z6I0R", + "question_id": "Z6I0R", + "video_file": "Charades/videos/Z6I0R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person playing with their phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "3SNSC", + "question_id": "3SNSC", + "video_file": "Charades/videos/3SNSC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take medicine from a pill bottle'." + }, + { + "from": "gpt", + "value": "From <22.2> to <29.5>." + } + ] + }, + { + "video_id": "3SNSC", + "question_id": "3SNSC", + "video_file": "Charades/videos/3SNSC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the book down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.8> to <21.9>." + } + ] + }, + { + "video_id": "IEXTB", + "question_id": "IEXTB", + "video_file": "Charades/videos/IEXTB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person laughing at a picture hanging on the wall' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <13.0>." + } + ] + }, + { + "video_id": "D09K3", + "question_id": "D09K3", + "video_file": "Charades/videos/D09K3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they begin opening a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.0> to <30.0>." + } + ] + }, + { + "video_id": "NBIR0", + "question_id": "NBIR0", + "video_file": "Charades/videos/NBIR0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person undresses' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <14.4>." + } + ] + }, + { + "video_id": "NBIR0", + "question_id": "NBIR0", + "video_file": "Charades/videos/NBIR0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person which they close the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.9> to <23.0>." + } + ] + }, + { + "video_id": "F9YMU", + "question_id": "F9YMU", + "video_file": "Charades/videos/F9YMU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person closes a door in a room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.1>." + } + ] + }, + { + "video_id": "F9YMU", + "question_id": "F9YMU", + "video_file": "Charades/videos/F9YMU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they close the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.1>." + } + ] + }, + { + "video_id": "F9YMU", + "question_id": "F9YMU", + "video_file": "Charades/videos/F9YMU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is closing the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.1>." + } + ] + }, + { + "video_id": "FXRMG", + "question_id": "FXRMG", + "video_file": "Charades/videos/FXRMG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person takes a book from their desk' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "FXRMG", + "question_id": "FXRMG", + "video_file": "Charades/videos/FXRMG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a door'." + }, + { + "from": "gpt", + "value": "From <22.9> to <28.1>." + } + ] + }, + { + "video_id": "FXRMG", + "question_id": "FXRMG", + "video_file": "Charades/videos/FXRMG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.6> to <29.3>." + } + ] + }, + { + "video_id": "7KAVM", + "question_id": "7KAVM", + "video_file": "Charades/videos/7KAVM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking from a bottle' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.1> to <13.7>." + } + ] + }, + { + "video_id": "LOH86", + "question_id": "LOH86", + "video_file": "Charades/videos/LOH86.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the camera down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "LOH86", + "question_id": "LOH86", + "video_file": "Charades/videos/LOH86.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a camera' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.3>." + } + ] + }, + { + "video_id": "LOH86", + "question_id": "LOH86", + "video_file": "Charades/videos/LOH86.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person plays with the camera' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.8>." + } + ] + }, + { + "video_id": "LOH86", + "question_id": "LOH86", + "video_file": "Charades/videos/LOH86.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks from a doorway to a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "LOH86", + "question_id": "LOH86", + "video_file": "Charades/videos/LOH86.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'i person is putting their camera on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <11.9>." + } + ] + }, + { + "video_id": "LOH86", + "question_id": "LOH86", + "video_file": "Charades/videos/LOH86.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is walking in through the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "LOH86", + "question_id": "LOH86", + "video_file": "Charades/videos/LOH86.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person playing with a camera in their hand' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.8>." + } + ] + }, + { + "video_id": "LFPLP", + "question_id": "LFPLP", + "video_file": "Charades/videos/LFPLP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens the refrigerator door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.8> to <16.2>." + } + ] + }, + { + "video_id": "LFPLP", + "question_id": "LFPLP", + "video_file": "Charades/videos/LFPLP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is cooking on the stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.1>." + } + ] + }, + { + "video_id": "LFPLP", + "question_id": "LFPLP", + "video_file": "Charades/videos/LFPLP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening the refrigerator door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <16.2>." + } + ] + }, + { + "video_id": "LFPLP", + "question_id": "LFPLP", + "video_file": "Charades/videos/LFPLP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is cooking' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.1>." + } + ] + }, + { + "video_id": "GXZHG", + "question_id": "GXZHG", + "video_file": "Charades/videos/GXZHG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person run to the couch' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <14.7>." + } + ] + }, + { + "video_id": "GXZHG", + "question_id": "GXZHG", + "video_file": "Charades/videos/GXZHG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts running to the sofa'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <14.7>." + } + ] + }, + { + "video_id": "1DL0E", + "question_id": "1DL0E", + "video_file": "Charades/videos/1DL0E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they pull out a camera to take a picture'." + }, + { + "from": "gpt", + "value": "From <19.7> to <32.0>." + } + ] + }, + { + "video_id": "1DL0E", + "question_id": "1DL0E", + "video_file": "Charades/videos/1DL0E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.7> to <32.0>." + } + ] + }, + { + "video_id": "7POUA", + "question_id": "7POUA", + "video_file": "Charades/videos/7POUA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begins to fix their hair in a mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <32.0>." + } + ] + }, + { + "video_id": "LJS9V", + "question_id": "LJS9V", + "video_file": "Charades/videos/LJS9V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting in a hallway smiling' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "LJS9V", + "question_id": "LJS9V", + "video_file": "Charades/videos/LJS9V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person smiling at the television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "8NGDR", + "question_id": "8NGDR", + "video_file": "Charades/videos/8NGDR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts watching a television program'." + }, + { + "from": "gpt", + "value": "From <32.0> to <44.0>." + } + ] + }, + { + "video_id": "8NGDR", + "question_id": "8NGDR", + "video_file": "Charades/videos/8NGDR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakening from a nap on a sofa' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.6> to <20.1>." + } + ] + }, + { + "video_id": "29B3I", + "question_id": "29B3I", + "video_file": "Charades/videos/29B3I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a broom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <16.1>." + } + ] + }, + { + "video_id": "29B3I", + "question_id": "29B3I", + "video_file": "Charades/videos/29B3I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person is holding a broom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <16.1>." + } + ] + }, + { + "video_id": "29B3I", + "question_id": "29B3I", + "video_file": "Charades/videos/29B3I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is holding a broom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <16.1>." + } + ] + }, + { + "video_id": "3CGO0", + "question_id": "3CGO0", + "video_file": "Charades/videos/3CGO0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person tidying some clothes in a box'." + }, + { + "from": "gpt", + "value": "From <11.1> to <16.2>." + } + ] + }, + { + "video_id": "670G1", + "question_id": "670G1", + "video_file": "Charades/videos/670G1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person is opening the cabinet door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "670G1", + "question_id": "670G1", + "video_file": "Charades/videos/670G1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "1ERM9", + "question_id": "1ERM9", + "video_file": "Charades/videos/1ERM9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes a picture of themselves' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.2> to <25.6>." + } + ] + }, + { + "video_id": "1ERM9", + "question_id": "1ERM9", + "video_file": "Charades/videos/1ERM9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is in the bathroom washing their hands' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "1ERM9", + "question_id": "1ERM9", + "video_file": "Charades/videos/1ERM9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is washing their hands in a sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "1ERM9", + "question_id": "1ERM9", + "video_file": "Charades/videos/1ERM9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a camera from a box'." + }, + { + "from": "gpt", + "value": "From <10.0> to <18.5>." + } + ] + }, + { + "video_id": "1ERM9", + "question_id": "1ERM9", + "video_file": "Charades/videos/1ERM9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a picture of them self'." + }, + { + "from": "gpt", + "value": "From <16.2> to <25.6>." + } + ] + }, + { + "video_id": "DUEEE", + "question_id": "DUEEE", + "video_file": "Charades/videos/DUEEE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the phone back into their pocket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <25.9>." + } + ] + }, + { + "video_id": "PCXYE", + "question_id": "PCXYE", + "video_file": "Charades/videos/PCXYE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person drinks from a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <14.3>." + } + ] + }, + { + "video_id": "PCXYE", + "question_id": "PCXYE", + "video_file": "Charades/videos/PCXYE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking from a white cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <14.3>." + } + ] + }, + { + "video_id": "D4RH8", + "question_id": "D4RH8", + "video_file": "Charades/videos/D4RH8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.6> to <30.0>." + } + ] + }, + { + "video_id": "D4RH8", + "question_id": "D4RH8", + "video_file": "Charades/videos/D4RH8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating a sandwich,' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <30.0>." + } + ] + }, + { + "video_id": "XPWYQ", + "question_id": "XPWYQ", + "video_file": "Charades/videos/XPWYQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put the books on a coffee table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <22.9>." + } + ] + }, + { + "video_id": "WH1S5", + "question_id": "WH1S5", + "video_file": "Charades/videos/WH1S5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is walking into the kitchen eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.4> to <25.2>." + } + ] + }, + { + "video_id": "0HD9F", + "question_id": "0HD9F", + "video_file": "Charades/videos/0HD9F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person close some cabinet doors'." + }, + { + "from": "gpt", + "value": "From <10.1> to <18.3>." + } + ] + }, + { + "video_id": "GL7ER", + "question_id": "GL7ER", + "video_file": "Charades/videos/GL7ER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begins eating' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.5> to <31.0>." + } + ] + }, + { + "video_id": "GL7ER", + "question_id": "GL7ER", + "video_file": "Charades/videos/GL7ER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts their food on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.6> to <8.4>." + } + ] + }, + { + "video_id": "GL7ER", + "question_id": "GL7ER", + "video_file": "Charades/videos/GL7ER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins to eat them' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.5> to <31.0>." + } + ] + }, + { + "video_id": "83D9H", + "question_id": "83D9H", + "video_file": "Charades/videos/83D9H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walks to a partially open door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.5> to <22.7>." + } + ] + }, + { + "video_id": "83D9H", + "question_id": "83D9H", + "video_file": "Charades/videos/83D9H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.5> to <22.7>." + } + ] + }, + { + "video_id": "PKRP5", + "question_id": "PKRP5", + "video_file": "Charades/videos/PKRP5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits down in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <11.0>." + } + ] + }, + { + "video_id": "1UXVA", + "question_id": "1UXVA", + "video_file": "Charades/videos/1UXVA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person is sitting down eating food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.3> to <16.7>." + } + ] + }, + { + "video_id": "1UXVA", + "question_id": "1UXVA", + "video_file": "Charades/videos/1UXVA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person drinks a glass of water'." + }, + { + "from": "gpt", + "value": "From <15.0> to <22.6>." + } + ] + }, + { + "video_id": "UYFET", + "question_id": "UYFET", + "video_file": "Charades/videos/UYFET.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turning the light off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.5> to <25.0>." + } + ] + }, + { + "video_id": "UYFET", + "question_id": "UYFET", + "video_file": "Charades/videos/UYFET.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turned off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.5> to <25.0>." + } + ] + }, + { + "video_id": "PKEZI", + "question_id": "PKEZI", + "video_file": "Charades/videos/PKEZI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person stand up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.5> to <40.0>." + } + ] + }, + { + "video_id": "PKEZI", + "question_id": "PKEZI", + "video_file": "Charades/videos/PKEZI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins undressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.1> to <40.0>." + } + ] + }, + { + "video_id": "I7DGO", + "question_id": "I7DGO", + "video_file": "Charades/videos/I7DGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a shot of medicine'." + }, + { + "from": "gpt", + "value": "From <20.2> to <32.0>." + } + ] + }, + { + "video_id": "X9Q1A", + "question_id": "X9Q1A", + "video_file": "Charades/videos/X9Q1A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person puts some clothes into a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.0> to <18.5>." + } + ] + }, + { + "video_id": "X9Q1A", + "question_id": "X9Q1A", + "video_file": "Charades/videos/X9Q1A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the clothing in a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <18.5>." + } + ] + }, + { + "video_id": "5C6IS", + "question_id": "5C6IS", + "video_file": "Charades/videos/5C6IS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take clothes from the bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <15.0>." + } + ] + }, + { + "video_id": "5C6IS", + "question_id": "5C6IS", + "video_file": "Charades/videos/5C6IS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walks away holding the clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.6> to <17.4>." + } + ] + }, + { + "video_id": "SMPEB", + "question_id": "SMPEB", + "video_file": "Charades/videos/SMPEB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.9>." + } + ] + }, + { + "video_id": "SMPEB", + "question_id": "SMPEB", + "video_file": "Charades/videos/SMPEB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person closes a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.9>." + } + ] + }, + { + "video_id": "SMPEB", + "question_id": "SMPEB", + "video_file": "Charades/videos/SMPEB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'this person closes a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.9>." + } + ] + }, + { + "video_id": "8HHI3", + "question_id": "8HHI3", + "video_file": "Charades/videos/8HHI3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person laughs as they open a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.9> to <28.0>." + } + ] + }, + { + "video_id": "8HHI3", + "question_id": "8HHI3", + "video_file": "Charades/videos/8HHI3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a box'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "8HHI3", + "question_id": "8HHI3", + "video_file": "Charades/videos/8HHI3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes some shoes from the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <7.9>." + } + ] + }, + { + "video_id": "DFD3T", + "question_id": "DFD3T", + "video_file": "Charades/videos/DFD3T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is throwing some clothes on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.8>." + } + ] + }, + { + "video_id": "UU5XI", + "question_id": "UU5XI", + "video_file": "Charades/videos/UU5XI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks through the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.2> to <31.0>." + } + ] + }, + { + "video_id": "TDD74", + "question_id": "TDD74", + "video_file": "Charades/videos/TDD74.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins to dress more warmly in the doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.3> to <20.5>." + } + ] + }, + { + "video_id": "TDD74", + "question_id": "TDD74", + "video_file": "Charades/videos/TDD74.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person finish dressing with some of the clothes in it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <20.5>." + } + ] + }, + { + "video_id": "G5G8K", + "question_id": "G5G8K", + "video_file": "Charades/videos/G5G8K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is holding a pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.7>." + } + ] + }, + { + "video_id": "G5G8K", + "question_id": "G5G8K", + "video_file": "Charades/videos/G5G8K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begin smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <22.1>." + } + ] + }, + { + "video_id": "8ASRQ", + "question_id": "8ASRQ", + "video_file": "Charades/videos/8ASRQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throwing a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <17.0>." + } + ] + }, + { + "video_id": "T342T", + "question_id": "T342T", + "video_file": "Charades/videos/T342T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.5> to <31.2>." + } + ] + }, + { + "video_id": "T342T", + "question_id": "T342T", + "video_file": "Charades/videos/T342T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the door of a bathroom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.5> to <31.2>." + } + ] + }, + { + "video_id": "T342T", + "question_id": "T342T", + "video_file": "Charades/videos/T342T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes some medicine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <11.6>." + } + ] + }, + { + "video_id": "I8QJN", + "question_id": "I8QJN", + "video_file": "Charades/videos/I8QJN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person laying on the floor stands' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.3>." + } + ] + }, + { + "video_id": "I8QJN", + "question_id": "I8QJN", + "video_file": "Charades/videos/I8QJN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person awakens on the floor of the kitchen' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.8> to <12.7>." + } + ] + }, + { + "video_id": "I8QJN", + "question_id": "I8QJN", + "video_file": "Charades/videos/I8QJN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person serve themselves a glass of water to drink'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.5> to <30.2>." + } + ] + }, + { + "video_id": "I8QJN", + "question_id": "I8QJN", + "video_file": "Charades/videos/I8QJN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they awaken' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.8> to <12.7>." + } + ] + }, + { + "video_id": "QRHJJ", + "question_id": "QRHJJ", + "video_file": "Charades/videos/QRHJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes a book from a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "OHVMG", + "question_id": "OHVMG", + "video_file": "Charades/videos/OHVMG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person wearing shoes is smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "2ZFNL", + "question_id": "2ZFNL", + "video_file": "Charades/videos/2ZFNL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throwing them on the floor'." + }, + { + "from": "gpt", + "value": "From <2.7> to <7.0>." + } + ] + }, + { + "video_id": "0F453", + "question_id": "0F453", + "video_file": "Charades/videos/0F453.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opens a refrigerator'." + }, + { + "from": "gpt", + "value": "From <0.3> to <7.9>." + } + ] + }, + { + "video_id": "0F453", + "question_id": "0F453", + "video_file": "Charades/videos/0F453.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person is opening refrigerator'." + }, + { + "from": "gpt", + "value": "From <0.3> to <7.9>." + } + ] + }, + { + "video_id": "NZ7VY", + "question_id": "NZ7VY", + "video_file": "Charades/videos/NZ7VY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sit on the sofa'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.8> to <29.7>." + } + ] + }, + { + "video_id": "NZ7VY", + "question_id": "NZ7VY", + "video_file": "Charades/videos/NZ7VY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the thermos into a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <25.6>." + } + ] + }, + { + "video_id": "NZ7VY", + "question_id": "NZ7VY", + "video_file": "Charades/videos/NZ7VY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person pours a cup of coffee from a thermos' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <14.3>." + } + ] + }, + { + "video_id": "NZ7VY", + "question_id": "NZ7VY", + "video_file": "Charades/videos/NZ7VY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the coffee on the shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <25.6>." + } + ] + }, + { + "video_id": "NZ7VY", + "question_id": "NZ7VY", + "video_file": "Charades/videos/NZ7VY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person pours a cup of coffee into a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <14.3>." + } + ] + }, + { + "video_id": "EZ3CG", + "question_id": "EZ3CG", + "video_file": "Charades/videos/EZ3CG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person proceeds to work/play on the laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.2> to <31.4>." + } + ] + }, + { + "video_id": "EZ3CG", + "question_id": "EZ3CG", + "video_file": "Charades/videos/EZ3CG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person playing on the laptop'." + }, + { + "from": "gpt", + "value": "From <11.2> to <31.4>." + } + ] + }, + { + "video_id": "EZ3CG", + "question_id": "EZ3CG", + "video_file": "Charades/videos/EZ3CG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person sits on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <31.7>." + } + ] + }, + { + "video_id": "INKPE", + "question_id": "INKPE", + "video_file": "Charades/videos/INKPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take a clean glass out of the dishwasher' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.7> to <12.6>." + } + ] + }, + { + "video_id": "INKPE", + "question_id": "INKPE", + "video_file": "Charades/videos/INKPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the cup on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.4> to <31.5>." + } + ] + }, + { + "video_id": "INKPE", + "question_id": "INKPE", + "video_file": "Charades/videos/INKPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person goes to the stove to cook'." + }, + { + "from": "gpt", + "value": "From <1.5> to <8.6>." + } + ] + }, + { + "video_id": "INKPE", + "question_id": "INKPE", + "video_file": "Charades/videos/INKPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put the clean glass on the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.4> to <31.5>." + } + ] + }, + { + "video_id": "E7FGK", + "question_id": "E7FGK", + "video_file": "Charades/videos/E7FGK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes some medicine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <23.0>." + } + ] + }, + { + "video_id": "S72J9", + "question_id": "S72J9", + "video_file": "Charades/videos/S72J9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes a look out a window'." + }, + { + "from": "gpt", + "value": "From <15.5> to <26.2>." + } + ] + }, + { + "video_id": "S72J9", + "question_id": "S72J9", + "video_file": "Charades/videos/S72J9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closing the dryer door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.1> to <11.5>." + } + ] + }, + { + "video_id": "S72J9", + "question_id": "S72J9", + "video_file": "Charades/videos/S72J9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person look out the window'." + }, + { + "from": "gpt", + "value": "From <15.5> to <26.2>." + } + ] + }, + { + "video_id": "S72J9", + "question_id": "S72J9", + "video_file": "Charades/videos/S72J9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person drinks a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.5> to <12.9>." + } + ] + }, + { + "video_id": "S72J9", + "question_id": "S72J9", + "video_file": "Charades/videos/S72J9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they close a cabinet door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <11.5>." + } + ] + }, + { + "video_id": "S72J9", + "question_id": "S72J9", + "video_file": "Charades/videos/S72J9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <11.5>." + } + ] + }, + { + "video_id": "S72J9", + "question_id": "S72J9", + "video_file": "Charades/videos/S72J9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person looks out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.5> to <26.2>." + } + ] + }, + { + "video_id": "X8Y9X", + "question_id": "X8Y9X", + "video_file": "Charades/videos/X8Y9X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drink from a cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.1> to <29.0>." + } + ] + }, + { + "video_id": "IUGG8", + "question_id": "IUGG8", + "video_file": "Charades/videos/IUGG8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person pours some water into a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.3> to <8.5>." + } + ] + }, + { + "video_id": "TAGS2", + "question_id": "TAGS2", + "video_file": "Charades/videos/TAGS2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is laughing on the phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.9> to <16.3>." + } + ] + }, + { + "video_id": "TAGS2", + "question_id": "TAGS2", + "video_file": "Charades/videos/TAGS2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closing a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <22.7>." + } + ] + }, + { + "video_id": "PCF14", + "question_id": "PCF14", + "video_file": "Charades/videos/PCF14.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take their shoes off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <12.6>." + } + ] + }, + { + "video_id": "PCF14", + "question_id": "PCF14", + "video_file": "Charades/videos/PCF14.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes off their shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.5> to <12.6>." + } + ] + }, + { + "video_id": "MOQIU", + "question_id": "MOQIU", + "video_file": "Charades/videos/MOQIU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is laughing as the watch tv' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <18.9>." + } + ] + }, + { + "video_id": "MOQIU", + "question_id": "MOQIU", + "video_file": "Charades/videos/MOQIU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person laughing at a video on a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <18.9>." + } + ] + }, + { + "video_id": "MOQIU", + "question_id": "MOQIU", + "video_file": "Charades/videos/MOQIU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person leaves talking on the phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <30.0>." + } + ] + }, + { + "video_id": "TLDA3", + "question_id": "TLDA3", + "video_file": "Charades/videos/TLDA3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person smiles as they put away groceries'." + }, + { + "from": "gpt", + "value": "From <1.5> to <8.6>." + } + ] + }, + { + "video_id": "R4BTI", + "question_id": "R4BTI", + "video_file": "Charades/videos/R4BTI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is in a entryway sneezing by the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <15.0>." + } + ] + }, + { + "video_id": "R4BTI", + "question_id": "R4BTI", + "video_file": "Charades/videos/R4BTI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person start undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.5> to <34.8>." + } + ] + }, + { + "video_id": "R4BTI", + "question_id": "R4BTI", + "video_file": "Charades/videos/R4BTI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put it into a bag on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <32.4> to <40.0>." + } + ] + }, + { + "video_id": "R4BTI", + "question_id": "R4BTI", + "video_file": "Charades/videos/R4BTI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks into the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "R4BTI", + "question_id": "R4BTI", + "video_file": "Charades/videos/R4BTI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put their clothes in a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.4> to <40.0>." + } + ] + }, + { + "video_id": "AST4H", + "question_id": "AST4H", + "video_file": "Charades/videos/AST4H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person fixes their hair in a mirror' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.9> to <25.8>." + } + ] + }, + { + "video_id": "AST4H", + "question_id": "AST4H", + "video_file": "Charades/videos/AST4H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person walks through a doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.0> to <21.9>." + } + ] + }, + { + "video_id": "6KJGX", + "question_id": "6KJGX", + "video_file": "Charades/videos/6KJGX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws it on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.1> to <12.6>." + } + ] + }, + { + "video_id": "6KJGX", + "question_id": "6KJGX", + "video_file": "Charades/videos/6KJGX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person throws a broom on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.2> to <12.8>." + } + ] + }, + { + "video_id": "43OGO", + "question_id": "43OGO", + "video_file": "Charades/videos/43OGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person another is putting shoes in a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.1> to <32.0>." + } + ] + }, + { + "video_id": "43OGO", + "question_id": "43OGO", + "video_file": "Charades/videos/43OGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person is putting shoes in a bag'." + }, + { + "from": "gpt", + "value": "From <18.1> to <32.0>." + } + ] + }, + { + "video_id": "A52N5", + "question_id": "A52N5", + "video_file": "Charades/videos/A52N5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person in a garage is awakens on a sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <9.1>." + } + ] + }, + { + "video_id": "A52N5", + "question_id": "A52N5", + "video_file": "Charades/videos/A52N5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person opens a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.1> to <20.6>." + } + ] + }, + { + "video_id": "06X2V", + "question_id": "06X2V", + "video_file": "Charades/videos/06X2V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person smiles at himself in a mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <10.6>." + } + ] + }, + { + "video_id": "06X2V", + "question_id": "06X2V", + "video_file": "Charades/videos/06X2V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a towel out of a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.2> to <22.6>." + } + ] + }, + { + "video_id": "BDFDE", + "question_id": "BDFDE", + "video_file": "Charades/videos/BDFDE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes some medicine' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.3> to <28.9>." + } + ] + }, + { + "video_id": "BDFDE", + "question_id": "BDFDE", + "video_file": "Charades/videos/BDFDE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take some medicine' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.3> to <28.9>." + } + ] + }, + { + "video_id": "JOPGB", + "question_id": "JOPGB", + "video_file": "Charades/videos/JOPGB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts shoes on' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.4> to <8.5>." + } + ] + }, + { + "video_id": "JOPGB", + "question_id": "JOPGB", + "video_file": "Charades/videos/JOPGB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person gets from the bed to sit in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "3EIRZ", + "question_id": "3EIRZ", + "video_file": "Charades/videos/3EIRZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <0.2> to <10.2>." + } + ] + }, + { + "video_id": "3EIRZ", + "question_id": "3EIRZ", + "video_file": "Charades/videos/3EIRZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating foot' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <10.2>." + } + ] + }, + { + "video_id": "5IW8A", + "question_id": "5IW8A", + "video_file": "Charades/videos/5IW8A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinks from a coffee cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.9> to <33.0>." + } + ] + }, + { + "video_id": "BZD0Q", + "question_id": "BZD0Q", + "video_file": "Charades/videos/BZD0Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person walks out through the same doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.7> to <10.2>." + } + ] + }, + { + "video_id": "BZD0Q", + "question_id": "BZD0Q", + "video_file": "Charades/videos/BZD0Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walking through a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <10.2>." + } + ] + }, + { + "video_id": "XFT37", + "question_id": "XFT37", + "video_file": "Charades/videos/XFT37.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is opening the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.9> to <16.8>." + } + ] + }, + { + "video_id": "XFT37", + "question_id": "XFT37", + "video_file": "Charades/videos/XFT37.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opened the front door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.9> to <16.8>." + } + ] + }, + { + "video_id": "35P3Q", + "question_id": "35P3Q", + "video_file": "Charades/videos/35P3Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a load of clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.8> to <30.7>." + } + ] + }, + { + "video_id": "35P3Q", + "question_id": "35P3Q", + "video_file": "Charades/videos/35P3Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is washing dishes in a sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <11.7>." + } + ] + }, + { + "video_id": "FPQRX", + "question_id": "FPQRX", + "video_file": "Charades/videos/FPQRX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person sits on the bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.1> to <27.0>." + } + ] + }, + { + "video_id": "FPQRX", + "question_id": "FPQRX", + "video_file": "Charades/videos/FPQRX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person awakens'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.1>." + } + ] + }, + { + "video_id": "FPQRX", + "question_id": "FPQRX", + "video_file": "Charades/videos/FPQRX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts getting dressed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.5> to <46.7>." + } + ] + }, + { + "video_id": "FPQRX", + "question_id": "FPQRX", + "video_file": "Charades/videos/FPQRX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws a blanket towards the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.6> to <10.8>." + } + ] + }, + { + "video_id": "SU8PN", + "question_id": "SU8PN", + "video_file": "Charades/videos/SU8PN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'they throw a book back at the other person' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.8> to <14.3>." + } + ] + }, + { + "video_id": "SU8PN", + "question_id": "SU8PN", + "video_file": "Charades/videos/SU8PN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person the people in the video were throwing a pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "SU8PN", + "question_id": "SU8PN", + "video_file": "Charades/videos/SU8PN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'that person is smiling because they think it is funny' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <21.6>." + } + ] + }, + { + "video_id": "7EBTI", + "question_id": "7EBTI", + "video_file": "Charades/videos/7EBTI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes a bite of sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.2> to <9.4>." + } + ] + }, + { + "video_id": "7EBTI", + "question_id": "7EBTI", + "video_file": "Charades/videos/7EBTI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walks through the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.5> to <32.0>." + } + ] + }, + { + "video_id": "9RKTE", + "question_id": "9RKTE", + "video_file": "Charades/videos/9RKTE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opening a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.0> to <25.2>." + } + ] + }, + { + "video_id": "74DPG", + "question_id": "74DPG", + "video_file": "Charades/videos/74DPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person sits at a chair next to the stove'." + }, + { + "from": "gpt", + "value": "From <5.7> to <15.8>." + } + ] + }, + { + "video_id": "61WIS", + "question_id": "61WIS", + "video_file": "Charades/videos/61WIS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they sit down on the sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "61WIS", + "question_id": "61WIS", + "video_file": "Charades/videos/61WIS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.2> to <25.5>." + } + ] + }, + { + "video_id": "61WIS", + "question_id": "61WIS", + "video_file": "Charades/videos/61WIS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person also putting a pillow onto the ground' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.6> to <16.1>." + } + ] + }, + { + "video_id": "61WIS", + "question_id": "61WIS", + "video_file": "Charades/videos/61WIS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person with a laptop sits on a sofa'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "61WIS", + "question_id": "61WIS", + "video_file": "Charades/videos/61WIS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take a pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.1> to <16.1>." + } + ] + }, + { + "video_id": "TJPFT", + "question_id": "TJPFT", + "video_file": "Charades/videos/TJPFT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person watching television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "TJPFT", + "question_id": "TJPFT", + "video_file": "Charades/videos/TJPFT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding cup sits down drinking wiping shirt' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <9.9>." + } + ] + }, + { + "video_id": "TJPFT", + "question_id": "TJPFT", + "video_file": "Charades/videos/TJPFT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is drinking a cup of cocoa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <9.9>." + } + ] + }, + { + "video_id": "XPQ0D", + "question_id": "XPQ0D", + "video_file": "Charades/videos/XPQ0D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person is running down the stairs grasping medicine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <20.7>." + } + ] + }, + { + "video_id": "XPQ0D", + "question_id": "XPQ0D", + "video_file": "Charades/videos/XPQ0D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs down the stars' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <20.7>." + } + ] + }, + { + "video_id": "4713W", + "question_id": "4713W", + "video_file": "Charades/videos/4713W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person while walking into through the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.0> to <9.7>." + } + ] + }, + { + "video_id": "4713W", + "question_id": "4713W", + "video_file": "Charades/videos/4713W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.3> to <8.5>." + } + ] + }, + { + "video_id": "4713W", + "question_id": "4713W", + "video_file": "Charades/videos/4713W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person started sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <20.8>." + } + ] + }, + { + "video_id": "VGTX8", + "question_id": "VGTX8", + "video_file": "Charades/videos/VGTX8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the same window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <29.3> to <35.0>." + } + ] + }, + { + "video_id": "E89O3", + "question_id": "E89O3", + "video_file": "Charades/videos/E89O3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person walks through a doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <12.4>." + } + ] + }, + { + "video_id": "E89O3", + "question_id": "E89O3", + "video_file": "Charades/videos/E89O3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts to get dressed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.2> to <16.5>." + } + ] + }, + { + "video_id": "G87XG", + "question_id": "G87XG", + "video_file": "Charades/videos/G87XG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a woman running in a kitchen' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.6> to <6.7>." + } + ] + }, + { + "video_id": "G87XG", + "question_id": "G87XG", + "video_file": "Charades/videos/G87XG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits down in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <12.5>." + } + ] + }, + { + "video_id": "G87XG", + "question_id": "G87XG", + "video_file": "Charades/videos/G87XG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they sit in a chair'." + }, + { + "from": "gpt", + "value": "From <7.3> to <12.5>." + } + ] + }, + { + "video_id": "G87XG", + "question_id": "G87XG", + "video_file": "Charades/videos/G87XG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs into the laundry room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.6> to <6.7>." + } + ] + }, + { + "video_id": "G87XG", + "question_id": "G87XG", + "video_file": "Charades/videos/G87XG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person runs into the kitchen holding a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.6> to <6.7>." + } + ] + }, + { + "video_id": "FMXZS", + "question_id": "FMXZS", + "video_file": "Charades/videos/FMXZS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person holding a bag of groceries in a kitchen' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.3> to <19.2>." + } + ] + }, + { + "video_id": "FMXZS", + "question_id": "FMXZS", + "video_file": "Charades/videos/FMXZS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a refrigerator door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "FMXZS", + "question_id": "FMXZS", + "video_file": "Charades/videos/FMXZS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.2> to <32.0>." + } + ] + }, + { + "video_id": "ROEB7", + "question_id": "ROEB7", + "video_file": "Charades/videos/ROEB7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person a phone runs to the end of a room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <10.9>." + } + ] + }, + { + "video_id": "ROEB7", + "question_id": "ROEB7", + "video_file": "Charades/videos/ROEB7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs into a recreation room holding a bag'." + }, + { + "from": "gpt", + "value": "From <4.0> to <10.9>." + } + ] + }, + { + "video_id": "2846J", + "question_id": "2846J", + "video_file": "Charades/videos/2846J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks through the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "2846J", + "question_id": "2846J", + "video_file": "Charades/videos/2846J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person enters the laundry room holding a box'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "2846J", + "question_id": "2846J", + "video_file": "Charades/videos/2846J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "CPG4Y", + "question_id": "CPG4Y", + "video_file": "Charades/videos/CPG4Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a bag of groceries' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "7R4YS", + "question_id": "7R4YS", + "video_file": "Charades/videos/7R4YS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they put on a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.2> to <10.1>." + } + ] + }, + { + "video_id": "EBMIO", + "question_id": "EBMIO", + "video_file": "Charades/videos/EBMIO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts a towel on the rack' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "DEJ35", + "question_id": "DEJ35", + "video_file": "Charades/videos/DEJ35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person lying on the couch who sits up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.1> to <28.0>." + } + ] + }, + { + "video_id": "DEJ35", + "question_id": "DEJ35", + "video_file": "Charades/videos/DEJ35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person awakens'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <8.9>." + } + ] + }, + { + "video_id": "DEJ35", + "question_id": "DEJ35", + "video_file": "Charades/videos/DEJ35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person dresses by putting on a jacket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <22.7>." + } + ] + }, + { + "video_id": "8N7NZ", + "question_id": "8N7NZ", + "video_file": "Charades/videos/8N7NZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they stand up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.4> to <11.7>." + } + ] + }, + { + "video_id": "8N7NZ", + "question_id": "8N7NZ", + "video_file": "Charades/videos/8N7NZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put the blanket away into a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.8> to <26.5>." + } + ] + }, + { + "video_id": "S29H6", + "question_id": "S29H6", + "video_file": "Charades/videos/S29H6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they stand up'." + }, + { + "from": "gpt", + "value": "From <3.4> to <11.1>." + } + ] + }, + { + "video_id": "S29H6", + "question_id": "S29H6", + "video_file": "Charades/videos/S29H6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is lying on the floor reading a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.2>." + } + ] + }, + { + "video_id": "S29H6", + "question_id": "S29H6", + "video_file": "Charades/videos/S29H6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.5> to <24.3>." + } + ] + }, + { + "video_id": "S29H6", + "question_id": "S29H6", + "video_file": "Charades/videos/S29H6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks to the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.3> to <26.7>." + } + ] + }, + { + "video_id": "S29H6", + "question_id": "S29H6", + "video_file": "Charades/videos/S29H6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person walk through the doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.3> to <26.7>." + } + ] + }, + { + "video_id": "S29H6", + "question_id": "S29H6", + "video_file": "Charades/videos/S29H6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open a door'." + }, + { + "from": "gpt", + "value": "From <12.5> to <24.3>." + } + ] + }, + { + "video_id": "B5NS3", + "question_id": "B5NS3", + "video_file": "Charades/videos/B5NS3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person is dressing themselves in front of a mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "B5NS3", + "question_id": "B5NS3", + "video_file": "Charades/videos/B5NS3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turn off the light'." + }, + { + "from": "gpt", + "value": "From <20.1> to <29.1>." + } + ] + }, + { + "video_id": "B5NS3", + "question_id": "B5NS3", + "video_file": "Charades/videos/B5NS3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person finishes dressing their bathroom'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "3QXPC", + "question_id": "3QXPC", + "video_file": "Charades/videos/3QXPC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <2.1> to <8.3>." + } + ] + }, + { + "video_id": "3QXPC", + "question_id": "3QXPC", + "video_file": "Charades/videos/3QXPC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walk through the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <7.0>." + } + ] + }, + { + "video_id": "3QXPC", + "question_id": "3QXPC", + "video_file": "Charades/videos/3QXPC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they take a bag from the cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.7> to <19.1>." + } + ] + }, + { + "video_id": "C2QK3", + "question_id": "C2QK3", + "video_file": "Charades/videos/C2QK3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "C2QK3", + "question_id": "C2QK3", + "video_file": "Charades/videos/C2QK3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens up a refrigerator'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "MOREL", + "question_id": "MOREL", + "video_file": "Charades/videos/MOREL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "MOREL", + "question_id": "MOREL", + "video_file": "Charades/videos/MOREL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'man opens a kitchen pantry door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "NPM0Z", + "question_id": "NPM0Z", + "video_file": "Charades/videos/NPM0Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts down a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <22.4>." + } + ] + }, + { + "video_id": "KAKSU", + "question_id": "KAKSU", + "video_file": "Charades/videos/KAKSU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns out the light'." + }, + { + "from": "gpt", + "value": "From <5.2> to <10.3>." + } + ] + }, + { + "video_id": "KAKSU", + "question_id": "KAKSU", + "video_file": "Charades/videos/KAKSU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is turning on a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <10.3>." + } + ] + }, + { + "video_id": "8FP8X", + "question_id": "8FP8X", + "video_file": "Charades/videos/8FP8X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the glass down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <17.0>." + } + ] + }, + { + "video_id": "8FP8X", + "question_id": "8FP8X", + "video_file": "Charades/videos/8FP8X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is holding a towel' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.8> to <13.1>." + } + ] + }, + { + "video_id": "G7LU9", + "question_id": "G7LU9", + "video_file": "Charades/videos/G7LU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes door'." + }, + { + "from": "gpt", + "value": "From <10.0> to <17.2>." + } + ] + }, + { + "video_id": "G7LU9", + "question_id": "G7LU9", + "video_file": "Charades/videos/G7LU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens doorknob' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <13.6>." + } + ] + }, + { + "video_id": "G7LU9", + "question_id": "G7LU9", + "video_file": "Charades/videos/G7LU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <13.6>." + } + ] + }, + { + "video_id": "DAIN7", + "question_id": "DAIN7", + "video_file": "Charades/videos/DAIN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the phone down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.9> to <32.3>." + } + ] + }, + { + "video_id": "9FBAP", + "question_id": "9FBAP", + "video_file": "Charades/videos/9FBAP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <7.4>." + } + ] + }, + { + "video_id": "9FBAP", + "question_id": "9FBAP", + "video_file": "Charades/videos/9FBAP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a bag out' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.5> to <9.7>." + } + ] + }, + { + "video_id": "6LUM5", + "question_id": "6LUM5", + "video_file": "Charades/videos/6LUM5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person sitting in a chair closes a laptop'." + }, + { + "from": "gpt", + "value": "From <8.4> to <14.4>." + } + ] + }, + { + "video_id": "6LUM5", + "question_id": "6LUM5", + "video_file": "Charades/videos/6LUM5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits in the chair'." + }, + { + "from": "gpt", + "value": "From <4.2> to <15.1>." + } + ] + }, + { + "video_id": "4FGLC", + "question_id": "4FGLC", + "video_file": "Charades/videos/4FGLC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person laughs' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.8> to <31.0>." + } + ] + }, + { + "video_id": "4FGLC", + "question_id": "4FGLC", + "video_file": "Charades/videos/4FGLC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'laughing at the person on the chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.8> to <31.0>." + } + ] + }, + { + "video_id": "EY8S6", + "question_id": "EY8S6", + "video_file": "Charades/videos/EY8S6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.0> to <18.0>." + } + ] + }, + { + "video_id": "PM2GT", + "question_id": "PM2GT", + "video_file": "Charades/videos/PM2GT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person puts a book on the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.0> to <9.9>." + } + ] + }, + { + "video_id": "AXS82", + "question_id": "AXS82", + "video_file": "Charades/videos/AXS82.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the refrigerator to look in' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.1> to <31.0>." + } + ] + }, + { + "video_id": "AXS82", + "question_id": "AXS82", + "video_file": "Charades/videos/AXS82.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.4> to <16.5>." + } + ] + }, + { + "video_id": "AXS82", + "question_id": "AXS82", + "video_file": "Charades/videos/AXS82.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they close the book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.7> to <19.2>." + } + ] + }, + { + "video_id": "KY28J", + "question_id": "KY28J", + "video_file": "Charades/videos/KY28J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person is watching tv on their laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.7> to <21.1>." + } + ] + }, + { + "video_id": "KY28J", + "question_id": "KY28J", + "video_file": "Charades/videos/KY28J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.4> to <15.2>." + } + ] + }, + { + "video_id": "KY28J", + "question_id": "KY28J", + "video_file": "Charades/videos/KY28J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is eating at their desk'." + }, + { + "from": "gpt", + "value": "From <8.4> to <15.2>." + } + ] + }, + { + "video_id": "PC0M6", + "question_id": "PC0M6", + "video_file": "Charades/videos/PC0M6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.6> to <14.0>." + } + ] + }, + { + "video_id": "PC0M6", + "question_id": "PC0M6", + "video_file": "Charades/videos/PC0M6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person closes the door behind them'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <16.4>." + } + ] + }, + { + "video_id": "0N7OQ", + "question_id": "0N7OQ", + "video_file": "Charades/videos/0N7OQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <33.4>." + } + ] + }, + { + "video_id": "9IT59", + "question_id": "9IT59", + "video_file": "Charades/videos/9IT59.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person holding a pillow walks through a hallway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.1>." + } + ] + }, + { + "video_id": "9IT59", + "question_id": "9IT59", + "video_file": "Charades/videos/9IT59.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is holding a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.1>." + } + ] + }, + { + "video_id": "696GE", + "question_id": "696GE", + "video_file": "Charades/videos/696GE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <22.2>." + } + ] + }, + { + "video_id": "EWES1", + "question_id": "EWES1", + "video_file": "Charades/videos/EWES1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sitting down to eat a sandwich'." + }, + { + "from": "gpt", + "value": "From <24.5> to <33.0>." + } + ] + }, + { + "video_id": "EWES1", + "question_id": "EWES1", + "video_file": "Charades/videos/EWES1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person begins to eat a sandwich'." + }, + { + "from": "gpt", + "value": "From <24.5> to <33.0>." + } + ] + }, + { + "video_id": "H1J19", + "question_id": "H1J19", + "video_file": "Charades/videos/H1J19.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put down the food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.8> to <23.2>." + } + ] + }, + { + "video_id": "H1J19", + "question_id": "H1J19", + "video_file": "Charades/videos/H1J19.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put the food down on a nearby table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.8> to <23.2>." + } + ] + }, + { + "video_id": "K7HQG", + "question_id": "K7HQG", + "video_file": "Charades/videos/K7HQG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person is putting a picture on the wall'." + }, + { + "from": "gpt", + "value": "From <8.0> to <23.1>." + } + ] + }, + { + "video_id": "OYNBC", + "question_id": "OYNBC", + "video_file": "Charades/videos/OYNBC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the fridge door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.4> to <22.2>." + } + ] + }, + { + "video_id": "M1TZR", + "question_id": "M1TZR", + "video_file": "Charades/videos/M1TZR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.8> to <26.3>." + } + ] + }, + { + "video_id": "M1TZR", + "question_id": "M1TZR", + "video_file": "Charades/videos/M1TZR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts dishes in sink' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.0> to <31.2>." + } + ] + }, + { + "video_id": "M1TZR", + "question_id": "M1TZR", + "video_file": "Charades/videos/M1TZR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a towel out of a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.5> to <36.2>." + } + ] + }, + { + "video_id": "M1TZR", + "question_id": "M1TZR", + "video_file": "Charades/videos/M1TZR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a towel from the shelf'." + }, + { + "from": "gpt", + "value": "From <27.5> to <36.2>." + } + ] + }, + { + "video_id": "M1TZR", + "question_id": "M1TZR", + "video_file": "Charades/videos/M1TZR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.8> to <26.3>." + } + ] + }, + { + "video_id": "M1TZR", + "question_id": "M1TZR", + "video_file": "Charades/videos/M1TZR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins eating' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.3> to <28.7>." + } + ] + }, + { + "video_id": "XEASG", + "question_id": "XEASG", + "video_file": "Charades/videos/XEASG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drink a cup of coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.7> to <26.1>." + } + ] + }, + { + "video_id": "XEASG", + "question_id": "XEASG", + "video_file": "Charades/videos/XEASG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinks from a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.7> to <26.1>." + } + ] + }, + { + "video_id": "M52K2", + "question_id": "M52K2", + "video_file": "Charades/videos/M52K2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person goes back by washing mashine puts clothes in bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <39.4> to <45.6>." + } + ] + }, + { + "video_id": "M52K2", + "question_id": "M52K2", + "video_file": "Charades/videos/M52K2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits down on floor in laundry room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <23.4>." + } + ] + }, + { + "video_id": "M52K2", + "question_id": "M52K2", + "video_file": "Charades/videos/M52K2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pours milk into cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.3> to <16.6>." + } + ] + }, + { + "video_id": "0YXN6", + "question_id": "0YXN6", + "video_file": "Charades/videos/0YXN6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.9> to <21.2>." + } + ] + }, + { + "video_id": "0YXN6", + "question_id": "0YXN6", + "video_file": "Charades/videos/0YXN6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person is standing up sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.6> to <27.9>." + } + ] + }, + { + "video_id": "BT7FV", + "question_id": "BT7FV", + "video_file": "Charades/videos/BT7FV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person reading through the book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.9> to <21.4>." + } + ] + }, + { + "video_id": "BT7FV", + "question_id": "BT7FV", + "video_file": "Charades/videos/BT7FV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens the book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.9> to <21.4>." + } + ] + }, + { + "video_id": "BT7FV", + "question_id": "BT7FV", + "video_file": "Charades/videos/BT7FV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person closes the book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.3> to <26.2>." + } + ] + }, + { + "video_id": "BT7FV", + "question_id": "BT7FV", + "video_file": "Charades/videos/BT7FV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.3> to <26.2>." + } + ] + }, + { + "video_id": "ROOLO", + "question_id": "ROOLO", + "video_file": "Charades/videos/ROOLO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <12.7>." + } + ] + }, + { + "video_id": "NKCXF", + "question_id": "NKCXF", + "video_file": "Charades/videos/NKCXF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts down the pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <15.3>." + } + ] + }, + { + "video_id": "CV5YJ", + "question_id": "CV5YJ", + "video_file": "Charades/videos/CV5YJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a laptop from a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.9> to <33.6>." + } + ] + }, + { + "video_id": "IU8YI", + "question_id": "IU8YI", + "video_file": "Charades/videos/IU8YI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person grabs a broom to begin tidying up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.6> to <7.5>." + } + ] + }, + { + "video_id": "NHXI3", + "question_id": "NHXI3", + "video_file": "Charades/videos/NHXI3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walks into the room which has the light on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.2> to <29.2>." + } + ] + }, + { + "video_id": "NHXI3", + "question_id": "NHXI3", + "video_file": "Charades/videos/NHXI3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a book from the table in the entryway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <13.4>." + } + ] + }, + { + "video_id": "NHXI3", + "question_id": "NHXI3", + "video_file": "Charades/videos/NHXI3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turned on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.2> to <29.2>." + } + ] + }, + { + "video_id": "NHXI3", + "question_id": "NHXI3", + "video_file": "Charades/videos/NHXI3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.2> to <29.2>." + } + ] + }, + { + "video_id": "NHXI3", + "question_id": "NHXI3", + "video_file": "Charades/videos/NHXI3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes out a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <13.4>." + } + ] + }, + { + "video_id": "NHXI3", + "question_id": "NHXI3", + "video_file": "Charades/videos/NHXI3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walks through another doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.6> to <27.6>." + } + ] + }, + { + "video_id": "NHXI3", + "question_id": "NHXI3", + "video_file": "Charades/videos/NHXI3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walks toward a doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <27.6>." + } + ] + }, + { + "video_id": "115MY", + "question_id": "115MY", + "video_file": "Charades/videos/115MY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.2> to <14.2>." + } + ] + }, + { + "video_id": "115MY", + "question_id": "115MY", + "video_file": "Charades/videos/115MY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person quickly awakens' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.3>." + } + ] + }, + { + "video_id": "MI0ZQ", + "question_id": "MI0ZQ", + "video_file": "Charades/videos/MI0ZQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person puts a towel on the counter' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "MI0ZQ", + "question_id": "MI0ZQ", + "video_file": "Charades/videos/MI0ZQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take off their shoes once inside the bathroom'." + }, + { + "from": "gpt", + "value": "From <15.1> to <25.8>." + } + ] + }, + { + "video_id": "MI0ZQ", + "question_id": "MI0ZQ", + "video_file": "Charades/videos/MI0ZQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person start undressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.1> to <14.4>." + } + ] + }, + { + "video_id": "MI0ZQ", + "question_id": "MI0ZQ", + "video_file": "Charades/videos/MI0ZQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person in the video is starting to undress' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.1> to <14.4>." + } + ] + }, + { + "video_id": "S7PLQ", + "question_id": "S7PLQ", + "video_file": "Charades/videos/S7PLQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits on a couch'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.6> to <33.0>." + } + ] + }, + { + "video_id": "S7PLQ", + "question_id": "S7PLQ", + "video_file": "Charades/videos/S7PLQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person closes the refrigerator door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <19.3>." + } + ] + }, + { + "video_id": "UZRQD", + "question_id": "UZRQD", + "video_file": "Charades/videos/UZRQD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the phone on a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.3> to <26.9>." + } + ] + }, + { + "video_id": "YTV2H", + "question_id": "YTV2H", + "video_file": "Charades/videos/YTV2H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turn off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.6> to <33.0>." + } + ] + }, + { + "video_id": "YTV2H", + "question_id": "YTV2H", + "video_file": "Charades/videos/YTV2H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the water onto the shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.8> to <17.5>." + } + ] + }, + { + "video_id": "CZPU9", + "question_id": "CZPU9", + "video_file": "Charades/videos/CZPU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes the medicine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.6> to <24.9>." + } + ] + }, + { + "video_id": "CZPU9", + "question_id": "CZPU9", + "video_file": "Charades/videos/CZPU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person put a towel on the doorknob'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "CZPU9", + "question_id": "CZPU9", + "video_file": "Charades/videos/CZPU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the cabinet door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.5> to <30.4>." + } + ] + }, + { + "video_id": "CZPU9", + "question_id": "CZPU9", + "video_file": "Charades/videos/CZPU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is putting a towel on the doorknob'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "CZPU9", + "question_id": "CZPU9", + "video_file": "Charades/videos/CZPU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.6> to <29.7>." + } + ] + }, + { + "video_id": "CZPU9", + "question_id": "CZPU9", + "video_file": "Charades/videos/CZPU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person started undressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.0> to <13.7>." + } + ] + }, + { + "video_id": "CZPU9", + "question_id": "CZPU9", + "video_file": "Charades/videos/CZPU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes out a box of medicine'." + }, + { + "from": "gpt", + "value": "From <18.6> to <24.9>." + } + ] + }, + { + "video_id": "CZPU9", + "question_id": "CZPU9", + "video_file": "Charades/videos/CZPU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts clothes on a doorknob' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.6>." + } + ] + }, + { + "video_id": "CZPU9", + "question_id": "CZPU9", + "video_file": "Charades/videos/CZPU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person he undresses' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <13.7>." + } + ] + }, + { + "video_id": "2RTH2", + "question_id": "2RTH2", + "video_file": "Charades/videos/2RTH2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turning off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.1> to <32.0>." + } + ] + }, + { + "video_id": "2RTH2", + "question_id": "2RTH2", + "video_file": "Charades/videos/2RTH2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they turn off the light'." + }, + { + "from": "gpt", + "value": "From <27.1> to <32.0>." + } + ] + }, + { + "video_id": "2RTH2", + "question_id": "2RTH2", + "video_file": "Charades/videos/2RTH2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts the towel down'." + }, + { + "from": "gpt", + "value": "From <11.6> to <17.0>." + } + ] + }, + { + "video_id": "2D98B", + "question_id": "2D98B", + "video_file": "Charades/videos/2D98B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they start opening a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <9.9>." + } + ] + }, + { + "video_id": "WETXR", + "question_id": "WETXR", + "video_file": "Charades/videos/WETXR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person appears to be sneezing repeatedly'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.3> to <13.9>." + } + ] + }, + { + "video_id": "I8I5F", + "question_id": "I8I5F", + "video_file": "Charades/videos/I8I5F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts to laugh happily' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.5> to <28.0>." + } + ] + }, + { + "video_id": "I8I5F", + "question_id": "I8I5F", + "video_file": "Charades/videos/I8I5F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is holding a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "I8I5F", + "question_id": "I8I5F", + "video_file": "Charades/videos/I8I5F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a pillow'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "82YF1", + "question_id": "82YF1", + "video_file": "Charades/videos/82YF1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts down a cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.9> to <25.9>." + } + ] + }, + { + "video_id": "82YF1", + "question_id": "82YF1", + "video_file": "Charades/videos/82YF1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinks from cup'." + }, + { + "from": "gpt", + "value": "From <14.8> to <21.6>." + } + ] + }, + { + "video_id": "WBS4I", + "question_id": "WBS4I", + "video_file": "Charades/videos/WBS4I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <5.4>." + } + ] + }, + { + "video_id": "WBS4I", + "question_id": "WBS4I", + "video_file": "Charades/videos/WBS4I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns on a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <5.4>." + } + ] + }, + { + "video_id": "OEE36", + "question_id": "OEE36", + "video_file": "Charades/videos/OEE36.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts the box on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <32.3> to <40.2>." + } + ] + }, + { + "video_id": "OEE36", + "question_id": "OEE36", + "video_file": "Charades/videos/OEE36.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is putting towels'." + }, + { + "from": "gpt", + "value": "From <21.1> to <29.1>." + } + ] + }, + { + "video_id": "UCT5X", + "question_id": "UCT5X", + "video_file": "Charades/videos/UCT5X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walk out through a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.7> to <15.4>." + } + ] + }, + { + "video_id": "G6GQP", + "question_id": "G6GQP", + "video_file": "Charades/videos/G6GQP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'laughing as the person stands in the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <15.1>." + } + ] + }, + { + "video_id": "SXMII", + "question_id": "SXMII", + "video_file": "Charades/videos/SXMII.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person awakens on a bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <5.0>." + } + ] + }, + { + "video_id": "SXMII", + "question_id": "SXMII", + "video_file": "Charades/videos/SXMII.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws a blanket on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.6> to <28.6>." + } + ] + }, + { + "video_id": "AZ5JS", + "question_id": "AZ5JS", + "video_file": "Charades/videos/AZ5JS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is holding a blanket in a doorway'." + }, + { + "from": "gpt", + "value": "From <29.1> to <32.0>." + } + ] + }, + { + "video_id": "AZ5JS", + "question_id": "AZ5JS", + "video_file": "Charades/videos/AZ5JS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.1> to <24.7>." + } + ] + }, + { + "video_id": "AZ5JS", + "question_id": "AZ5JS", + "video_file": "Charades/videos/AZ5JS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.7> to <32.0>." + } + ] + }, + { + "video_id": "DVPL2", + "question_id": "DVPL2", + "video_file": "Charades/videos/DVPL2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinks from a cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.3> to <20.8>." + } + ] + }, + { + "video_id": "DVPL2", + "question_id": "DVPL2", + "video_file": "Charades/videos/DVPL2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person drinks from the cup'." + }, + { + "from": "gpt", + "value": "From <13.3> to <20.8>." + } + ] + }, + { + "video_id": "7FTBS", + "question_id": "7FTBS", + "video_file": "Charades/videos/7FTBS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person sits in a chair at the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <39.2> to <50.0>." + } + ] + }, + { + "video_id": "7FTBS", + "question_id": "7FTBS", + "video_file": "Charades/videos/7FTBS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opens the refrigerator'." + }, + { + "from": "gpt", + "value": "From <20.0> to <25.8>." + } + ] + }, + { + "video_id": "7FTBS", + "question_id": "7FTBS", + "video_file": "Charades/videos/7FTBS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person is sitting at table drinking coffee' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <39.2> to <50.0>." + } + ] + }, + { + "video_id": "WZY7C", + "question_id": "WZY7C", + "video_file": "Charades/videos/WZY7C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs through the door with a towel'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "WZY7C", + "question_id": "WZY7C", + "video_file": "Charades/videos/WZY7C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs into their study' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "S8X4A", + "question_id": "S8X4A", + "video_file": "Charades/videos/S8X4A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person in a bathroom starts sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <8.3>." + } + ] + }, + { + "video_id": "38TA9", + "question_id": "38TA9", + "video_file": "Charades/videos/38TA9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "38TA9", + "question_id": "38TA9", + "video_file": "Charades/videos/38TA9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'while the person was opening the front doorknob' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "38TA9", + "question_id": "38TA9", + "video_file": "Charades/videos/38TA9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a closet door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "HKLB1", + "question_id": "HKLB1", + "video_file": "Charades/videos/HKLB1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is undressing in front of a mirror'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <18.2>." + } + ] + }, + { + "video_id": "HKLB1", + "question_id": "HKLB1", + "video_file": "Charades/videos/HKLB1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person undresses in front of a mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.3> to <18.2>." + } + ] + }, + { + "video_id": "ZISOA", + "question_id": "ZISOA", + "video_file": "Charades/videos/ZISOA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a camera' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "RZLAZ", + "question_id": "RZLAZ", + "video_file": "Charades/videos/RZLAZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs up the stairs' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "Q2U7O", + "question_id": "Q2U7O", + "video_file": "Charades/videos/Q2U7O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes a glass of water from a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.2> to <8.2>." + } + ] + }, + { + "video_id": "Q2U7O", + "question_id": "Q2U7O", + "video_file": "Charades/videos/Q2U7O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.9> to <15.6>." + } + ] + }, + { + "video_id": "H6BT7", + "question_id": "H6BT7", + "video_file": "Charades/videos/H6BT7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person closes the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <30.2>." + } + ] + }, + { + "video_id": "H6BT7", + "question_id": "H6BT7", + "video_file": "Charades/videos/H6BT7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the box'." + }, + { + "from": "gpt", + "value": "From <18.0> to <30.2>." + } + ] + }, + { + "video_id": "H6BT7", + "question_id": "H6BT7", + "video_file": "Charades/videos/H6BT7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person lays on the floor'." + }, + { + "from": "gpt", + "value": "From <33.1> to <54.0>." + } + ] + }, + { + "video_id": "H6BT7", + "question_id": "H6BT7", + "video_file": "Charades/videos/H6BT7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks through a doorway holding a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <28.9>." + } + ] + }, + { + "video_id": "H6BT7", + "question_id": "H6BT7", + "video_file": "Charades/videos/H6BT7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens the box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "G2J63", + "question_id": "G2J63", + "video_file": "Charades/videos/G2J63.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throwing towel on floor'." + }, + { + "from": "gpt", + "value": "From <14.4> to <22.7>." + } + ] + }, + { + "video_id": "MHI9B", + "question_id": "MHI9B", + "video_file": "Charades/videos/MHI9B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person pours coffee in a cup in a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <23.8>." + } + ] + }, + { + "video_id": "MHI9B", + "question_id": "MHI9B", + "video_file": "Charades/videos/MHI9B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pours coffee into a glass at the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.0> to <23.8>." + } + ] + }, + { + "video_id": "ZQAMW", + "question_id": "ZQAMW", + "video_file": "Charades/videos/ZQAMW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sneezing into a glass in a living room'." + }, + { + "from": "gpt", + "value": "From <20.5> to <34.0>." + } + ] + }, + { + "video_id": "F1QR1", + "question_id": "F1QR1", + "video_file": "Charades/videos/F1QR1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a door walks out' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.8> to <23.4>." + } + ] + }, + { + "video_id": "KD3SK", + "question_id": "KD3SK", + "video_file": "Charades/videos/KD3SK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person where they take a blanket from the shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.8> to <23.4>." + } + ] + }, + { + "video_id": "KD3SK", + "question_id": "KD3SK", + "video_file": "Charades/videos/KD3SK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they take a blanket from a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <23.4>." + } + ] + }, + { + "video_id": "2Z8HO", + "question_id": "2Z8HO", + "video_file": "Charades/videos/2Z8HO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a pair of shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.8>." + } + ] + }, + { + "video_id": "2Z8HO", + "question_id": "2Z8HO", + "video_file": "Charades/videos/2Z8HO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts down a pair of shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <15.7>." + } + ] + }, + { + "video_id": "5LAQN", + "question_id": "5LAQN", + "video_file": "Charades/videos/5LAQN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is drinking a cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.1>." + } + ] + }, + { + "video_id": "QQD2X", + "question_id": "QQD2X", + "video_file": "Charades/videos/QQD2X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is running into the bathroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "QQD2X", + "question_id": "QQD2X", + "video_file": "Charades/videos/QQD2X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens up the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.5> to <17.2>." + } + ] + }, + { + "video_id": "QQD2X", + "question_id": "QQD2X", + "video_file": "Charades/videos/QQD2X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.6> to <10.9>." + } + ] + }, + { + "video_id": "QQD2X", + "question_id": "QQD2X", + "video_file": "Charades/videos/QQD2X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turning on a light'." + }, + { + "from": "gpt", + "value": "From <4.6> to <10.9>." + } + ] + }, + { + "video_id": "QQD2X", + "question_id": "QQD2X", + "video_file": "Charades/videos/QQD2X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person runs into the bathroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "XX7ZJ", + "question_id": "XX7ZJ", + "video_file": "Charades/videos/XX7ZJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts smiling'." + }, + { + "from": "gpt", + "value": "From <12.0> to <17.0>." + } + ] + }, + { + "video_id": "5LFGO", + "question_id": "5LFGO", + "video_file": "Charades/videos/5LFGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person fixes a vacuum' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.6>." + } + ] + }, + { + "video_id": "5LFGO", + "question_id": "5LFGO", + "video_file": "Charades/videos/5LFGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they finish dressing by putting on a cap'." + }, + { + "from": "gpt", + "value": "From <17.1> to <24.4>." + } + ] + }, + { + "video_id": "OIXA4", + "question_id": "OIXA4", + "video_file": "Charades/videos/OIXA4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person is putting a glass on the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <21.3>." + } + ] + }, + { + "video_id": "G8E71", + "question_id": "G8E71", + "video_file": "Charades/videos/G8E71.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is undressing in a bathroom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "G8E71", + "question_id": "G8E71", + "video_file": "Charades/videos/G8E71.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "1SMZZ", + "question_id": "1SMZZ", + "video_file": "Charades/videos/1SMZZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a pillow from the bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <25.0>." + } + ] + }, + { + "video_id": "HL5OP", + "question_id": "HL5OP", + "video_file": "Charades/videos/HL5OP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is laughing'." + }, + { + "from": "gpt", + "value": "From <14.6> to <20.4>." + } + ] + }, + { + "video_id": "HL5OP", + "question_id": "HL5OP", + "video_file": "Charades/videos/HL5OP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throwing food in the trash bin' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.4> to <21.7>." + } + ] + }, + { + "video_id": "4F76V", + "question_id": "4F76V", + "video_file": "Charades/videos/4F76V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person cooking' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <10.6>." + } + ] + }, + { + "video_id": "4F76V", + "question_id": "4F76V", + "video_file": "Charades/videos/4F76V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is cooking'." + }, + { + "from": "gpt", + "value": "From <3.1> to <10.6>." + } + ] + }, + { + "video_id": "4F76V", + "question_id": "4F76V", + "video_file": "Charades/videos/4F76V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.9> to <27.6>." + } + ] + }, + { + "video_id": "EW6RX", + "question_id": "EW6RX", + "video_file": "Charades/videos/EW6RX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eat a sandwich'." + }, + { + "from": "gpt", + "value": "From <2.5> to <7.3>." + } + ] + }, + { + "video_id": "EW6RX", + "question_id": "EW6RX", + "video_file": "Charades/videos/EW6RX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.5> to <7.3>." + } + ] + }, + { + "video_id": "LGCTV", + "question_id": "LGCTV", + "video_file": "Charades/videos/LGCTV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <20.5>." + } + ] + }, + { + "video_id": "LGCTV", + "question_id": "LGCTV", + "video_file": "Charades/videos/LGCTV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begins tidying some clothes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <33.0>." + } + ] + }, + { + "video_id": "HEV65", + "question_id": "HEV65", + "video_file": "Charades/videos/HEV65.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they open the bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.6> to <12.0>." + } + ] + }, + { + "video_id": "HEV65", + "question_id": "HEV65", + "video_file": "Charades/videos/HEV65.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.8> to <23.7>." + } + ] + }, + { + "video_id": "XNXW6", + "question_id": "XNXW6", + "video_file": "Charades/videos/XNXW6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person laughing with a camera' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.5> to <18.0>." + } + ] + }, + { + "video_id": "RJ27X", + "question_id": "RJ27X", + "video_file": "Charades/videos/RJ27X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the food down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.2> to <22.3>." + } + ] + }, + { + "video_id": "RJ27X", + "question_id": "RJ27X", + "video_file": "Charades/videos/RJ27X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person is cooking' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "RJ27X", + "question_id": "RJ27X", + "video_file": "Charades/videos/RJ27X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person holding food walks to a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.8> to <21.9>." + } + ] + }, + { + "video_id": "LVTRJ", + "question_id": "LVTRJ", + "video_file": "Charades/videos/LVTRJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a drink out of the cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.8> to <42.4>." + } + ] + }, + { + "video_id": "LVTRJ", + "question_id": "LVTRJ", + "video_file": "Charades/videos/LVTRJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts a pair of shoes in the sink' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.4>." + } + ] + }, + { + "video_id": "LVTRJ", + "question_id": "LVTRJ", + "video_file": "Charades/videos/LVTRJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts to eat the sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.3> to <37.8>." + } + ] + }, + { + "video_id": "AEA8P", + "question_id": "AEA8P", + "video_file": "Charades/videos/AEA8P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "AEA8P", + "question_id": "AEA8P", + "video_file": "Charades/videos/AEA8P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person is in the basement eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "AEA8P", + "question_id": "AEA8P", + "video_file": "Charades/videos/AEA8P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting at at table eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "Q9Q9L", + "question_id": "Q9Q9L", + "video_file": "Charades/videos/Q9Q9L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person entering house they take a drink from glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <32.0>." + } + ] + }, + { + "video_id": "Q9Q9L", + "question_id": "Q9Q9L", + "video_file": "Charades/videos/Q9Q9L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting on a chair in their garage' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "IYNUU", + "question_id": "IYNUU", + "video_file": "Charades/videos/IYNUU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a picture of themselves' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.5> to <29.8>." + } + ] + }, + { + "video_id": "IYNUU", + "question_id": "IYNUU", + "video_file": "Charades/videos/IYNUU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes their phone from their pocket'." + }, + { + "from": "gpt", + "value": "From <18.1> to <23.9>." + } + ] + }, + { + "video_id": "IYNUU", + "question_id": "IYNUU", + "video_file": "Charades/videos/IYNUU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person use the camera to take a picture of themself' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.5> to <29.8>." + } + ] + }, + { + "video_id": "57WV7", + "question_id": "57WV7", + "video_file": "Charades/videos/57WV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.7> to <7.4>." + } + ] + }, + { + "video_id": "57WV7", + "question_id": "57WV7", + "video_file": "Charades/videos/57WV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits on the chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <14.9>." + } + ] + }, + { + "video_id": "57WV7", + "question_id": "57WV7", + "video_file": "Charades/videos/57WV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting in a chair in the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <14.9>." + } + ] + }, + { + "video_id": "DV31C", + "question_id": "DV31C", + "video_file": "Charades/videos/DV31C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes the glass off the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.1> to <29.1>." + } + ] + }, + { + "video_id": "6ZKZS", + "question_id": "6ZKZS", + "video_file": "Charades/videos/6ZKZS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pouring water into a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "6ZKZS", + "question_id": "6ZKZS", + "video_file": "Charades/videos/6ZKZS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sneezing into a blanket in a bedroom'." + }, + { + "from": "gpt", + "value": "From <19.5> to <26.0>." + } + ] + }, + { + "video_id": "6ZKZS", + "question_id": "6ZKZS", + "video_file": "Charades/videos/6ZKZS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they drink the water out of the glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.4> to <8.1>." + } + ] + }, + { + "video_id": "6ZKZS", + "question_id": "6ZKZS", + "video_file": "Charades/videos/6ZKZS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take some medicine'." + }, + { + "from": "gpt", + "value": "From <10.5> to <20.4>." + } + ] + }, + { + "video_id": "UJMNV", + "question_id": "UJMNV", + "video_file": "Charades/videos/UJMNV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person is sneezing in a towel' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "MC6J7", + "question_id": "MC6J7", + "video_file": "Charades/videos/MC6J7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take out a vacuum' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.3> to <32.6>." + } + ] + }, + { + "video_id": "MC6J7", + "question_id": "MC6J7", + "video_file": "Charades/videos/MC6J7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens the door of a wardrobe' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "MC6J7", + "question_id": "MC6J7", + "video_file": "Charades/videos/MC6J7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they throw the clothes in the closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.5> to <17.7>." + } + ] + }, + { + "video_id": "MC6J7", + "question_id": "MC6J7", + "video_file": "Charades/videos/MC6J7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walk away smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.3> to <33.1>." + } + ] + }, + { + "video_id": "MC6J7", + "question_id": "MC6J7", + "video_file": "Charades/videos/MC6J7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the clothes in the wardrobe'." + }, + { + "from": "gpt", + "value": "From <8.5> to <17.2>." + } + ] + }, + { + "video_id": "IC89C", + "question_id": "IC89C", + "video_file": "Charades/videos/IC89C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is cooking on a stove in the kitchen' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <15.6>." + } + ] + }, + { + "video_id": "IC89C", + "question_id": "IC89C", + "video_file": "Charades/videos/IC89C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts fixing a vacuum' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.6> to <46.7>." + } + ] + }, + { + "video_id": "IC89C", + "question_id": "IC89C", + "video_file": "Charades/videos/IC89C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is cooking' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.4> to <15.6>." + } + ] + }, + { + "video_id": "5MDL5", + "question_id": "5MDL5", + "video_file": "Charades/videos/5MDL5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts a broom into a closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "5MDL5", + "question_id": "5MDL5", + "video_file": "Charades/videos/5MDL5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person puts a broom down inside their closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "G1A2O", + "question_id": "G1A2O", + "video_file": "Charades/videos/G1A2O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person talking on the phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.8> to <29.7>." + } + ] + }, + { + "video_id": "G1A2O", + "question_id": "G1A2O", + "video_file": "Charades/videos/G1A2O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting eating something' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <9.7>." + } + ] + }, + { + "video_id": "G1A2O", + "question_id": "G1A2O", + "video_file": "Charades/videos/G1A2O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating something' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.9> to <9.7>." + } + ] + }, + { + "video_id": "G1A2O", + "question_id": "G1A2O", + "video_file": "Charades/videos/G1A2O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating in front of a television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.9> to <9.7>." + } + ] + }, + { + "video_id": "1ZB73", + "question_id": "1ZB73", + "video_file": "Charades/videos/1ZB73.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person laughing at something'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.9>." + } + ] + }, + { + "video_id": "NWGVT", + "question_id": "NWGVT", + "video_file": "Charades/videos/NWGVT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pouring it into a cup to make some coffee' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.2> to <13.8>." + } + ] + }, + { + "video_id": "S38BY", + "question_id": "S38BY", + "video_file": "Charades/videos/S38BY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person smiling in the entryway mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.8> to <19.7>." + } + ] + }, + { + "video_id": "1YH1Y", + "question_id": "1YH1Y", + "video_file": "Charades/videos/1YH1Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes some towels' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.5> to <7.7>." + } + ] + }, + { + "video_id": "1YH1Y", + "question_id": "1YH1Y", + "video_file": "Charades/videos/1YH1Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begin dressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "1YH1Y", + "question_id": "1YH1Y", + "video_file": "Charades/videos/1YH1Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a bag'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.3>." + } + ] + }, + { + "video_id": "1YH1Y", + "question_id": "1YH1Y", + "video_file": "Charades/videos/1YH1Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take some clothes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.3> to <10.9>." + } + ] + }, + { + "video_id": "TANB8", + "question_id": "TANB8", + "video_file": "Charades/videos/TANB8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts something on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <15.0>." + } + ] + }, + { + "video_id": "TANB8", + "question_id": "TANB8", + "video_file": "Charades/videos/TANB8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sneezing'." + }, + { + "from": "gpt", + "value": "From <14.6> to <22.5>." + } + ] + }, + { + "video_id": "GY70Y", + "question_id": "GY70Y", + "video_file": "Charades/videos/GY70Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking from a cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.6> to <23.0>." + } + ] + }, + { + "video_id": "SVO89", + "question_id": "SVO89", + "video_file": "Charades/videos/SVO89.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a book off of a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <21.2>." + } + ] + }, + { + "video_id": "SVO89", + "question_id": "SVO89", + "video_file": "Charades/videos/SVO89.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door behind him' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <25.6>." + } + ] + }, + { + "video_id": "D7D89", + "question_id": "D7D89", + "video_file": "Charades/videos/D7D89.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is looking at a picture on a cabinet'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "D7D89", + "question_id": "D7D89", + "video_file": "Charades/videos/D7D89.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person looks at a picture on a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "7OPHI", + "question_id": "7OPHI", + "video_file": "Charades/videos/7OPHI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opening a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "7OPHI", + "question_id": "7OPHI", + "video_file": "Charades/videos/7OPHI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens their laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "7OPHI", + "question_id": "7OPHI", + "video_file": "Charades/videos/7OPHI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person closes their laptop'." + }, + { + "from": "gpt", + "value": "From <25.0> to <34.0>." + } + ] + }, + { + "video_id": "BPDTY", + "question_id": "BPDTY", + "video_file": "Charades/videos/BPDTY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens the bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.1> to <18.7>." + } + ] + }, + { + "video_id": "CCTIO", + "question_id": "CCTIO", + "video_file": "Charades/videos/CCTIO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.8> to <12.9>." + } + ] + }, + { + "video_id": "CCTIO", + "question_id": "CCTIO", + "video_file": "Charades/videos/CCTIO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits on sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <8.9>." + } + ] + }, + { + "video_id": "FPNT1", + "question_id": "FPNT1", + "video_file": "Charades/videos/FPNT1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person throws the book on the floor'." + }, + { + "from": "gpt", + "value": "From <6.6> to <11.2>." + } + ] + }, + { + "video_id": "DGMDT", + "question_id": "DGMDT", + "video_file": "Charades/videos/DGMDT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put the bag down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.8> to <26.8>." + } + ] + }, + { + "video_id": "DGMDT", + "question_id": "DGMDT", + "video_file": "Charades/videos/DGMDT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts bag back on table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.8> to <26.8>." + } + ] + }, + { + "video_id": "AZZ3W", + "question_id": "AZZ3W", + "video_file": "Charades/videos/AZZ3W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put on their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.1> to <28.9>." + } + ] + }, + { + "video_id": "AZZ3W", + "question_id": "AZZ3W", + "video_file": "Charades/videos/AZZ3W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person throws some dirty clothes from a chair'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "AZZ3W", + "question_id": "AZZ3W", + "video_file": "Charades/videos/AZZ3W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person finally puts on shoes'." + }, + { + "from": "gpt", + "value": "From <24.1> to <28.9>." + } + ] + }, + { + "video_id": "AZZ3W", + "question_id": "AZZ3W", + "video_file": "Charades/videos/AZZ3W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person begins dressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.5> to <13.7>." + } + ] + }, + { + "video_id": "AZZ3W", + "question_id": "AZZ3W", + "video_file": "Charades/videos/AZZ3W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they get dressed'." + }, + { + "from": "gpt", + "value": "From <8.5> to <13.7>." + } + ] + }, + { + "video_id": "AZZ3W", + "question_id": "AZZ3W", + "video_file": "Charades/videos/AZZ3W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sit in the chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.0> to <26.9>." + } + ] + }, + { + "video_id": "5ZHOX", + "question_id": "5ZHOX", + "video_file": "Charades/videos/5ZHOX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'sneezing as the person does so' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.4> to <18.3>." + } + ] + }, + { + "video_id": "X0CVU", + "question_id": "X0CVU", + "video_file": "Charades/videos/X0CVU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looks out the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "XDFPV", + "question_id": "XDFPV", + "video_file": "Charades/videos/XDFPV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person two people sit on chairs in a living room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "XDFPV", + "question_id": "XDFPV", + "video_file": "Charades/videos/XDFPV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.7> to <18.5>." + } + ] + }, + { + "video_id": "79WPY", + "question_id": "79WPY", + "video_file": "Charades/videos/79WPY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is watching television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <22.2>." + } + ] + }, + { + "video_id": "79WPY", + "question_id": "79WPY", + "video_file": "Charades/videos/79WPY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is opening a refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.2> to <28.8>." + } + ] + }, + { + "video_id": "79WPY", + "question_id": "79WPY", + "video_file": "Charades/videos/79WPY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person closes the door to a refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.8> to <33.0>." + } + ] + }, + { + "video_id": "79WPY", + "question_id": "79WPY", + "video_file": "Charades/videos/79WPY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person watches television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <22.2>." + } + ] + }, + { + "video_id": "0K0LP", + "question_id": "0K0LP", + "video_file": "Charades/videos/0K0LP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opened a bag of chips'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.4> to <21.7>." + } + ] + }, + { + "video_id": "0K0LP", + "question_id": "0K0LP", + "video_file": "Charades/videos/0K0LP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person watching television instead' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.1> to <13.9>." + } + ] + }, + { + "video_id": "X8I9J", + "question_id": "X8I9J", + "video_file": "Charades/videos/X8I9J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a laughing person comes in through the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.1>." + } + ] + }, + { + "video_id": "X8I9J", + "question_id": "X8I9J", + "video_file": "Charades/videos/X8I9J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks through a door laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.1>." + } + ] + }, + { + "video_id": "X8I9J", + "question_id": "X8I9J", + "video_file": "Charades/videos/X8I9J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they begin undressing starting with their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <22.7>." + } + ] + }, + { + "video_id": "B9F4P", + "question_id": "B9F4P", + "video_file": "Charades/videos/B9F4P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person is closing the door to the bedroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.0> to <17.3>." + } + ] + }, + { + "video_id": "B9F4P", + "question_id": "B9F4P", + "video_file": "Charades/videos/B9F4P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person opens a closet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.0> to <17.3>." + } + ] + }, + { + "video_id": "OW9LO", + "question_id": "OW9LO", + "video_file": "Charades/videos/OW9LO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eat food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.7> to <30.0>." + } + ] + }, + { + "video_id": "OW9LO", + "question_id": "OW9LO", + "video_file": "Charades/videos/OW9LO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating food remains' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.7> to <30.0>." + } + ] + }, + { + "video_id": "OW9LO", + "question_id": "OW9LO", + "video_file": "Charades/videos/OW9LO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eat some snack food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.7> to <30.0>." + } + ] + }, + { + "video_id": "LC1NU", + "question_id": "LC1NU", + "video_file": "Charades/videos/LC1NU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is lying on the floor eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <0.0> to <19.0>." + } + ] + }, + { + "video_id": "LC1NU", + "question_id": "LC1NU", + "video_file": "Charades/videos/LC1NU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks down the hallway to turn off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <24.4>." + } + ] + }, + { + "video_id": "K0Y6X", + "question_id": "K0Y6X", + "video_file": "Charades/videos/K0Y6X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person pauses to turn on a light'." + }, + { + "from": "gpt", + "value": "From <2.2> to <6.5>." + } + ] + }, + { + "video_id": "K0Y6X", + "question_id": "K0Y6X", + "video_file": "Charades/videos/K0Y6X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person turns on the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <6.5>." + } + ] + }, + { + "video_id": "K0Y6X", + "question_id": "K0Y6X", + "video_file": "Charades/videos/K0Y6X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens their laptop to work on it'." + }, + { + "from": "gpt", + "value": "From <14.9> to <20.7>." + } + ] + }, + { + "video_id": "UKMK6", + "question_id": "UKMK6", + "video_file": "Charades/videos/UKMK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the pillow in a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <42.2> to <54.0>." + } + ] + }, + { + "video_id": "UKMK6", + "question_id": "UKMK6", + "video_file": "Charades/videos/UKMK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person snuggled with the pillow for a few seconds' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.9> to <42.1>." + } + ] + }, + { + "video_id": "UDGP2", + "question_id": "UDGP2", + "video_file": "Charades/videos/UDGP2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins cooking it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.8> to <32.0>." + } + ] + }, + { + "video_id": "UDGP2", + "question_id": "UDGP2", + "video_file": "Charades/videos/UDGP2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins cooking on a stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.8> to <32.0>." + } + ] + }, + { + "video_id": "UDGP2", + "question_id": "UDGP2", + "video_file": "Charades/videos/UDGP2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a refrigerator door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "NCBNH", + "question_id": "NCBNH", + "video_file": "Charades/videos/NCBNH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the glass of water on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.2> to <34.4>." + } + ] + }, + { + "video_id": "NCBNH", + "question_id": "NCBNH", + "video_file": "Charades/videos/NCBNH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs back to the camera' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <36.0> to <41.0>." + } + ] + }, + { + "video_id": "NCBNH", + "question_id": "NCBNH", + "video_file": "Charades/videos/NCBNH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a towel' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.5> to <34.0>." + } + ] + }, + { + "video_id": "UQVO7", + "question_id": "UQVO7", + "video_file": "Charades/videos/UQVO7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they are putting a pillow into a bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.7> to <16.7>." + } + ] + }, + { + "video_id": "3S70A", + "question_id": "3S70A", + "video_file": "Charades/videos/3S70A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting them on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.4> to <24.6>." + } + ] + }, + { + "video_id": "3S70A", + "question_id": "3S70A", + "video_file": "Charades/videos/3S70A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is awakening near a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.3> to <13.9>." + } + ] + }, + { + "video_id": "2EPFZ", + "question_id": "2EPFZ", + "video_file": "Charades/videos/2EPFZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins to put on their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.1> to <31.3>." + } + ] + }, + { + "video_id": "2EPFZ", + "question_id": "2EPFZ", + "video_file": "Charades/videos/2EPFZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits in a chair to put shoes on' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.1> to <31.3>." + } + ] + }, + { + "video_id": "0XP8L", + "question_id": "0XP8L", + "video_file": "Charades/videos/0XP8L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they go sit at a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.4> to <37.0>." + } + ] + }, + { + "video_id": "0XP8L", + "question_id": "0XP8L", + "video_file": "Charades/videos/0XP8L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person work on a laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.7> to <37.0>." + } + ] + }, + { + "video_id": "0XP8L", + "question_id": "0XP8L", + "video_file": "Charades/videos/0XP8L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes out food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.2> to <15.0>." + } + ] + }, + { + "video_id": "0XP8L", + "question_id": "0XP8L", + "video_file": "Charades/videos/0XP8L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is putting food into a pan' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.8> to <21.0>." + } + ] + }, + { + "video_id": "0XP8L", + "question_id": "0XP8L", + "video_file": "Charades/videos/0XP8L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person cooking it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <20.4>." + } + ] + }, + { + "video_id": "V7LTE", + "question_id": "V7LTE", + "video_file": "Charades/videos/V7LTE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is dressing on the stairs'." + }, + { + "from": "gpt", + "value": "From <0.8> to <11.3>." + } + ] + }, + { + "video_id": "V7LTE", + "question_id": "V7LTE", + "video_file": "Charades/videos/V7LTE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opens a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.9> to <23.0>." + } + ] + }, + { + "video_id": "40FKN", + "question_id": "40FKN", + "video_file": "Charades/videos/40FKN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person drinks a cup of coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "8W0XJ", + "question_id": "8W0XJ", + "video_file": "Charades/videos/8W0XJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "8W0XJ", + "question_id": "8W0XJ", + "video_file": "Charades/videos/8W0XJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person closes the garage window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "FIBCE", + "question_id": "FIBCE", + "video_file": "Charades/videos/FIBCE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turn off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.2> to <28.4>." + } + ] + }, + { + "video_id": "FIBCE", + "question_id": "FIBCE", + "video_file": "Charades/videos/FIBCE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they turn off the light'." + }, + { + "from": "gpt", + "value": "From <23.2> to <28.4>." + } + ] + }, + { + "video_id": "XRI58", + "question_id": "XRI58", + "video_file": "Charades/videos/XRI58.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.2> to <18.9>." + } + ] + }, + { + "video_id": "XRI58", + "question_id": "XRI58", + "video_file": "Charades/videos/XRI58.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats food from a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.2> to <18.9>." + } + ] + }, + { + "video_id": "XRI58", + "question_id": "XRI58", + "video_file": "Charades/videos/XRI58.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs across the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.4> to <24.6>." + } + ] + }, + { + "video_id": "XRI58", + "question_id": "XRI58", + "video_file": "Charades/videos/XRI58.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person holding a paper runs across a room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.4> to <24.6>." + } + ] + }, + { + "video_id": "XRI58", + "question_id": "XRI58", + "video_file": "Charades/videos/XRI58.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'sits in a chair .the person takes a bottle'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <12.5>." + } + ] + }, + { + "video_id": "XRI58", + "question_id": "XRI58", + "video_file": "Charades/videos/XRI58.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.4> to <24.6>." + } + ] + }, + { + "video_id": "64IAU", + "question_id": "64IAU", + "video_file": "Charades/videos/64IAU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person is opening the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <23.8>." + } + ] + }, + { + "video_id": "64IAU", + "question_id": "64IAU", + "video_file": "Charades/videos/64IAU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts laughing'." + }, + { + "from": "gpt", + "value": "From <24.0> to <32.0>." + } + ] + }, + { + "video_id": "64IAU", + "question_id": "64IAU", + "video_file": "Charades/videos/64IAU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looking out into an open window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.5> to <5.1>." + } + ] + }, + { + "video_id": "64IAU", + "question_id": "64IAU", + "video_file": "Charades/videos/64IAU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looks out the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <5.1>." + } + ] + }, + { + "video_id": "33I4D", + "question_id": "33I4D", + "video_file": "Charades/videos/33I4D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'stirs it .the person takes a drink of the cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <38.6> to <47.6>." + } + ] + }, + { + "video_id": "33I4D", + "question_id": "33I4D", + "video_file": "Charades/videos/33I4D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take a drink from the cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <38.6> to <47.6>." + } + ] + }, + { + "video_id": "33I4D", + "question_id": "33I4D", + "video_file": "Charades/videos/33I4D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put the cup on the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.8> to <32.6>." + } + ] + }, + { + "video_id": "2YJUQ", + "question_id": "2YJUQ", + "video_file": "Charades/videos/2YJUQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.1> to <23.5>." + } + ] + }, + { + "video_id": "2YJUQ", + "question_id": "2YJUQ", + "video_file": "Charades/videos/2YJUQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pours a glass of water'." + }, + { + "from": "gpt", + "value": "From <7.2> to <15.7>." + } + ] + }, + { + "video_id": "2YJUQ", + "question_id": "2YJUQ", + "video_file": "Charades/videos/2YJUQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person looks through a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.6> to <31.6>." + } + ] + }, + { + "video_id": "2YJUQ", + "question_id": "2YJUQ", + "video_file": "Charades/videos/2YJUQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sits down at a dining room table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.3> to <9.7>." + } + ] + }, + { + "video_id": "2YJUQ", + "question_id": "2YJUQ", + "video_file": "Charades/videos/2YJUQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looks out the window'." + }, + { + "from": "gpt", + "value": "From <23.6> to <31.6>." + } + ] + }, + { + "video_id": "2YJUQ", + "question_id": "2YJUQ", + "video_file": "Charades/videos/2YJUQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is pouring water into a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.2> to <15.7>." + } + ] + }, + { + "video_id": "YO9NC", + "question_id": "YO9NC", + "video_file": "Charades/videos/YO9NC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person picks up a book to read'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.5> to <31.0>." + } + ] + }, + { + "video_id": "YO9NC", + "question_id": "YO9NC", + "video_file": "Charades/videos/YO9NC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put their clothes in the closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.6> to <19.9>." + } + ] + }, + { + "video_id": "YO9NC", + "question_id": "YO9NC", + "video_file": "Charades/videos/YO9NC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they start undressing by taking their jacket off' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.6> to <11.6>." + } + ] + }, + { + "video_id": "YO9NC", + "question_id": "YO9NC", + "video_file": "Charades/videos/YO9NC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begins to undress' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <11.6>." + } + ] + }, + { + "video_id": "YO9NC", + "question_id": "YO9NC", + "video_file": "Charades/videos/YO9NC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person start watching the book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.5> to <31.0>." + } + ] + }, + { + "video_id": "JJGEU", + "question_id": "JJGEU", + "video_file": "Charades/videos/JJGEU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats food from a plate' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.2> to <17.6>." + } + ] + }, + { + "video_id": "K14NJ", + "question_id": "K14NJ", + "video_file": "Charades/videos/K14NJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits on the floor'." + }, + { + "from": "gpt", + "value": "From <13.9> to <25.7>." + } + ] + }, + { + "video_id": "K14NJ", + "question_id": "K14NJ", + "video_file": "Charades/videos/K14NJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they sit on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.9> to <25.7>." + } + ] + }, + { + "video_id": "EFSIN", + "question_id": "EFSIN", + "video_file": "Charades/videos/EFSIN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is washing dishes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.9>." + } + ] + }, + { + "video_id": "YIV35", + "question_id": "YIV35", + "video_file": "Charades/videos/YIV35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a bite of the sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.2> to <27.5>." + } + ] + }, + { + "video_id": "YIV35", + "question_id": "YIV35", + "video_file": "Charades/videos/YIV35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.6> to <7.3>." + } + ] + }, + { + "video_id": "YIV35", + "question_id": "YIV35", + "video_file": "Charades/videos/YIV35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts the book down sneezes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.0> to <23.2>." + } + ] + }, + { + "video_id": "YIV35", + "question_id": "YIV35", + "video_file": "Charades/videos/YIV35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.6> to <25.6>." + } + ] + }, + { + "video_id": "X0VAO", + "question_id": "X0VAO", + "video_file": "Charades/videos/X0VAO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a camera' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.2> to <32.0>." + } + ] + }, + { + "video_id": "DUAOJ", + "question_id": "DUAOJ", + "video_file": "Charades/videos/DUAOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sit down in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.9> to <16.6>." + } + ] + }, + { + "video_id": "DUAOJ", + "question_id": "DUAOJ", + "video_file": "Charades/videos/DUAOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a sneezing person walks into a kitchen eating a cookie' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.6> to <13.7>." + } + ] + }, + { + "video_id": "DUAOJ", + "question_id": "DUAOJ", + "video_file": "Charades/videos/DUAOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turn on a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "DUAOJ", + "question_id": "DUAOJ", + "video_file": "Charades/videos/DUAOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.9> to <16.6>." + } + ] + }, + { + "video_id": "DUAOJ", + "question_id": "DUAOJ", + "video_file": "Charades/videos/DUAOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a man sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <13.7>." + } + ] + }, + { + "video_id": "DUAOJ", + "question_id": "DUAOJ", + "video_file": "Charades/videos/DUAOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns on a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "DUAOJ", + "question_id": "DUAOJ", + "video_file": "Charades/videos/DUAOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turning a light on'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "DUAOJ", + "question_id": "DUAOJ", + "video_file": "Charades/videos/DUAOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <13.7>." + } + ] + }, + { + "video_id": "DUAOJ", + "question_id": "DUAOJ", + "video_file": "Charades/videos/DUAOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eat something' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.8> to <18.5>." + } + ] + }, + { + "video_id": "UUPEV", + "question_id": "UUPEV", + "video_file": "Charades/videos/UUPEV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a bag of groceries' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "UUPEV", + "question_id": "UUPEV", + "video_file": "Charades/videos/UUPEV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "8SAEO", + "question_id": "8SAEO", + "video_file": "Charades/videos/8SAEO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person undressed in a bathroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.2> to <9.5>." + } + ] + }, + { + "video_id": "QXT9W", + "question_id": "QXT9W", + "video_file": "Charades/videos/QXT9W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'opens it .the person lays the book on a stand'." + }, + { + "from": "gpt", + "value": "From <16.6> to <24.2>." + } + ] + }, + { + "video_id": "QXT9W", + "question_id": "QXT9W", + "video_file": "Charades/videos/QXT9W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person reads a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <38.9>." + } + ] + }, + { + "video_id": "QXT9W", + "question_id": "QXT9W", + "video_file": "Charades/videos/QXT9W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person look out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <43.1> to <62.7>." + } + ] + }, + { + "video_id": "QXT9W", + "question_id": "QXT9W", + "video_file": "Charades/videos/QXT9W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person sits down on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.7> to <40.9>." + } + ] + }, + { + "video_id": "WCLGY", + "question_id": "WCLGY", + "video_file": "Charades/videos/WCLGY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throwing clothes on the bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.4> to <22.2>." + } + ] + }, + { + "video_id": "WCLGY", + "question_id": "WCLGY", + "video_file": "Charades/videos/WCLGY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is throwing clothes on a bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.4> to <22.2>." + } + ] + }, + { + "video_id": "Z9B4Y", + "question_id": "Z9B4Y", + "video_file": "Charades/videos/Z9B4Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put their book on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.9> to <16.6>." + } + ] + }, + { + "video_id": "NY5LX", + "question_id": "NY5LX", + "video_file": "Charades/videos/NY5LX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting clothes in a bag in a wardrobe' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.8> to <36.0>." + } + ] + }, + { + "video_id": "K0Z9V", + "question_id": "K0Z9V", + "video_file": "Charades/videos/K0Z9V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is cooking in the kitchen'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "K0Z9V", + "question_id": "K0Z9V", + "video_file": "Charades/videos/K0Z9V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking on the stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "K0Z9V", + "question_id": "K0Z9V", + "video_file": "Charades/videos/K0Z9V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'this person is cooking in the kitchen' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "K0Z9V", + "question_id": "K0Z9V", + "video_file": "Charades/videos/K0Z9V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts sneezing'." + }, + { + "from": "gpt", + "value": "From <8.1> to <15.8>." + } + ] + }, + { + "video_id": "K0Z9V", + "question_id": "K0Z9V", + "video_file": "Charades/videos/K0Z9V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person starts sneezing repeatedly from the pepper' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <15.8>." + } + ] + }, + { + "video_id": "LK3LF", + "question_id": "LK3LF", + "video_file": "Charades/videos/LK3LF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put a pillow on to the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.9> to <25.4>." + } + ] + }, + { + "video_id": "6EHR7", + "question_id": "6EHR7", + "video_file": "Charades/videos/6EHR7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts on shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.0> to <36.0>." + } + ] + }, + { + "video_id": "6EHR7", + "question_id": "6EHR7", + "video_file": "Charades/videos/6EHR7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts on some shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.0> to <36.0>." + } + ] + }, + { + "video_id": "3W6TL", + "question_id": "3W6TL", + "video_file": "Charades/videos/3W6TL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <12.1> to <19.9>." + } + ] + }, + { + "video_id": "3W6TL", + "question_id": "3W6TL", + "video_file": "Charades/videos/3W6TL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a pillow off of the bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "3W6TL", + "question_id": "3W6TL", + "video_file": "Charades/videos/3W6TL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <19.9>." + } + ] + }, + { + "video_id": "3W6TL", + "question_id": "3W6TL", + "video_file": "Charades/videos/3W6TL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes a pillow to the wardrobe' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "DUHWU", + "question_id": "DUHWU", + "video_file": "Charades/videos/DUHWU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person someone runs up to their door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.2>." + } + ] + }, + { + "video_id": "DUHWU", + "question_id": "DUHWU", + "video_file": "Charades/videos/DUHWU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is running to open the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.3> to <16.9>." + } + ] + }, + { + "video_id": "82X3Z", + "question_id": "82X3Z", + "video_file": "Charades/videos/82X3Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is watching a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.1>." + } + ] + }, + { + "video_id": "82X3Z", + "question_id": "82X3Z", + "video_file": "Charades/videos/82X3Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating some food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.3> to <24.7>." + } + ] + }, + { + "video_id": "82X3Z", + "question_id": "82X3Z", + "video_file": "Charades/videos/82X3Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they eat some food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.3> to <24.7>." + } + ] + }, + { + "video_id": "TD3HC", + "question_id": "TD3HC", + "video_file": "Charades/videos/TD3HC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person throws a blanket on their bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "TD3HC", + "question_id": "TD3HC", + "video_file": "Charades/videos/TD3HC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is throwing a blanket on a bed'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "NSF4N", + "question_id": "NSF4N", + "video_file": "Charades/videos/NSF4N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person cook it on the stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.2> to <37.0>." + } + ] + }, + { + "video_id": "NSF4N", + "question_id": "NSF4N", + "video_file": "Charades/videos/NSF4N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens the refrigerator'." + }, + { + "from": "gpt", + "value": "From <2.1> to <9.0>." + } + ] + }, + { + "video_id": "NSF4N", + "question_id": "NSF4N", + "video_file": "Charades/videos/NSF4N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person begins cooking the items on the stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.2> to <37.0>." + } + ] + }, + { + "video_id": "RGIZ7", + "question_id": "RGIZ7", + "video_file": "Charades/videos/RGIZ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person works on a laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.1>." + } + ] + }, + { + "video_id": "RGIZ7", + "question_id": "RGIZ7", + "video_file": "Charades/videos/RGIZ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person closes a laptop'." + }, + { + "from": "gpt", + "value": "From <9.2> to <16.3>." + } + ] + }, + { + "video_id": "RGIZ7", + "question_id": "RGIZ7", + "video_file": "Charades/videos/RGIZ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looking out the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.4> to <25.0>." + } + ] + }, + { + "video_id": "RGIZ7", + "question_id": "RGIZ7", + "video_file": "Charades/videos/RGIZ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they look out a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.4> to <25.0>." + } + ] + }, + { + "video_id": "583F7", + "question_id": "583F7", + "video_file": "Charades/videos/583F7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a bag of chips out'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.3>." + } + ] + }, + { + "video_id": "583F7", + "question_id": "583F7", + "video_file": "Charades/videos/583F7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opened the bag'." + }, + { + "from": "gpt", + "value": "From <10.1> to <21.9>." + } + ] + }, + { + "video_id": "US6P9", + "question_id": "US6P9", + "video_file": "Charades/videos/US6P9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person turns on a light'." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.7>." + } + ] + }, + { + "video_id": "US6P9", + "question_id": "US6P9", + "video_file": "Charades/videos/US6P9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.7>." + } + ] + }, + { + "video_id": "US6P9", + "question_id": "US6P9", + "video_file": "Charades/videos/US6P9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person he puts the picture down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.0> to <22.8>." + } + ] + }, + { + "video_id": "QIYU4", + "question_id": "QIYU4", + "video_file": "Charades/videos/QIYU4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turning off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.6> to <6.4>." + } + ] + }, + { + "video_id": "QIYU4", + "question_id": "QIYU4", + "video_file": "Charades/videos/QIYU4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the door to a closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.0> to <26.0>." + } + ] + }, + { + "video_id": "QIYU4", + "question_id": "QIYU4", + "video_file": "Charades/videos/QIYU4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they turn off the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.6> to <6.4>." + } + ] + }, + { + "video_id": "IBNKL", + "question_id": "IBNKL", + "video_file": "Charades/videos/IBNKL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'there is a person cookingon the stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "IBNKL", + "question_id": "IBNKL", + "video_file": "Charades/videos/IBNKL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts some dishes in the sink'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.8> to <13.5>." + } + ] + }, + { + "video_id": "IBNKL", + "question_id": "IBNKL", + "video_file": "Charades/videos/IBNKL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'that person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <10.7>." + } + ] + }, + { + "video_id": "IBNKL", + "question_id": "IBNKL", + "video_file": "Charades/videos/IBNKL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person continue cooking' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "IBNKL", + "question_id": "IBNKL", + "video_file": "Charades/videos/IBNKL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is cooking at a stove'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "5OIQ7", + "question_id": "5OIQ7", + "video_file": "Charades/videos/5OIQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person awakens in a bedroom holding a phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "7BQ1O", + "question_id": "7BQ1O", + "video_file": "Charades/videos/7BQ1O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they both laugh' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.1> to <31.0>." + } + ] + }, + { + "video_id": "7BQ1O", + "question_id": "7BQ1O", + "video_file": "Charades/videos/7BQ1O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <6.6>." + } + ] + }, + { + "video_id": "7BQ1O", + "question_id": "7BQ1O", + "video_file": "Charades/videos/7BQ1O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turn on the lights'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <6.6>." + } + ] + }, + { + "video_id": "7BQ1O", + "question_id": "7BQ1O", + "video_file": "Charades/videos/7BQ1O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they open the garage door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <12.5>." + } + ] + }, + { + "video_id": "Y4MW3", + "question_id": "Y4MW3", + "video_file": "Charades/videos/Y4MW3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person a phone is taken out'." + }, + { + "from": "gpt", + "value": "From <13.6> to <19.8>." + } + ] + }, + { + "video_id": "Y4MW3", + "question_id": "Y4MW3", + "video_file": "Charades/videos/Y4MW3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person watching themselves in a mirror'." + }, + { + "from": "gpt", + "value": "From <9.1> to <18.1>." + } + ] + }, + { + "video_id": "NKNVR", + "question_id": "NKNVR", + "video_file": "Charades/videos/NKNVR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating a cookie' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.7> to <34.1>." + } + ] + }, + { + "video_id": "NKNVR", + "question_id": "NKNVR", + "video_file": "Charades/videos/NKNVR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a second person eats food'." + }, + { + "from": "gpt", + "value": "From <24.7> to <34.1>." + } + ] + }, + { + "video_id": "VX82P", + "question_id": "VX82P", + "video_file": "Charades/videos/VX82P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sneezing on the food'." + }, + { + "from": "gpt", + "value": "From <7.3> to <19.8>." + } + ] + }, + { + "video_id": "ELBLK", + "question_id": "ELBLK", + "video_file": "Charades/videos/ELBLK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts eating it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.1> to <31.0>." + } + ] + }, + { + "video_id": "ZUU2X", + "question_id": "ZUU2X", + "video_file": "Charades/videos/ZUU2X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts a bag on the sofa' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.6> to <25.3>." + } + ] + }, + { + "video_id": "ZUU2X", + "question_id": "ZUU2X", + "video_file": "Charades/videos/ZUU2X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the second person puts a bag next to the bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.6> to <25.3>." + } + ] + }, + { + "video_id": "ZUU2X", + "question_id": "ZUU2X", + "video_file": "Charades/videos/ZUU2X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns on the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.7> to <20.0>." + } + ] + }, + { + "video_id": "ZUU2X", + "question_id": "ZUU2X", + "video_file": "Charades/videos/ZUU2X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person playing with a phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.8>." + } + ] + }, + { + "video_id": "7ZCXJ", + "question_id": "7ZCXJ", + "video_file": "Charades/videos/7ZCXJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person so they begin undressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.7> to <23.5>." + } + ] + }, + { + "video_id": "9GS13", + "question_id": "9GS13", + "video_file": "Charades/videos/9GS13.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person washes hands' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.5> to <57.5>." + } + ] + }, + { + "video_id": "9GS13", + "question_id": "9GS13", + "video_file": "Charades/videos/9GS13.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person dumps the clothes in the washer' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <22.4>." + } + ] + }, + { + "video_id": "9GS13", + "question_id": "9GS13", + "video_file": "Charades/videos/9GS13.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts clothes in a dishwasher'." + }, + { + "from": "gpt", + "value": "From <7.2> to <20.1>." + } + ] + }, + { + "video_id": "9GS13", + "question_id": "9GS13", + "video_file": "Charades/videos/9GS13.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person washes hands in the sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.5> to <57.5>." + } + ] + }, + { + "video_id": "FQGQS", + "question_id": "FQGQS", + "video_file": "Charades/videos/FQGQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a closet door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "FQGQS", + "question_id": "FQGQS", + "video_file": "Charades/videos/FQGQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws a pair of shoes inside' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.7> to <5.3>." + } + ] + }, + { + "video_id": "QUGLP", + "question_id": "QUGLP", + "video_file": "Charades/videos/QUGLP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.7> to <10.2>." + } + ] + }, + { + "video_id": "4JQKA", + "question_id": "4JQKA", + "video_file": "Charades/videos/4JQKA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the cabinet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.6> to <33.9>." + } + ] + }, + { + "video_id": "4JQKA", + "question_id": "4JQKA", + "video_file": "Charades/videos/4JQKA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person undresses in front of a bathroom mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.8> to <12.0>." + } + ] + }, + { + "video_id": "4JQKA", + "question_id": "4JQKA", + "video_file": "Charades/videos/4JQKA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the mirrored -door medicine cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <15.0>." + } + ] + }, + { + "video_id": "4JQKA", + "question_id": "4JQKA", + "video_file": "Charades/videos/4JQKA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open the mirror to reveal a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.0> to <15.0>." + } + ] + }, + { + "video_id": "KQYR0", + "question_id": "KQYR0", + "video_file": "Charades/videos/KQYR0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts them on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.9> to <11.9>." + } + ] + }, + { + "video_id": "KQYR0", + "question_id": "KQYR0", + "video_file": "Charades/videos/KQYR0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take the clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.5> to <25.0>." + } + ] + }, + { + "video_id": "KQYR0", + "question_id": "KQYR0", + "video_file": "Charades/videos/KQYR0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person picks up a book puts it on a desk' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.0> to <23.4>." + } + ] + }, + { + "video_id": "ATUOW", + "question_id": "ATUOW", + "video_file": "Charades/videos/ATUOW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts papers on their shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <20.0>." + } + ] + }, + { + "video_id": "ATUOW", + "question_id": "ATUOW", + "video_file": "Charades/videos/ATUOW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put it on the shelf'." + }, + { + "from": "gpt", + "value": "From <12.3> to <20.0>." + } + ] + }, + { + "video_id": "QXCUP", + "question_id": "QXCUP", + "video_file": "Charades/videos/QXCUP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a few pictures with a camera' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "QXCUP", + "question_id": "QXCUP", + "video_file": "Charades/videos/QXCUP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts clothes away in a closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.6> to <38.4>." + } + ] + }, + { + "video_id": "QXCUP", + "question_id": "QXCUP", + "video_file": "Charades/videos/QXCUP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person close door'." + }, + { + "from": "gpt", + "value": "From <34.4> to <40.0>." + } + ] + }, + { + "video_id": "QXCUP", + "question_id": "QXCUP", + "video_file": "Charades/videos/QXCUP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting the camera down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <19.8>." + } + ] + }, + { + "video_id": "QXCUP", + "question_id": "QXCUP", + "video_file": "Charades/videos/QXCUP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts the camera down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.0> to <19.8>." + } + ] + }, + { + "video_id": "BQ6UI", + "question_id": "BQ6UI", + "video_file": "Charades/videos/BQ6UI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person washes their hands in the sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.6> to <21.7>." + } + ] + }, + { + "video_id": "BQ6UI", + "question_id": "BQ6UI", + "video_file": "Charades/videos/BQ6UI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person going into the bathroom to wash hands'." + }, + { + "from": "gpt", + "value": "From <7.6> to <21.7>." + } + ] + }, + { + "video_id": "8M1TG", + "question_id": "8M1TG", + "video_file": "Charades/videos/8M1TG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person another closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <23.5>." + } + ] + }, + { + "video_id": "8M1TG", + "question_id": "8M1TG", + "video_file": "Charades/videos/8M1TG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting a broom away' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.4> to <20.6>." + } + ] + }, + { + "video_id": "8M1TG", + "question_id": "8M1TG", + "video_file": "Charades/videos/8M1TG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'when another person walks over to close a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <23.5>." + } + ] + }, + { + "video_id": "VXNGC", + "question_id": "VXNGC", + "video_file": "Charades/videos/VXNGC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they are smiling as they work on their laptop'." + }, + { + "from": "gpt", + "value": "From <5.3> to <17.3>." + } + ] + }, + { + "video_id": "VXNGC", + "question_id": "VXNGC", + "video_file": "Charades/videos/VXNGC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is smiling at a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <17.3>." + } + ] + }, + { + "video_id": "GFH4E", + "question_id": "GFH4E", + "video_file": "Charades/videos/GFH4E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looking out a window'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.6>." + } + ] + }, + { + "video_id": "GFH4E", + "question_id": "GFH4E", + "video_file": "Charades/videos/GFH4E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is holding a cup'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.7>." + } + ] + }, + { + "video_id": "GFH4E", + "question_id": "GFH4E", + "video_file": "Charades/videos/GFH4E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.7>." + } + ] + }, + { + "video_id": "68YR6", + "question_id": "68YR6", + "video_file": "Charades/videos/68YR6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person opens the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.2> to <19.5>." + } + ] + }, + { + "video_id": "68YR6", + "question_id": "68YR6", + "video_file": "Charades/videos/68YR6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts washing the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.0> to <32.0>." + } + ] + }, + { + "video_id": "9QVWC", + "question_id": "9QVWC", + "video_file": "Charades/videos/9QVWC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person start sneezing'." + }, + { + "from": "gpt", + "value": "From <19.7> to <33.0>." + } + ] + }, + { + "video_id": "9QVWC", + "question_id": "9QVWC", + "video_file": "Charades/videos/9QVWC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is standing fixing a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.7>." + } + ] + }, + { + "video_id": "LQK90", + "question_id": "LQK90", + "video_file": "Charades/videos/LQK90.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pours a drink into a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.9> to <31.0>." + } + ] + }, + { + "video_id": "4NRPM", + "question_id": "4NRPM", + "video_file": "Charades/videos/4NRPM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they take their shoes that they are wearing off'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <27.5>." + } + ] + }, + { + "video_id": "20YRU", + "question_id": "20YRU", + "video_file": "Charades/videos/20YRU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a box'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "Z8HPQ", + "question_id": "Z8HPQ", + "video_file": "Charades/videos/Z8HPQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person leaving the fridge door open'." + }, + { + "from": "gpt", + "value": "From <7.9> to <14.1>." + } + ] + }, + { + "video_id": "Z8HPQ", + "question_id": "Z8HPQ", + "video_file": "Charades/videos/Z8HPQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opens the refrigerator'." + }, + { + "from": "gpt", + "value": "From <7.9> to <14.1>." + } + ] + }, + { + "video_id": "Z8HPQ", + "question_id": "Z8HPQ", + "video_file": "Charades/videos/Z8HPQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person sees the open refrigerator door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.9> to <14.1>." + } + ] + }, + { + "video_id": "Z8HPQ", + "question_id": "Z8HPQ", + "video_file": "Charades/videos/Z8HPQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the refrigerator'." + }, + { + "from": "gpt", + "value": "From <7.9> to <14.1>." + } + ] + }, + { + "video_id": "YZI2P", + "question_id": "YZI2P", + "video_file": "Charades/videos/YZI2P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they close the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.7> to <10.2>." + } + ] + }, + { + "video_id": "YZI2P", + "question_id": "YZI2P", + "video_file": "Charades/videos/YZI2P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is seen closing the refrigerator door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.7> to <10.2>." + } + ] + }, + { + "video_id": "JISS4", + "question_id": "JISS4", + "video_file": "Charades/videos/JISS4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person stands by a door laughing throwing a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "JISS4", + "question_id": "JISS4", + "video_file": "Charades/videos/JISS4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throwing a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "JISS4", + "question_id": "JISS4", + "video_file": "Charades/videos/JISS4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person laughs' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.9>." + } + ] + }, + { + "video_id": "FM5D5", + "question_id": "FM5D5", + "video_file": "Charades/videos/FM5D5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a bag of clothes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <17.5>." + } + ] + }, + { + "video_id": "FM5D5", + "question_id": "FM5D5", + "video_file": "Charades/videos/FM5D5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person throws down the bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <11.8>." + } + ] + }, + { + "video_id": "FM5D5", + "question_id": "FM5D5", + "video_file": "Charades/videos/FM5D5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts bag down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <17.5>." + } + ] + }, + { + "video_id": "2S6V7", + "question_id": "2S6V7", + "video_file": "Charades/videos/2S6V7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes their laptop from the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.5> to <30.1>." + } + ] + }, + { + "video_id": "2S6V7", + "question_id": "2S6V7", + "video_file": "Charades/videos/2S6V7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they take a laptop from the counter'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.5> to <30.1>." + } + ] + }, + { + "video_id": "2S6V7", + "question_id": "2S6V7", + "video_file": "Charades/videos/2S6V7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person begins laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <26.5>." + } + ] + }, + { + "video_id": "T8V10", + "question_id": "T8V10", + "video_file": "Charades/videos/T8V10.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take a drink from a cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.2> to <25.4>." + } + ] + }, + { + "video_id": "RHAJX", + "question_id": "RHAJX", + "video_file": "Charades/videos/RHAJX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put their shoes on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <15.8>." + } + ] + }, + { + "video_id": "RHAJX", + "question_id": "RHAJX", + "video_file": "Charades/videos/RHAJX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is awakening from laying on a bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "RHAJX", + "question_id": "RHAJX", + "video_file": "Charades/videos/RHAJX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take a laptop from a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.2> to <29.4>." + } + ] + }, + { + "video_id": "65M4K", + "question_id": "65M4K", + "video_file": "Charades/videos/65M4K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person getting food out cooking the food on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.5> to <32.0>." + } + ] + }, + { + "video_id": "65M4K", + "question_id": "65M4K", + "video_file": "Charades/videos/65M4K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walking in the room open a refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "65M4K", + "question_id": "65M4K", + "video_file": "Charades/videos/65M4K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking at a stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.5> to <32.0>." + } + ] + }, + { + "video_id": "8L08M", + "question_id": "8L08M", + "video_file": "Charades/videos/8L08M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits down on the bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.8> to <18.6>." + } + ] + }, + { + "video_id": "8L08M", + "question_id": "8L08M", + "video_file": "Charades/videos/8L08M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding the shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "8L08M", + "question_id": "8L08M", + "video_file": "Charades/videos/8L08M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is standing in a bedroom holding their shoes'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "8L08M", + "question_id": "8L08M", + "video_file": "Charades/videos/8L08M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they dress'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <16.0>." + } + ] + }, + { + "video_id": "FKZQN", + "question_id": "FKZQN", + "video_file": "Charades/videos/FKZQN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts dressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <24.6>." + } + ] + }, + { + "video_id": "FKZQN", + "question_id": "FKZQN", + "video_file": "Charades/videos/FKZQN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person started dressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <24.6>." + } + ] + }, + { + "video_id": "OA9W5", + "question_id": "OA9W5", + "video_file": "Charades/videos/OA9W5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person in a closet is smiling'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "OA9W5", + "question_id": "OA9W5", + "video_file": "Charades/videos/OA9W5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opening a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.6> to <18.6>." + } + ] + }, + { + "video_id": "OA9W5", + "question_id": "OA9W5", + "video_file": "Charades/videos/OA9W5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is sitting in a closet eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "OA9W5", + "question_id": "OA9W5", + "video_file": "Charades/videos/OA9W5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "884C0", + "question_id": "884C0", + "video_file": "Charades/videos/884C0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person holding a broom walks into the kitchen' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.6>." + } + ] + }, + { + "video_id": "884C0", + "question_id": "884C0", + "video_file": "Charades/videos/884C0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sneezing at the same time' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "884C0", + "question_id": "884C0", + "video_file": "Charades/videos/884C0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks in a room holding a broom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.6>." + } + ] + }, + { + "video_id": "RRFE7", + "question_id": "RRFE7", + "video_file": "Charades/videos/RRFE7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is laughing'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.1>." + } + ] + }, + { + "video_id": "PM107", + "question_id": "PM107", + "video_file": "Charades/videos/PM107.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts watching out of the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.8> to <33.0>." + } + ] + }, + { + "video_id": "PM107", + "question_id": "PM107", + "video_file": "Charades/videos/PM107.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looks through a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.8> to <33.0>." + } + ] + }, + { + "video_id": "PM107", + "question_id": "PM107", + "video_file": "Charades/videos/PM107.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is holding a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.4> to <18.2>." + } + ] + }, + { + "video_id": "9JZCC", + "question_id": "9JZCC", + "video_file": "Charades/videos/9JZCC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person smiling at something on their phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.8> to <18.5>." + } + ] + }, + { + "video_id": "9JZCC", + "question_id": "9JZCC", + "video_file": "Charades/videos/9JZCC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person lies down on the bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.8> to <31.0>." + } + ] + }, + { + "video_id": "4CSXJ", + "question_id": "4CSXJ", + "video_file": "Charades/videos/4CSXJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person throws the pillow down the staircase'." + }, + { + "from": "gpt", + "value": "From <5.7> to <10.9>." + } + ] + }, + { + "video_id": "4CSXJ", + "question_id": "4CSXJ", + "video_file": "Charades/videos/4CSXJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws the pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.7> to <10.9>." + } + ] + }, + { + "video_id": "4CSXJ", + "question_id": "4CSXJ", + "video_file": "Charades/videos/4CSXJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting on the staircase holding a pillow'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.0>." + } + ] + }, + { + "video_id": "DHVYY", + "question_id": "DHVYY", + "video_file": "Charades/videos/DHVYY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is washing dishes with a towel' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.3> to <14.5>." + } + ] + }, + { + "video_id": "DHVYY", + "question_id": "DHVYY", + "video_file": "Charades/videos/DHVYY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person washing dishes'." + }, + { + "from": "gpt", + "value": "From <7.3> to <14.5>." + } + ] + }, + { + "video_id": "MQ9PZ", + "question_id": "MQ9PZ", + "video_file": "Charades/videos/MQ9PZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they close the cabinet door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "MQ9PZ", + "question_id": "MQ9PZ", + "video_file": "Charades/videos/MQ9PZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening up a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.1> to <11.8>." + } + ] + }, + { + "video_id": "MQ9PZ", + "question_id": "MQ9PZ", + "video_file": "Charades/videos/MQ9PZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they close the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "MQ9PZ", + "question_id": "MQ9PZ", + "video_file": "Charades/videos/MQ9PZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open up a book'." + }, + { + "from": "gpt", + "value": "From <6.1> to <11.8>." + } + ] + }, + { + "video_id": "8J5O9", + "question_id": "8J5O9", + "video_file": "Charades/videos/8J5O9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating out of a bowl' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <31.0>." + } + ] + }, + { + "video_id": "YXN50", + "question_id": "YXN50", + "video_file": "Charades/videos/YXN50.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the water on the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <25.2>." + } + ] + }, + { + "video_id": "FGY8R", + "question_id": "FGY8R", + "video_file": "Charades/videos/FGY8R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <17.7>." + } + ] + }, + { + "video_id": "FGY8R", + "question_id": "FGY8R", + "video_file": "Charades/videos/FGY8R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person working on the laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "FGY8R", + "question_id": "FGY8R", + "video_file": "Charades/videos/FGY8R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the laptop down on the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <24.0>." + } + ] + }, + { + "video_id": "FGY8R", + "question_id": "FGY8R", + "video_file": "Charades/videos/FGY8R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens their laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "P0DXX", + "question_id": "P0DXX", + "video_file": "Charades/videos/P0DXX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person puts some trash in a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.8> to <39.5>." + } + ] + }, + { + "video_id": "P0DXX", + "question_id": "P0DXX", + "video_file": "Charades/videos/P0DXX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the box next to a television'." + }, + { + "from": "gpt", + "value": "From <31.8> to <39.5>." + } + ] + }, + { + "video_id": "P0DXX", + "question_id": "P0DXX", + "video_file": "Charades/videos/P0DXX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person throws clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <34.8>." + } + ] + }, + { + "video_id": "CSRZM", + "question_id": "CSRZM", + "video_file": "Charades/videos/CSRZM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes a sandwich out of the bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.2> to <21.5>." + } + ] + }, + { + "video_id": "CSRZM", + "question_id": "CSRZM", + "video_file": "Charades/videos/CSRZM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes out a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <21.5>." + } + ] + }, + { + "video_id": "CSRZM", + "question_id": "CSRZM", + "video_file": "Charades/videos/CSRZM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take a sandwich out of the bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <21.5>." + } + ] + }, + { + "video_id": "X8XQE", + "question_id": "X8XQE", + "video_file": "Charades/videos/X8XQE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person cooks something on the stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.8> to <29.7>." + } + ] + }, + { + "video_id": "X8XQE", + "question_id": "X8XQE", + "video_file": "Charades/videos/X8XQE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person cooking dinner' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.8> to <29.7>." + } + ] + }, + { + "video_id": "X8XQE", + "question_id": "X8XQE", + "video_file": "Charades/videos/X8XQE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a picture with their cellphone camera' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <31.6> to <40.3>." + } + ] + }, + { + "video_id": "TOUXU", + "question_id": "TOUXU", + "video_file": "Charades/videos/TOUXU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the front door'." + }, + { + "from": "gpt", + "value": "From <5.1> to <10.9>." + } + ] + }, + { + "video_id": "TOUXU", + "question_id": "TOUXU", + "video_file": "Charades/videos/TOUXU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <13.7>." + } + ] + }, + { + "video_id": "APVAD", + "question_id": "APVAD", + "video_file": "Charades/videos/APVAD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing in the basement' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.8>." + } + ] + }, + { + "video_id": "APVAD", + "question_id": "APVAD", + "video_file": "Charades/videos/APVAD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs out' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.1> to <31.0>." + } + ] + }, + { + "video_id": "APVAD", + "question_id": "APVAD", + "video_file": "Charades/videos/APVAD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person runs out the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.1> to <31.0>." + } + ] + }, + { + "video_id": "K0X80", + "question_id": "K0X80", + "video_file": "Charades/videos/K0X80.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person leaving the door open' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.1> to <8.6>." + } + ] + }, + { + "video_id": "K0X80", + "question_id": "K0X80", + "video_file": "Charades/videos/K0X80.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <8.6>." + } + ] + }, + { + "video_id": "H4ZD8", + "question_id": "H4ZD8", + "video_file": "Charades/videos/H4ZD8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begin to undress'." + }, + { + "from": "gpt", + "value": "From <8.0> to <15.0>." + } + ] + }, + { + "video_id": "H4ZD8", + "question_id": "H4ZD8", + "video_file": "Charades/videos/H4ZD8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person looks at the picture'." + }, + { + "from": "gpt", + "value": "From <4.4> to <11.5>." + } + ] + }, + { + "video_id": "H4ZD8", + "question_id": "H4ZD8", + "video_file": "Charades/videos/H4ZD8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looking at a picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.4> to <11.5>." + } + ] + }, + { + "video_id": "H4ZD8", + "question_id": "H4ZD8", + "video_file": "Charades/videos/H4ZD8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a book from the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.2> to <19.0>." + } + ] + }, + { + "video_id": "H4ZD8", + "question_id": "H4ZD8", + "video_file": "Charades/videos/H4ZD8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a picture from the table'." + }, + { + "from": "gpt", + "value": "From <4.0> to <10.0>." + } + ] + }, + { + "video_id": "H4ZD8", + "question_id": "H4ZD8", + "video_file": "Charades/videos/H4ZD8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is undressing in a dining room'." + }, + { + "from": "gpt", + "value": "From <8.0> to <15.0>." + } + ] + }, + { + "video_id": "CN108", + "question_id": "CN108", + "video_file": "Charades/videos/CN108.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <17.6>." + } + ] + }, + { + "video_id": "Z7QTA", + "question_id": "Z7QTA", + "video_file": "Charades/videos/Z7QTA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person box are put on a table in the closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.7> to <23.2>." + } + ] + }, + { + "video_id": "VTY0J", + "question_id": "VTY0J", + "video_file": "Charades/videos/VTY0J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sit down at a table'." + }, + { + "from": "gpt", + "value": "From <19.1> to <35.0>." + } + ] + }, + { + "video_id": "VTY0J", + "question_id": "VTY0J", + "video_file": "Charades/videos/VTY0J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits in a chair with a phone'." + }, + { + "from": "gpt", + "value": "From <20.1> to <35.0>." + } + ] + }, + { + "video_id": "VTY0J", + "question_id": "VTY0J", + "video_file": "Charades/videos/VTY0J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person opens a refrigerator to get food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.7>." + } + ] + }, + { + "video_id": "VTY0J", + "question_id": "VTY0J", + "video_file": "Charades/videos/VTY0J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person start playing on their phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.1> to <35.0>." + } + ] + }, + { + "video_id": "W75CV", + "question_id": "W75CV", + "video_file": "Charades/videos/W75CV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a smiling person awakens on the stairs holding a towel' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.6> to <23.0>." + } + ] + }, + { + "video_id": "LB4AR", + "question_id": "LB4AR", + "video_file": "Charades/videos/LB4AR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes off his shoes'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.0>." + } + ] + }, + { + "video_id": "LB4AR", + "question_id": "LB4AR", + "video_file": "Charades/videos/LB4AR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes out two bottles from the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.8> to <24.0>." + } + ] + }, + { + "video_id": "LB4AR", + "question_id": "LB4AR", + "video_file": "Charades/videos/LB4AR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person proceeds to open the refrigerator door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.6> to <16.7>." + } + ] + }, + { + "video_id": "LB4AR", + "question_id": "LB4AR", + "video_file": "Charades/videos/LB4AR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person throws their shoes through the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "6R7QY", + "question_id": "6R7QY", + "video_file": "Charades/videos/6R7QY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person stands on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "6R7QY", + "question_id": "6R7QY", + "video_file": "Charades/videos/6R7QY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person another is putting dishes in a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.4> to <31.0>." + } + ] + }, + { + "video_id": "6R7QY", + "question_id": "6R7QY", + "video_file": "Charades/videos/6R7QY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person someone stands on a chair to replace a lightbulb'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "QEFFJ", + "question_id": "QEFFJ", + "video_file": "Charades/videos/QEFFJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drink from a glass that was on the floor'." + }, + { + "from": "gpt", + "value": "From <12.6> to <25.0>." + } + ] + }, + { + "video_id": "QEFFJ", + "question_id": "QEFFJ", + "video_file": "Charades/videos/QEFFJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person drinks from a glass'." + }, + { + "from": "gpt", + "value": "From <12.6> to <25.0>." + } + ] + }, + { + "video_id": "TDHM0", + "question_id": "TDHM0", + "video_file": "Charades/videos/TDHM0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sneezing into a phone in the basement' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <13.7>." + } + ] + }, + { + "video_id": "MLK57", + "question_id": "MLK57", + "video_file": "Charades/videos/MLK57.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throwing a blanket on the floor'." + }, + { + "from": "gpt", + "value": "From <18.9> to <27.0>." + } + ] + }, + { + "video_id": "MLK57", + "question_id": "MLK57", + "video_file": "Charades/videos/MLK57.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person awakens in a bedroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.9> to <18.9>." + } + ] + }, + { + "video_id": "1NNFV", + "question_id": "1NNFV", + "video_file": "Charades/videos/1NNFV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person pours a cup of coffee' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.0>." + } + ] + }, + { + "video_id": "XET54", + "question_id": "XET54", + "video_file": "Charades/videos/XET54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person enters the kitchen holding a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <15.7>." + } + ] + }, + { + "video_id": "XET54", + "question_id": "XET54", + "video_file": "Charades/videos/XET54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is throwing a pillow onto the sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.9> to <17.9>." + } + ] + }, + { + "video_id": "XET54", + "question_id": "XET54", + "video_file": "Charades/videos/XET54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closing the door behind her'." + }, + { + "from": "gpt", + "value": "From <15.3> to <23.5>." + } + ] + }, + { + "video_id": "XET54", + "question_id": "XET54", + "video_file": "Charades/videos/XET54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closing the door behind them'." + }, + { + "from": "gpt", + "value": "From <15.3> to <23.5>." + } + ] + }, + { + "video_id": "XET54", + "question_id": "XET54", + "video_file": "Charades/videos/XET54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts a pillow on a sofa' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.4> to <18.7>." + } + ] + }, + { + "video_id": "9NV2G", + "question_id": "9NV2G", + "video_file": "Charades/videos/9NV2G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person awakens in bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.5> to <10.1>." + } + ] + }, + { + "video_id": "EBBKQ", + "question_id": "EBBKQ", + "video_file": "Charades/videos/EBBKQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sneezing into blanket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.7> to <15.0>." + } + ] + }, + { + "video_id": "EBBKQ", + "question_id": "EBBKQ", + "video_file": "Charades/videos/EBBKQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is watching a television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "RO9RP", + "question_id": "RO9RP", + "video_file": "Charades/videos/RO9RP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put the glass on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.5> to <27.1>." + } + ] + }, + { + "video_id": "RO9RP", + "question_id": "RO9RP", + "video_file": "Charades/videos/RO9RP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is pouring something in a glass'." + }, + { + "from": "gpt", + "value": "From <5.0> to <15.1>." + } + ] + }, + { + "video_id": "RO9RP", + "question_id": "RO9RP", + "video_file": "Charades/videos/RO9RP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pour it into the glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <15.1>." + } + ] + }, + { + "video_id": "RO9RP", + "question_id": "RO9RP", + "video_file": "Charades/videos/RO9RP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "DWBS3", + "question_id": "DWBS3", + "video_file": "Charades/videos/DWBS3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating some food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "DWBS3", + "question_id": "DWBS3", + "video_file": "Charades/videos/DWBS3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person is eating'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "6SWTR", + "question_id": "6SWTR", + "video_file": "Charades/videos/6SWTR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person read a book'." + }, + { + "from": "gpt", + "value": "From <18.9> to <27.0>." + } + ] + }, + { + "video_id": "S1LY3", + "question_id": "S1LY3", + "video_file": "Charades/videos/S1LY3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person watching a video of some kind on a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.3> to <38.0>." + } + ] + }, + { + "video_id": "S1LY3", + "question_id": "S1LY3", + "video_file": "Charades/videos/S1LY3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person watching a video on a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.3> to <38.0>." + } + ] + }, + { + "video_id": "5OB3W", + "question_id": "5OB3W", + "video_file": "Charades/videos/5OB3W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes bags' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.7> to <26.9>." + } + ] + }, + { + "video_id": "5OB3W", + "question_id": "5OB3W", + "video_file": "Charades/videos/5OB3W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they begin opening bags' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "MNC10", + "question_id": "MNC10", + "video_file": "Charades/videos/MNC10.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person throws a broom in the corner'." + }, + { + "from": "gpt", + "value": "From <7.6> to <14.0>." + } + ] + }, + { + "video_id": "XUYCA", + "question_id": "XUYCA", + "video_file": "Charades/videos/XUYCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sits in a chair eating' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.2>." + } + ] + }, + { + "video_id": "XUYCA", + "question_id": "XUYCA", + "video_file": "Charades/videos/XUYCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walks to a shelf to tidy up books'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.4> to <31.0>." + } + ] + }, + { + "video_id": "XUYCA", + "question_id": "XUYCA", + "video_file": "Charades/videos/XUYCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person was doen eating'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.2>." + } + ] + }, + { + "video_id": "9MOQ7", + "question_id": "9MOQ7", + "video_file": "Charades/videos/9MOQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put their shoes on' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.9> to <20.2>." + } + ] + }, + { + "video_id": "9MOQ7", + "question_id": "9MOQ7", + "video_file": "Charades/videos/9MOQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting on a bed sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "9MOQ7", + "question_id": "9MOQ7", + "video_file": "Charades/videos/9MOQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts their shoes on' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <20.2>." + } + ] + }, + { + "video_id": "ZKIBW", + "question_id": "ZKIBW", + "video_file": "Charades/videos/ZKIBW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person opens a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "9FEV7", + "question_id": "9FEV7", + "video_file": "Charades/videos/9FEV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes some clothes from a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "PIRMO", + "question_id": "PIRMO", + "video_file": "Charades/videos/PIRMO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes a cup off the shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.5> to <8.6>." + } + ] + }, + { + "video_id": "PIRMO", + "question_id": "PIRMO", + "video_file": "Charades/videos/PIRMO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking a glass of water the find a box'." + }, + { + "from": "gpt", + "value": "From <4.5> to <11.2>." + } + ] + }, + { + "video_id": "PIRMO", + "question_id": "PIRMO", + "video_file": "Charades/videos/PIRMO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person in a closet putting a blanket away'." + }, + { + "from": "gpt", + "value": "From <0.4> to <7.7>." + } + ] + }, + { + "video_id": "PIRMO", + "question_id": "PIRMO", + "video_file": "Charades/videos/PIRMO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person putting a towel on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <8.9>." + } + ] + }, + { + "video_id": "PIRMO", + "question_id": "PIRMO", + "video_file": "Charades/videos/PIRMO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a box off the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.7> to <17.1>." + } + ] + }, + { + "video_id": "USZCJ", + "question_id": "USZCJ", + "video_file": "Charades/videos/USZCJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens door to room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.3>." + } + ] + }, + { + "video_id": "YSBN0", + "question_id": "YSBN0", + "video_file": "Charades/videos/YSBN0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person stands at a stove loaded with dishes cooking' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "D67MQ", + "question_id": "D67MQ", + "video_file": "Charades/videos/D67MQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person is at the laundry sink sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.3> to <21.2>." + } + ] + }, + { + "video_id": "TCCQL", + "question_id": "TCCQL", + "video_file": "Charades/videos/TCCQL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person drinks from a bottle'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <25.0>." + } + ] + }, + { + "video_id": "TCCQL", + "question_id": "TCCQL", + "video_file": "Charades/videos/TCCQL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is holding a medicine bottle'." + }, + { + "from": "gpt", + "value": "From <17.0> to <28.0>." + } + ] + }, + { + "video_id": "FVMVY", + "question_id": "FVMVY", + "video_file": "Charades/videos/FVMVY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person sits in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <34.6> to <38.0>." + } + ] + }, + { + "video_id": "FVMVY", + "question_id": "FVMVY", + "video_file": "Charades/videos/FVMVY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a broom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <34.6> to <38.0>." + } + ] + }, + { + "video_id": "FVMVY", + "question_id": "FVMVY", + "video_file": "Charades/videos/FVMVY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits down in a chair'." + }, + { + "from": "gpt", + "value": "From <34.6> to <38.0>." + } + ] + }, + { + "video_id": "FVMVY", + "question_id": "FVMVY", + "video_file": "Charades/videos/FVMVY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is putting clothes in the washer'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.0>." + } + ] + }, + { + "video_id": "FVMVY", + "question_id": "FVMVY", + "video_file": "Charades/videos/FVMVY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person throwing clothes into the washing machine pours detergent' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <34.6> to <38.0>." + } + ] + }, + { + "video_id": "OINMN", + "question_id": "OINMN", + "video_file": "Charades/videos/OINMN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the door to a closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <33.1> to <40.0>." + } + ] + }, + { + "video_id": "OINMN", + "question_id": "OINMN", + "video_file": "Charades/videos/OINMN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws a box onto a shelf'." + }, + { + "from": "gpt", + "value": "From <27.4> to <35.5>." + } + ] + }, + { + "video_id": "MVPQF", + "question_id": "MVPQF", + "video_file": "Charades/videos/MVPQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the bags on the bottom shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <10.4>." + } + ] + }, + { + "video_id": "MVPQF", + "question_id": "MVPQF", + "video_file": "Charades/videos/MVPQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person puts clothes on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.9> to <10.4>." + } + ] + }, + { + "video_id": "PE41X", + "question_id": "PE41X", + "video_file": "Charades/videos/PE41X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a broom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "PE41X", + "question_id": "PE41X", + "video_file": "Charades/videos/PE41X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting on a chair holding a broom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "PE41X", + "question_id": "PE41X", + "video_file": "Charades/videos/PE41X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is in their kitchen sitting on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.1> to <31.0>." + } + ] + }, + { + "video_id": "AEZDA", + "question_id": "AEZDA", + "video_file": "Charades/videos/AEZDA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is eating some food from a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "AEZDA", + "question_id": "AEZDA", + "video_file": "Charades/videos/AEZDA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person eats out of a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "AD9MO", + "question_id": "AD9MO", + "video_file": "Charades/videos/AD9MO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turn off a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.7> to <27.0>." + } + ] + }, + { + "video_id": "AD9MO", + "question_id": "AD9MO", + "video_file": "Charades/videos/AD9MO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person off a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.7> to <27.0>." + } + ] + }, + { + "video_id": "AD9MO", + "question_id": "AD9MO", + "video_file": "Charades/videos/AD9MO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is putting a broom in the pantry'." + }, + { + "from": "gpt", + "value": "From <1.5> to <10.2>." + } + ] + }, + { + "video_id": "AD9MO", + "question_id": "AD9MO", + "video_file": "Charades/videos/AD9MO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a broom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "AD9MO", + "question_id": "AD9MO", + "video_file": "Charades/videos/AD9MO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is opening a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.4> to <9.5>." + } + ] + }, + { + "video_id": "AD9MO", + "question_id": "AD9MO", + "video_file": "Charades/videos/AD9MO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they close the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <25.6>." + } + ] + }, + { + "video_id": "Q3M8J", + "question_id": "Q3M8J", + "video_file": "Charades/videos/Q3M8J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits at a chair at a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.4> to <31.0>." + } + ] + }, + { + "video_id": "Q3M8J", + "question_id": "Q3M8J", + "video_file": "Charades/videos/Q3M8J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person sits in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.4> to <31.0>." + } + ] + }, + { + "video_id": "3DO3W", + "question_id": "3DO3W", + "video_file": "Charades/videos/3DO3W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the towel on the chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <19.7>." + } + ] + }, + { + "video_id": "6VT4U", + "question_id": "6VT4U", + "video_file": "Charades/videos/6VT4U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts it on a shelf'." + }, + { + "from": "gpt", + "value": "From <8.0> to <13.4>." + } + ] + }, + { + "video_id": "6VT4U", + "question_id": "6VT4U", + "video_file": "Charades/videos/6VT4U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person plays on a phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "3CFFJ", + "question_id": "3CFFJ", + "video_file": "Charades/videos/3CFFJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinks milk from a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.4> to <34.0>." + } + ] + }, + { + "video_id": "OZSB3", + "question_id": "OZSB3", + "video_file": "Charades/videos/OZSB3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opening a bag'." + }, + { + "from": "gpt", + "value": "From <20.6> to <25.5>." + } + ] + }, + { + "video_id": "14VCB", + "question_id": "14VCB", + "video_file": "Charades/videos/14VCB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking something from a cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "EXB02", + "question_id": "EXB02", + "video_file": "Charades/videos/EXB02.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person points at the wall smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <15.6>." + } + ] + }, + { + "video_id": "2NZ5C", + "question_id": "2NZ5C", + "video_file": "Charades/videos/2NZ5C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is undressing under the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.1> to <16.7>." + } + ] + }, + { + "video_id": "2NZ5C", + "question_id": "2NZ5C", + "video_file": "Charades/videos/2NZ5C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begin dressing by a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <16.7>." + } + ] + }, + { + "video_id": "2NZ5C", + "question_id": "2NZ5C", + "video_file": "Charades/videos/2NZ5C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person undresses'." + }, + { + "from": "gpt", + "value": "From <11.1> to <16.7>." + } + ] + }, + { + "video_id": "2NZ5C", + "question_id": "2NZ5C", + "video_file": "Charades/videos/2NZ5C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person dressing in front of the window on the stairs' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.1> to <16.7>." + } + ] + }, + { + "video_id": "2NZ5C", + "question_id": "2NZ5C", + "video_file": "Charades/videos/2NZ5C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person get dressed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <16.7>." + } + ] + }, + { + "video_id": "Z6LYG", + "question_id": "Z6LYG", + "video_file": "Charades/videos/Z6LYG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pause to eat it'." + }, + { + "from": "gpt", + "value": "From <6.7> to <27.1>." + } + ] + }, + { + "video_id": "Z6LYG", + "question_id": "Z6LYG", + "video_file": "Charades/videos/Z6LYG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes some food on a plate out to eat' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.7> to <27.1>." + } + ] + }, + { + "video_id": "Z6LYG", + "question_id": "Z6LYG", + "video_file": "Charades/videos/Z6LYG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <8.2>." + } + ] + }, + { + "video_id": "96GC9", + "question_id": "96GC9", + "video_file": "Charades/videos/96GC9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person the other eats some food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.2> to <19.3>." + } + ] + }, + { + "video_id": "UWL8I", + "question_id": "UWL8I", + "video_file": "Charades/videos/UWL8I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person run back down carrying a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.6> to <24.2>." + } + ] + }, + { + "video_id": "UWL8I", + "question_id": "UWL8I", + "video_file": "Charades/videos/UWL8I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a laughing person runs down the stairs holding a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.5> to <27.3>." + } + ] + }, + { + "video_id": "IKZN3", + "question_id": "IKZN3", + "video_file": "Charades/videos/IKZN3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person stand up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.8> to <34.8>." + } + ] + }, + { + "video_id": "IKZN3", + "question_id": "IKZN3", + "video_file": "Charades/videos/IKZN3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <11.9>." + } + ] + }, + { + "video_id": "ED0VL", + "question_id": "ED0VL", + "video_file": "Charades/videos/ED0VL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person putting trash in a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <23.3>." + } + ] + }, + { + "video_id": "A0CZE", + "question_id": "A0CZE", + "video_file": "Charades/videos/A0CZE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open a refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <18.5>." + } + ] + }, + { + "video_id": "A0CZE", + "question_id": "A0CZE", + "video_file": "Charades/videos/A0CZE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks through a doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <8.1>." + } + ] + }, + { + "video_id": "NAF1C", + "question_id": "NAF1C", + "video_file": "Charades/videos/NAF1C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a closet door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.4> to <36.7>." + } + ] + }, + { + "video_id": "NAF1C", + "question_id": "NAF1C", + "video_file": "Charades/videos/NAF1C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person looks at a picture' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.6> to <31.7>." + } + ] + }, + { + "video_id": "NAF1C", + "question_id": "NAF1C", + "video_file": "Charades/videos/NAF1C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person laughs at the picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.6> to <34.4>." + } + ] + }, + { + "video_id": "GN912", + "question_id": "GN912", + "video_file": "Charades/videos/GN912.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.4> to <21.0>." + } + ] + }, + { + "video_id": "UBVV7", + "question_id": "UBVV7", + "video_file": "Charades/videos/UBVV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes some food out of a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <18.4>." + } + ] + }, + { + "video_id": "UBVV7", + "question_id": "UBVV7", + "video_file": "Charades/videos/UBVV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts box in closet'." + }, + { + "from": "gpt", + "value": "From <19.2> to <27.1>." + } + ] + }, + { + "video_id": "UBVV7", + "question_id": "UBVV7", + "video_file": "Charades/videos/UBVV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats small food item' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.5> to <25.6>." + } + ] + }, + { + "video_id": "UBVV7", + "question_id": "UBVV7", + "video_file": "Charades/videos/UBVV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes the box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.9> to <35.7>." + } + ] + }, + { + "video_id": "UBVV7", + "question_id": "UBVV7", + "video_file": "Charades/videos/UBVV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.5> to <25.6>." + } + ] + }, + { + "video_id": "UBVV7", + "question_id": "UBVV7", + "video_file": "Charades/videos/UBVV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.2> to <27.1>." + } + ] + }, + { + "video_id": "F87KZ", + "question_id": "F87KZ", + "video_file": "Charades/videos/F87KZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening a refrigerator in a kitchen' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <23.7>." + } + ] + }, + { + "video_id": "F87KZ", + "question_id": "F87KZ", + "video_file": "Charades/videos/F87KZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.4> to <23.7>." + } + ] + }, + { + "video_id": "E06A2", + "question_id": "E06A2", + "video_file": "Charades/videos/E06A2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins sneezing'." + }, + { + "from": "gpt", + "value": "From <4.0> to <13.1>." + } + ] + }, + { + "video_id": "MGBLJ", + "question_id": "MGBLJ", + "video_file": "Charades/videos/MGBLJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person standing to look out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.2> to <32.0>." + } + ] + }, + { + "video_id": "MGBLJ", + "question_id": "MGBLJ", + "video_file": "Charades/videos/MGBLJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looks out the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <32.0>." + } + ] + }, + { + "video_id": "FV684", + "question_id": "FV684", + "video_file": "Charades/videos/FV684.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person is smiling in the mirror'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.2>." + } + ] + }, + { + "video_id": "FV684", + "question_id": "FV684", + "video_file": "Charades/videos/FV684.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is looking at themselves in a mirror smiling'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.2>." + } + ] + }, + { + "video_id": "FVPMC", + "question_id": "FVPMC", + "video_file": "Charades/videos/FVPMC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.2> to <32.2>." + } + ] + }, + { + "video_id": "FVPMC", + "question_id": "FVPMC", + "video_file": "Charades/videos/FVPMC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts eating' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.2> to <31.3>." + } + ] + }, + { + "video_id": "FVPMC", + "question_id": "FVPMC", + "video_file": "Charades/videos/FVPMC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is putting a broom inside the pantry door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.6> to <22.3>." + } + ] + }, + { + "video_id": "FVPMC", + "question_id": "FVPMC", + "video_file": "Charades/videos/FVPMC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person holds a broom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.7>." + } + ] + }, + { + "video_id": "FVPMC", + "question_id": "FVPMC", + "video_file": "Charades/videos/FVPMC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the doors of the cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "FVPMC", + "question_id": "FVPMC", + "video_file": "Charades/videos/FVPMC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats the sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.2> to <32.2>." + } + ] + }, + { + "video_id": "FVPMC", + "question_id": "FVPMC", + "video_file": "Charades/videos/FVPMC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the cabinet doors' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <27.9>." + } + ] + }, + { + "video_id": "FJIMX", + "question_id": "FJIMX", + "video_file": "Charades/videos/FJIMX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is smiling'." + }, + { + "from": "gpt", + "value": "From <0.6> to <11.2>." + } + ] + }, + { + "video_id": "FJIMX", + "question_id": "FJIMX", + "video_file": "Charades/videos/FJIMX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person laughing in front of a mirror'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <14.7>." + } + ] + }, + { + "video_id": "FJIMX", + "question_id": "FJIMX", + "video_file": "Charades/videos/FJIMX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting clothes in a box in a bedroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.2> to <31.0>." + } + ] + }, + { + "video_id": "8YKGP", + "question_id": "8YKGP", + "video_file": "Charades/videos/8YKGP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts things away in a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <35.7> to <40.3>." + } + ] + }, + { + "video_id": "Q36FS", + "question_id": "Q36FS", + "video_file": "Charades/videos/Q36FS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person decides to look out of the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.6> to <22.0>." + } + ] + }, + { + "video_id": "Q36FS", + "question_id": "Q36FS", + "video_file": "Charades/videos/Q36FS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person begins laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <14.5>." + } + ] + }, + { + "video_id": "Q36FS", + "question_id": "Q36FS", + "video_file": "Charades/videos/Q36FS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looks out of the window'." + }, + { + "from": "gpt", + "value": "From <11.6> to <22.0>." + } + ] + }, + { + "video_id": "Q36FS", + "question_id": "Q36FS", + "video_file": "Charades/videos/Q36FS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looks out the window'." + }, + { + "from": "gpt", + "value": "From <11.6> to <22.0>." + } + ] + }, + { + "video_id": "85LEB", + "question_id": "85LEB", + "video_file": "Charades/videos/85LEB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <22.0>." + } + ] + }, + { + "video_id": "85LEB", + "question_id": "85LEB", + "video_file": "Charades/videos/85LEB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person working on their laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.8> to <20.0>." + } + ] + }, + { + "video_id": "85LEB", + "question_id": "85LEB", + "video_file": "Charades/videos/85LEB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they drink a bottle of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "R8OTI", + "question_id": "R8OTI", + "video_file": "Charades/videos/R8OTI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks through the doorway of a pantry' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "XO97L", + "question_id": "XO97L", + "video_file": "Charades/videos/XO97L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person sits on a sofa with a blanket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.1> to <23.8>." + } + ] + }, + { + "video_id": "F4BJJ", + "question_id": "F4BJJ", + "video_file": "Charades/videos/F4BJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes a book out of the bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <13.8>." + } + ] + }, + { + "video_id": "F4BJJ", + "question_id": "F4BJJ", + "video_file": "Charades/videos/F4BJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs to a desk' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "F4BJJ", + "question_id": "F4BJJ", + "video_file": "Charades/videos/F4BJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a smiling person runs through the doorway'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "F4BJJ", + "question_id": "F4BJJ", + "video_file": "Charades/videos/F4BJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a book out of their bookbag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.6> to <13.8>." + } + ] + }, + { + "video_id": "F4BJJ", + "question_id": "F4BJJ", + "video_file": "Charades/videos/F4BJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts a bag down on the desk'." + }, + { + "from": "gpt", + "value": "From <7.7> to <12.6>." + } + ] + }, + { + "video_id": "JI3ZF", + "question_id": "JI3ZF", + "video_file": "Charades/videos/JI3ZF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person holds up a towel' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <14.3>." + } + ] + }, + { + "video_id": "T0HLX", + "question_id": "T0HLX", + "video_file": "Charades/videos/T0HLX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put them on a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.7> to <28.5>." + } + ] + }, + { + "video_id": "T0HLX", + "question_id": "T0HLX", + "video_file": "Charades/videos/T0HLX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person puts shoes on a shelf in a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.7> to <28.5>." + } + ] + }, + { + "video_id": "EUMRR", + "question_id": "EUMRR", + "video_file": "Charades/videos/EUMRR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is washing their hands in the bathroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.9>." + } + ] + }, + { + "video_id": "9ILM6", + "question_id": "9ILM6", + "video_file": "Charades/videos/9ILM6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking a cup of coffee in the pantry' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.7> to <30.0>." + } + ] + }, + { + "video_id": "9ILM6", + "question_id": "9ILM6", + "video_file": "Charades/videos/9ILM6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a bag of food'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.2>." + } + ] + }, + { + "video_id": "9ILM6", + "question_id": "9ILM6", + "video_file": "Charades/videos/9ILM6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the cup of coffee on the counter'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.8> to <26.0>." + } + ] + }, + { + "video_id": "WQWT0", + "question_id": "WQWT0", + "video_file": "Charades/videos/WQWT0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sitting at a desk eating' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <25.0>." + } + ] + }, + { + "video_id": "WQWT0", + "question_id": "WQWT0", + "video_file": "Charades/videos/WQWT0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the sandwich on the plate' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <16.9>." + } + ] + }, + { + "video_id": "WQWT0", + "question_id": "WQWT0", + "video_file": "Charades/videos/WQWT0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "JD3KW", + "question_id": "JD3KW", + "video_file": "Charades/videos/JD3KW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is in the basement tidying up a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.3> to <17.1>." + } + ] + }, + { + "video_id": "JD3KW", + "question_id": "JD3KW", + "video_file": "Charades/videos/JD3KW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <34.7> to <43.1>." + } + ] + }, + { + "video_id": "4X5P6", + "question_id": "4X5P6", + "video_file": "Charades/videos/4X5P6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing in a doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <10.5>." + } + ] + }, + { + "video_id": "LI6LV", + "question_id": "LI6LV", + "video_file": "Charades/videos/LI6LV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person started undressing'." + }, + { + "from": "gpt", + "value": "From <2.7> to <15.2>." + } + ] + }, + { + "video_id": "LI6LV", + "question_id": "LI6LV", + "video_file": "Charades/videos/LI6LV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person put a towel on the doorknob'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "X0FW6", + "question_id": "X0FW6", + "video_file": "Charades/videos/X0FW6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is running through a living room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "X0FW6", + "question_id": "X0FW6", + "video_file": "Charades/videos/X0FW6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person sits on a couch' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.2> to <5.8>." + } + ] + }, + { + "video_id": "X0FW6", + "question_id": "X0FW6", + "video_file": "Charades/videos/X0FW6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs into the kitchen' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "0EJAG", + "question_id": "0EJAG", + "video_file": "Charades/videos/0EJAG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a laughing person has a blanket wrapped around their shoulders' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.1> to <11.5>." + } + ] + }, + { + "video_id": "HCM54", + "question_id": "HCM54", + "video_file": "Charades/videos/HCM54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is standing there smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <10.5>." + } + ] + }, + { + "video_id": "HCM54", + "question_id": "HCM54", + "video_file": "Charades/videos/HCM54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person smiling at a picture on the wall'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <10.5>." + } + ] + }, + { + "video_id": "F1OMV", + "question_id": "F1OMV", + "video_file": "Charades/videos/F1OMV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person close the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.7> to <17.2>." + } + ] + }, + { + "video_id": "F1OMV", + "question_id": "F1OMV", + "video_file": "Charades/videos/F1OMV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turning off a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.2> to <15.2>." + } + ] + }, + { + "video_id": "VUXLB", + "question_id": "VUXLB", + "video_file": "Charades/videos/VUXLB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the glass back' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.0> to <27.0>." + } + ] + }, + { + "video_id": "VUXLB", + "question_id": "VUXLB", + "video_file": "Charades/videos/VUXLB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person putting the coffee down on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.8> to <29.4>." + } + ] + }, + { + "video_id": "3K2QE", + "question_id": "3K2QE", + "video_file": "Charades/videos/3K2QE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.9> to <19.9>." + } + ] + }, + { + "video_id": "1KJI0", + "question_id": "1KJI0", + "video_file": "Charades/videos/1KJI0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is cooking food on a stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.8>." + } + ] + }, + { + "video_id": "1KJI0", + "question_id": "1KJI0", + "video_file": "Charades/videos/1KJI0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a refrigerator to pour milk' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.6> to <13.2>." + } + ] + }, + { + "video_id": "1KJI0", + "question_id": "1KJI0", + "video_file": "Charades/videos/1KJI0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking something on the stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.8>." + } + ] + }, + { + "video_id": "YIK5I", + "question_id": "YIK5I", + "video_file": "Charades/videos/YIK5I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <28.5>." + } + ] + }, + { + "video_id": "YIK5I", + "question_id": "YIK5I", + "video_file": "Charades/videos/YIK5I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting at their desk watching television'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.5>." + } + ] + }, + { + "video_id": "YIK5I", + "question_id": "YIK5I", + "video_file": "Charades/videos/YIK5I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening his front door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.1> to <28.5>." + } + ] + }, + { + "video_id": "YIK5I", + "question_id": "YIK5I", + "video_file": "Charades/videos/YIK5I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opening the door to air it out' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.1> to <28.5>." + } + ] + }, + { + "video_id": "5X6S7", + "question_id": "5X6S7", + "video_file": "Charades/videos/5X6S7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sitting down laughing at their desk'." + }, + { + "from": "gpt", + "value": "From <1.6> to <7.9>." + } + ] + }, + { + "video_id": "5X6S7", + "question_id": "5X6S7", + "video_file": "Charades/videos/5X6S7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting in a chair smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.3>." + } + ] + }, + { + "video_id": "5O0YS", + "question_id": "5O0YS", + "video_file": "Charades/videos/5O0YS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs out of the room through an entryway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.8> to <22.0>." + } + ] + }, + { + "video_id": "5O0YS", + "question_id": "5O0YS", + "video_file": "Charades/videos/5O0YS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs out room with pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.8> to <22.0>." + } + ] + }, + { + "video_id": "5O0YS", + "question_id": "5O0YS", + "video_file": "Charades/videos/5O0YS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they start running to the door'." + }, + { + "from": "gpt", + "value": "From <15.8> to <22.0>." + } + ] + }, + { + "video_id": "WT405", + "question_id": "WT405", + "video_file": "Charades/videos/WT405.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person began watching the laptop'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "WT405", + "question_id": "WT405", + "video_file": "Charades/videos/WT405.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begin watching a show on a laptop'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "WT405", + "question_id": "WT405", + "video_file": "Charades/videos/WT405.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put the glass in the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <22.6>." + } + ] + }, + { + "video_id": "WT405", + "question_id": "WT405", + "video_file": "Charades/videos/WT405.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person pours a glass of water that they' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.1> to <16.7>." + } + ] + }, + { + "video_id": "ZFQNZ", + "question_id": "ZFQNZ", + "video_file": "Charades/videos/ZFQNZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person cooking in their kitchen opens their refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <23.1>." + } + ] + }, + { + "video_id": "ZFQNZ", + "question_id": "ZFQNZ", + "video_file": "Charades/videos/ZFQNZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.8>." + } + ] + }, + { + "video_id": "ZFQNZ", + "question_id": "ZFQNZ", + "video_file": "Charades/videos/ZFQNZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person start cooking on the stove'." + }, + { + "from": "gpt", + "value": "From <14.0> to <23.1>." + } + ] + }, + { + "video_id": "ZFQNZ", + "question_id": "ZFQNZ", + "video_file": "Charades/videos/ZFQNZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begin cooking some food at a stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <23.1>." + } + ] + }, + { + "video_id": "ZFQNZ", + "question_id": "ZFQNZ", + "video_file": "Charades/videos/ZFQNZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take out some food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <15.3>." + } + ] + }, + { + "video_id": "C7P86", + "question_id": "C7P86", + "video_file": "Charades/videos/C7P86.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person undresses'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.1>." + } + ] + }, + { + "video_id": "99596", + "question_id": "99596", + "video_file": "Charades/videos/99596.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <35.1> to <42.4>." + } + ] + }, + { + "video_id": "99596", + "question_id": "99596", + "video_file": "Charades/videos/99596.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begin dressing'." + }, + { + "from": "gpt", + "value": "From <28.2> to <39.5>." + } + ] + }, + { + "video_id": "99596", + "question_id": "99596", + "video_file": "Charades/videos/99596.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is awakening under a blanket'." + }, + { + "from": "gpt", + "value": "From <14.6> to <30.4>." + } + ] + }, + { + "video_id": "GBDJX", + "question_id": "GBDJX", + "video_file": "Charades/videos/GBDJX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats some'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.5> to <24.0>." + } + ] + }, + { + "video_id": "GBDJX", + "question_id": "GBDJX", + "video_file": "Charades/videos/GBDJX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <14.8>." + } + ] + }, + { + "video_id": "GBDJX", + "question_id": "GBDJX", + "video_file": "Charades/videos/GBDJX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person eats the food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.5> to <24.0>." + } + ] + }, + { + "video_id": "GBDJX", + "question_id": "GBDJX", + "video_file": "Charades/videos/GBDJX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a closet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "GPR89", + "question_id": "GPR89", + "video_file": "Charades/videos/GPR89.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person takes a stack of papers off a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <44.8>." + } + ] + }, + { + "video_id": "XSY2P", + "question_id": "XSY2P", + "video_file": "Charades/videos/XSY2P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts his laptop in a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.4> to <26.5>." + } + ] + }, + { + "video_id": "UFNPU", + "question_id": "UFNPU", + "video_file": "Charades/videos/UFNPU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens the door to the laundry room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <7.3>." + } + ] + }, + { + "video_id": "UFNPU", + "question_id": "UFNPU", + "video_file": "Charades/videos/UFNPU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the door behind them' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <11.8>." + } + ] + }, + { + "video_id": "2XG25", + "question_id": "2XG25", + "video_file": "Charades/videos/2XG25.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person awakens in their bedroom.they look in a mirror'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "Y0CA0", + "question_id": "Y0CA0", + "video_file": "Charades/videos/Y0CA0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person appears to be cooking something' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.6> to <22.3>." + } + ] + }, + { + "video_id": "Y0CA0", + "question_id": "Y0CA0", + "video_file": "Charades/videos/Y0CA0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is cooking on a stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.6> to <22.3>." + } + ] + }, + { + "video_id": "Y0CA0", + "question_id": "Y0CA0", + "video_file": "Charades/videos/Y0CA0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closing cabinets' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <12.4>." + } + ] + }, + { + "video_id": "Y0CA0", + "question_id": "Y0CA0", + "video_file": "Charades/videos/Y0CA0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is closing the cabinet'." + }, + { + "from": "gpt", + "value": "From <6.5> to <12.4>." + } + ] + }, + { + "video_id": "KRLT7", + "question_id": "KRLT7", + "video_file": "Charades/videos/KRLT7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take a picture of the clothes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <22.8>." + } + ] + }, + { + "video_id": "KRLT7", + "question_id": "KRLT7", + "video_file": "Charades/videos/KRLT7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is throwing clothes on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.7> to <12.5>." + } + ] + }, + { + "video_id": "FEKJQ", + "question_id": "FEKJQ", + "video_file": "Charades/videos/FEKJQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <14.8>." + } + ] + }, + { + "video_id": "EHCUE", + "question_id": "EHCUE", + "video_file": "Charades/videos/EHCUE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting in a chair at a desk'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.5> to <26.6>." + } + ] + }, + { + "video_id": "02CYP", + "question_id": "02CYP", + "video_file": "Charades/videos/02CYP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting on a couch watching television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.7>." + } + ] + }, + { + "video_id": "CICFF", + "question_id": "CICFF", + "video_file": "Charades/videos/CICFF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person someone walks through a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "CICFF", + "question_id": "CICFF", + "video_file": "Charades/videos/CICFF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take a picture with their phone'." + }, + { + "from": "gpt", + "value": "From <5.6> to <15.8>." + } + ] + }, + { + "video_id": "CICFF", + "question_id": "CICFF", + "video_file": "Charades/videos/CICFF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a phone from the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.6> to <15.8>." + } + ] + }, + { + "video_id": "WIPMO", + "question_id": "WIPMO", + "video_file": "Charades/videos/WIPMO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person closes the book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <38.4> to <43.0>." + } + ] + }, + { + "video_id": "Y49WR", + "question_id": "Y49WR", + "video_file": "Charades/videos/Y49WR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.4> to <23.0>." + } + ] + }, + { + "video_id": "Y49WR", + "question_id": "Y49WR", + "video_file": "Charades/videos/Y49WR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person getting dressed takes a picture in mirror'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.4> to <23.0>." + } + ] + }, + { + "video_id": "MJZJM", + "question_id": "MJZJM", + "video_file": "Charades/videos/MJZJM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person turns on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <10.5>." + } + ] + }, + { + "video_id": "17RPG", + "question_id": "17RPG", + "video_file": "Charades/videos/17RPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is standing in a hallway smiling'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "17RPG", + "question_id": "17RPG", + "video_file": "Charades/videos/17RPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating the sandwich they are holding' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.7> to <33.0>." + } + ] + }, + { + "video_id": "17RPG", + "question_id": "17RPG", + "video_file": "Charades/videos/17RPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person watch themselves eat'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.7> to <33.0>." + } + ] + }, + { + "video_id": "17RPG", + "question_id": "17RPG", + "video_file": "Charades/videos/17RPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take off their sweater' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.7> to <33.0>." + } + ] + }, + { + "video_id": "17RPG", + "question_id": "17RPG", + "video_file": "Charades/videos/17RPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <27.7> to <33.0>." + } + ] + }, + { + "video_id": "CZPFO", + "question_id": "CZPFO", + "video_file": "Charades/videos/CZPFO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.1> to <36.6>." + } + ] + }, + { + "video_id": "CZPFO", + "question_id": "CZPFO", + "video_file": "Charades/videos/CZPFO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens up a closet door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.1> to <36.6>." + } + ] + }, + { + "video_id": "YVOM4", + "question_id": "YVOM4", + "video_file": "Charades/videos/YVOM4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs back out' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.1> to <18.3>." + } + ] + }, + { + "video_id": "YVOM4", + "question_id": "YVOM4", + "video_file": "Charades/videos/YVOM4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person run in' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.1> to <18.3>." + } + ] + }, + { + "video_id": "DLQIT", + "question_id": "DLQIT", + "video_file": "Charades/videos/DLQIT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person briefly looked at a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.6> to <35.5>." + } + ] + }, + { + "video_id": "DLQIT", + "question_id": "DLQIT", + "video_file": "Charades/videos/DLQIT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <12.9>." + } + ] + }, + { + "video_id": "DLQIT", + "question_id": "DLQIT", + "video_file": "Charades/videos/DLQIT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person poured water into a glass took a drink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.4> to <18.3>." + } + ] + }, + { + "video_id": "DLQIT", + "question_id": "DLQIT", + "video_file": "Charades/videos/DLQIT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they start pouring themselves a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.4> to <18.3>." + } + ] + }, + { + "video_id": "MNJWH", + "question_id": "MNJWH", + "video_file": "Charades/videos/MNJWH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a book'." + }, + { + "from": "gpt", + "value": "From <1.8> to <12.1>." + } + ] + }, + { + "video_id": "MNJWH", + "question_id": "MNJWH", + "video_file": "Charades/videos/MNJWH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holds a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.8>." + } + ] + }, + { + "video_id": "SPDF6", + "question_id": "SPDF6", + "video_file": "Charades/videos/SPDF6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.9> to <25.9>." + } + ] + }, + { + "video_id": "SPDF6", + "question_id": "SPDF6", + "video_file": "Charades/videos/SPDF6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is throwing clothes into a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <13.9>." + } + ] + }, + { + "video_id": "TGUKM", + "question_id": "TGUKM", + "video_file": "Charades/videos/TGUKM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'pick it up .the person takes a flash light out' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.0> to <16.5>." + } + ] + }, + { + "video_id": "TGUKM", + "question_id": "TGUKM", + "video_file": "Charades/videos/TGUKM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person examining it with a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <16.5>." + } + ] + }, + { + "video_id": "52WDV", + "question_id": "52WDV", + "video_file": "Charades/videos/52WDV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is dressing in a walk-in closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.7> to <14.9>." + } + ] + }, + { + "video_id": "52WDV", + "question_id": "52WDV", + "video_file": "Charades/videos/52WDV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person watching their reflection in a mirror' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.5> to <21.0>." + } + ] + }, + { + "video_id": "5DYQR", + "question_id": "5DYQR", + "video_file": "Charades/videos/5DYQR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.3> to <9.8>." + } + ] + }, + { + "video_id": "W6TFQ", + "question_id": "W6TFQ", + "video_file": "Charades/videos/W6TFQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person the they sneeze as they open the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.6> to <33.0>." + } + ] + }, + { + "video_id": "W6TFQ", + "question_id": "W6TFQ", + "video_file": "Charades/videos/W6TFQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.6> to <33.0>." + } + ] + }, + { + "video_id": "Y2TNP", + "question_id": "Y2TNP", + "video_file": "Charades/videos/Y2TNP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <6.1>." + } + ] + }, + { + "video_id": "Y2TNP", + "question_id": "Y2TNP", + "video_file": "Charades/videos/Y2TNP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns on a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <6.1>." + } + ] + }, + { + "video_id": "Y2TNP", + "question_id": "Y2TNP", + "video_file": "Charades/videos/Y2TNP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person leaving a light on in the entryway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <6.1>." + } + ] + }, + { + "video_id": "ZATTZ", + "question_id": "ZATTZ", + "video_file": "Charades/videos/ZATTZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is holding some dishes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "ZATTZ", + "question_id": "ZATTZ", + "video_file": "Charades/videos/ZATTZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins cooking at the stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.2> to <34.0>." + } + ] + }, + { + "video_id": "VXF18", + "question_id": "VXF18", + "video_file": "Charades/videos/VXF18.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person a glass sits down on a couch'." + }, + { + "from": "gpt", + "value": "From <2.2> to <7.4>." + } + ] + }, + { + "video_id": "VXF18", + "question_id": "VXF18", + "video_file": "Charades/videos/VXF18.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts drinking from a cup'." + }, + { + "from": "gpt", + "value": "From <35.3> to <43.7>." + } + ] + }, + { + "video_id": "VXF18", + "question_id": "VXF18", + "video_file": "Charades/videos/VXF18.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person sits down on the couch' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <7.4>." + } + ] + }, + { + "video_id": "VXF18", + "question_id": "VXF18", + "video_file": "Charades/videos/VXF18.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person eats some of the food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.7> to <15.0>." + } + ] + }, + { + "video_id": "VXF18", + "question_id": "VXF18", + "video_file": "Charades/videos/VXF18.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person is holding a drink cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <35.3> to <43.7>." + } + ] + }, + { + "video_id": "VXF18", + "question_id": "VXF18", + "video_file": "Charades/videos/VXF18.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is eating food from a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.7> to <15.0>." + } + ] + }, + { + "video_id": "IHGNV", + "question_id": "IHGNV", + "video_file": "Charades/videos/IHGNV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens the bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <18.2>." + } + ] + }, + { + "video_id": "IHGNV", + "question_id": "IHGNV", + "video_file": "Charades/videos/IHGNV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is holding a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "3MLVF", + "question_id": "3MLVF", + "video_file": "Charades/videos/3MLVF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking food on the stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.6>." + } + ] + }, + { + "video_id": "3MLVF", + "question_id": "3MLVF", + "video_file": "Charades/videos/3MLVF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinks from the coke bottle' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.7> to <32.0>." + } + ] + }, + { + "video_id": "3MLVF", + "question_id": "3MLVF", + "video_file": "Charades/videos/3MLVF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is cooking with a pot on the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.6>." + } + ] + }, + { + "video_id": "3MLVF", + "question_id": "3MLVF", + "video_file": "Charades/videos/3MLVF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they open the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <16.6>." + } + ] + }, + { + "video_id": "3MLVF", + "question_id": "3MLVF", + "video_file": "Charades/videos/3MLVF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person the open the refrigerator remove a soda'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <16.6>." + } + ] + }, + { + "video_id": "3MLVF", + "question_id": "3MLVF", + "video_file": "Charades/videos/3MLVF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is cooking on a stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.6>." + } + ] + }, + { + "video_id": "UNGPY", + "question_id": "UNGPY", + "video_file": "Charades/videos/UNGPY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <20.2>." + } + ] + }, + { + "video_id": "5ZRFL", + "question_id": "5ZRFL", + "video_file": "Charades/videos/5ZRFL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens their bedroom window'." + }, + { + "from": "gpt", + "value": "From <6.7> to <12.6>." + } + ] + }, + { + "video_id": "5ZRFL", + "question_id": "5ZRFL", + "video_file": "Charades/videos/5ZRFL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a camera from their nightstand' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.6> to <32.0>." + } + ] + }, + { + "video_id": "SUYMQ", + "question_id": "SUYMQ", + "video_file": "Charades/videos/SUYMQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a pair of shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <17.3>." + } + ] + }, + { + "video_id": "SUYMQ", + "question_id": "SUYMQ", + "video_file": "Charades/videos/SUYMQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding shoes in their hands'." + }, + { + "from": "gpt", + "value": "From <0.0> to <17.3>." + } + ] + }, + { + "video_id": "SUYMQ", + "question_id": "SUYMQ", + "video_file": "Charades/videos/SUYMQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person stand up'." + }, + { + "from": "gpt", + "value": "From <9.7> to <17.1>." + } + ] + }, + { + "video_id": "SUYMQ", + "question_id": "SUYMQ", + "video_file": "Charades/videos/SUYMQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they begin to stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <17.1>." + } + ] + }, + { + "video_id": "XIYIK", + "question_id": "XIYIK", + "video_file": "Charades/videos/XIYIK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.3> to <20.5>." + } + ] + }, + { + "video_id": "XIYIK", + "question_id": "XIYIK", + "video_file": "Charades/videos/XIYIK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a towel from a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <23.8>." + } + ] + }, + { + "video_id": "E1TUD", + "question_id": "E1TUD", + "video_file": "Charades/videos/E1TUD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is dressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <21.2>." + } + ] + }, + { + "video_id": "E1TUD", + "question_id": "E1TUD", + "video_file": "Charades/videos/E1TUD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begins dressing again' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <21.2>." + } + ] + }, + { + "video_id": "E1TUD", + "question_id": "E1TUD", + "video_file": "Charades/videos/E1TUD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is undressing in the bathroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "BVX8X", + "question_id": "BVX8X", + "video_file": "Charades/videos/BVX8X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person goes to open the refrigerator for something' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.3> to <26.7>." + } + ] + }, + { + "video_id": "5OMSL", + "question_id": "5OMSL", + "video_file": "Charades/videos/5OMSL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take a box'." + }, + { + "from": "gpt", + "value": "From <26.1> to <31.9>." + } + ] + }, + { + "video_id": "5OMSL", + "question_id": "5OMSL", + "video_file": "Charades/videos/5OMSL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person take a box off the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.1> to <31.9>." + } + ] + }, + { + "video_id": "U6QKX", + "question_id": "U6QKX", + "video_file": "Charades/videos/U6QKX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting it on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.2> to <14.6>." + } + ] + }, + { + "video_id": "1N1UT", + "question_id": "1N1UT", + "video_file": "Charades/videos/1N1UT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich closes a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.3> to <30.7>." + } + ] + }, + { + "video_id": "MS3E8", + "question_id": "MS3E8", + "video_file": "Charades/videos/MS3E8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts putting groceries away from a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "MS3E8", + "question_id": "MS3E8", + "video_file": "Charades/videos/MS3E8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person carrying a holding a box'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "9JIOI", + "question_id": "9JIOI", + "video_file": "Charades/videos/9JIOI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes it to the stove to cook' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.1> to <14.1>." + } + ] + }, + { + "video_id": "9JIOI", + "question_id": "9JIOI", + "video_file": "Charades/videos/9JIOI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the refrigerator door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <17.0>." + } + ] + }, + { + "video_id": "9JIOI", + "question_id": "9JIOI", + "video_file": "Charades/videos/9JIOI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks to the refrigerator to close the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.2> to <17.0>." + } + ] + }, + { + "video_id": "9JIOI", + "question_id": "9JIOI", + "video_file": "Charades/videos/9JIOI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person is cooking on the stove near the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.1> to <14.1>." + } + ] + }, + { + "video_id": "YPZQ2", + "question_id": "YPZQ2", + "video_file": "Charades/videos/YPZQ2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.3> to <15.5>." + } + ] + }, + { + "video_id": "1WAOU", + "question_id": "1WAOU", + "video_file": "Charades/videos/1WAOU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is walking around drinking out of a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "T03KF", + "question_id": "T03KF", + "video_file": "Charades/videos/T03KF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begin eating it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.6> to <32.0>." + } + ] + }, + { + "video_id": "T03KF", + "question_id": "T03KF", + "video_file": "Charades/videos/T03KF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating some food'." + }, + { + "from": "gpt", + "value": "From <21.6> to <32.0>." + } + ] + }, + { + "video_id": "T03KF", + "question_id": "T03KF", + "video_file": "Charades/videos/T03KF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes clothes out of a cabinet'." + }, + { + "from": "gpt", + "value": "From <10.6> to <14.9>." + } + ] + }, + { + "video_id": "T03KF", + "question_id": "T03KF", + "video_file": "Charades/videos/T03KF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throwing clothes on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.6> to <19.6>." + } + ] + }, + { + "video_id": "T03KF", + "question_id": "T03KF", + "video_file": "Charades/videos/T03KF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws them on the floor'." + }, + { + "from": "gpt", + "value": "From <10.6> to <19.6>." + } + ] + }, + { + "video_id": "T03KF", + "question_id": "T03KF", + "video_file": "Charades/videos/T03KF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <32.0>." + } + ] + }, + { + "video_id": "UDRZ6", + "question_id": "UDRZ6", + "video_file": "Charades/videos/UDRZ6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is laughing on the stairs'." + }, + { + "from": "gpt", + "value": "From <5.0> to <15.0>." + } + ] + }, + { + "video_id": "R0OLR", + "question_id": "R0OLR", + "video_file": "Charades/videos/R0OLR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes towel off'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.3>." + } + ] + }, + { + "video_id": "R0OLR", + "question_id": "R0OLR", + "video_file": "Charades/videos/R0OLR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws the blanket on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.7> to <24.9>." + } + ] + }, + { + "video_id": "R0OLR", + "question_id": "R0OLR", + "video_file": "Charades/videos/R0OLR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person stops & starts sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.4> to <19.4>." + } + ] + }, + { + "video_id": "R0OLR", + "question_id": "R0OLR", + "video_file": "Charades/videos/R0OLR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is running down the hallway holding a blanket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.7> to <26.9>." + } + ] + }, + { + "video_id": "R0OLR", + "question_id": "R0OLR", + "video_file": "Charades/videos/R0OLR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person wrapped in a blanket runs into the room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.7> to <26.9>." + } + ] + }, + { + "video_id": "W8H3C", + "question_id": "W8H3C", + "video_file": "Charades/videos/W8H3C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is holding a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.1>." + } + ] + }, + { + "video_id": "CN43X", + "question_id": "CN43X", + "video_file": "Charades/videos/CN43X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.0> to <10.4>." + } + ] + }, + { + "video_id": "O8N4M", + "question_id": "O8N4M", + "video_file": "Charades/videos/O8N4M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakens on the couch in the living room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.6> to <17.8>." + } + ] + }, + { + "video_id": "V70F7", + "question_id": "V70F7", + "video_file": "Charades/videos/V70F7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is fixing his hair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.8> to <12.1>." + } + ] + }, + { + "video_id": "V70F7", + "question_id": "V70F7", + "video_file": "Charades/videos/V70F7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <25.6>." + } + ] + }, + { + "video_id": "47Y1X", + "question_id": "47Y1X", + "video_file": "Charades/videos/47Y1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person eating food in pantry smiling at the camera' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "9NWSD", + "question_id": "9NWSD", + "video_file": "Charades/videos/9NWSD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the box on top of the dresser' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.0> to <29.0>." + } + ] + }, + { + "video_id": "SDXZ8", + "question_id": "SDXZ8", + "video_file": "Charades/videos/SDXZ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they take a vacuum' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <33.5> to <38.8>." + } + ] + }, + { + "video_id": "SDXZ8", + "question_id": "SDXZ8", + "video_file": "Charades/videos/SDXZ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put the dish away' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.4> to <27.2>." + } + ] + }, + { + "video_id": "SDXZ8", + "question_id": "SDXZ8", + "video_file": "Charades/videos/SDXZ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person washes the dishes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <24.7>." + } + ] + }, + { + "video_id": "SDXZ8", + "question_id": "SDXZ8", + "video_file": "Charades/videos/SDXZ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person washes some dishes in a sink' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.3> to <24.7>." + } + ] + }, + { + "video_id": "4ZN66", + "question_id": "4ZN66", + "video_file": "Charades/videos/4ZN66.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes out their phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.5> to <33.2>." + } + ] + }, + { + "video_id": "4ZN66", + "question_id": "4ZN66", + "video_file": "Charades/videos/4ZN66.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a phone out their pocket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.5> to <33.2>." + } + ] + }, + { + "video_id": "V149B", + "question_id": "V149B", + "video_file": "Charades/videos/V149B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person washing windows with a cloth' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.3> to <32.0>." + } + ] + }, + { + "video_id": "V149B", + "question_id": "V149B", + "video_file": "Charades/videos/V149B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person washing a window in an entryway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.3> to <32.0>." + } + ] + }, + { + "video_id": "UL7YC", + "question_id": "UL7YC", + "video_file": "Charades/videos/UL7YC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a glass of soda'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.9> to <17.4>." + } + ] + }, + { + "video_id": "5TG2E", + "question_id": "5TG2E", + "video_file": "Charades/videos/5TG2E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person after putting on shoes from a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <9.7>." + } + ] + }, + { + "video_id": "5TG2E", + "question_id": "5TG2E", + "video_file": "Charades/videos/5TG2E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'that person puts on their tennis shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <9.7>." + } + ] + }, + { + "video_id": "AYMCZ", + "question_id": "AYMCZ", + "video_file": "Charades/videos/AYMCZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is putting a pillow into a washer'." + }, + { + "from": "gpt", + "value": "From <4.4> to <13.5>." + } + ] + }, + { + "video_id": "AYMCZ", + "question_id": "AYMCZ", + "video_file": "Charades/videos/AYMCZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person working on the laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.8> to <30.0>." + } + ] + }, + { + "video_id": "AYMCZ", + "question_id": "AYMCZ", + "video_file": "Charades/videos/AYMCZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opening a laptop'." + }, + { + "from": "gpt", + "value": "From <18.7> to <25.2>." + } + ] + }, + { + "video_id": "AYMCZ", + "question_id": "AYMCZ", + "video_file": "Charades/videos/AYMCZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person finally opening a laptop in a laundry room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <25.2>." + } + ] + }, + { + "video_id": "FN0F8", + "question_id": "FN0F8", + "video_file": "Charades/videos/FN0F8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting the piece of paper away' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.3> to <36.8>." + } + ] + }, + { + "video_id": "SBW7O", + "question_id": "SBW7O", + "video_file": "Charades/videos/SBW7O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person looks out a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.5>." + } + ] + }, + { + "video_id": "SBW7O", + "question_id": "SBW7O", + "video_file": "Charades/videos/SBW7O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the door'." + }, + { + "from": "gpt", + "value": "From <27.1> to <32.0>." + } + ] + }, + { + "video_id": "SBW7O", + "question_id": "SBW7O", + "video_file": "Charades/videos/SBW7O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.1> to <32.0>." + } + ] + }, + { + "video_id": "KZ746", + "question_id": "KZ746", + "video_file": "Charades/videos/KZ746.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <8.2>." + } + ] + }, + { + "video_id": "XF41L", + "question_id": "XF41L", + "video_file": "Charades/videos/XF41L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opening a closet door'." + }, + { + "from": "gpt", + "value": "From <2.2> to <9.3>." + } + ] + }, + { + "video_id": "XF41L", + "question_id": "XF41L", + "video_file": "Charades/videos/XF41L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opened a closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.2> to <9.3>." + } + ] + }, + { + "video_id": "XF41L", + "question_id": "XF41L", + "video_file": "Charades/videos/XF41L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <23.6>." + } + ] + }, + { + "video_id": "XF41L", + "question_id": "XF41L", + "video_file": "Charades/videos/XF41L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.4> to <30.8>." + } + ] + }, + { + "video_id": "XF41L", + "question_id": "XF41L", + "video_file": "Charades/videos/XF41L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes off shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <23.6>." + } + ] + }, + { + "video_id": "XF41L", + "question_id": "XF41L", + "video_file": "Charades/videos/XF41L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the phone on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.0> to <20.1>." + } + ] + }, + { + "video_id": "NHD1J", + "question_id": "NHD1J", + "video_file": "Charades/videos/NHD1J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <21.4>." + } + ] + }, + { + "video_id": "NHD1J", + "question_id": "NHD1J", + "video_file": "Charades/videos/NHD1J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.7> to <21.4>." + } + ] + }, + { + "video_id": "NHD1J", + "question_id": "NHD1J", + "video_file": "Charades/videos/NHD1J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is undressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.2> to <20.7>." + } + ] + }, + { + "video_id": "NHD1J", + "question_id": "NHD1J", + "video_file": "Charades/videos/NHD1J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens up laptop'." + }, + { + "from": "gpt", + "value": "From <14.7> to <21.4>." + } + ] + }, + { + "video_id": "NHD1J", + "question_id": "NHD1J", + "video_file": "Charades/videos/NHD1J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person watching television' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <14.2>." + } + ] + }, + { + "video_id": "Z5FNI", + "question_id": "Z5FNI", + "video_file": "Charades/videos/Z5FNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.3> to <26.1>." + } + ] + }, + { + "video_id": "Z5FNI", + "question_id": "Z5FNI", + "video_file": "Charades/videos/Z5FNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turning off the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.5> to <19.0>." + } + ] + }, + { + "video_id": "Z5FNI", + "question_id": "Z5FNI", + "video_file": "Charades/videos/Z5FNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.1> to <22.7>." + } + ] + }, + { + "video_id": "OB1XO", + "question_id": "OB1XO", + "video_file": "Charades/videos/OB1XO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts on their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.7> to <31.4>." + } + ] + }, + { + "video_id": "99WON", + "question_id": "99WON", + "video_file": "Charades/videos/99WON.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits in a chair at the top' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.3> to <11.8>." + } + ] + }, + { + "video_id": "99WON", + "question_id": "99WON", + "video_file": "Charades/videos/99WON.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs up the stairs'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "99WON", + "question_id": "99WON", + "video_file": "Charades/videos/99WON.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs up a flight of stairs' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "WVB7F", + "question_id": "WVB7F", + "video_file": "Charades/videos/WVB7F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes some medicine' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.8> to <33.3>." + } + ] + }, + { + "video_id": "WVB7F", + "question_id": "WVB7F", + "video_file": "Charades/videos/WVB7F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a blanket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.5> to <18.1>." + } + ] + }, + { + "video_id": "WVB7F", + "question_id": "WVB7F", + "video_file": "Charades/videos/WVB7F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a bottle of medicine'." + }, + { + "from": "gpt", + "value": "From <27.8> to <33.3>." + } + ] + }, + { + "video_id": "WVB7F", + "question_id": "WVB7F", + "video_file": "Charades/videos/WVB7F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "0TIHC", + "question_id": "0TIHC", + "video_file": "Charades/videos/0TIHC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws their clothes onto the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "0TIHC", + "question_id": "0TIHC", + "video_file": "Charades/videos/0TIHC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws clothes on the floor'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "0TIHC", + "question_id": "0TIHC", + "video_file": "Charades/videos/0TIHC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person undresses' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.3>." + } + ] + }, + { + "video_id": "0TIHC", + "question_id": "0TIHC", + "video_file": "Charades/videos/0TIHC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes a water bottle' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.0> to <27.7>." + } + ] + }, + { + "video_id": "NJ0X1", + "question_id": "NJ0X1", + "video_file": "Charades/videos/NJ0X1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.5> to <24.0>." + } + ] + }, + { + "video_id": "G537E", + "question_id": "G537E", + "video_file": "Charades/videos/G537E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sneezing on a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.3> to <22.6>." + } + ] + }, + { + "video_id": "G537E", + "question_id": "G537E", + "video_file": "Charades/videos/G537E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person finally putting away groceries in a pantry'." + }, + { + "from": "gpt", + "value": "From <23.9> to <40.0>." + } + ] + }, + { + "video_id": "BPH56", + "question_id": "BPH56", + "video_file": "Charades/videos/BPH56.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'laughing as the person collects the dishes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.4>." + } + ] + }, + { + "video_id": "BPH56", + "question_id": "BPH56", + "video_file": "Charades/videos/BPH56.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns the light on' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.1> to <29.1>." + } + ] + }, + { + "video_id": "BPH56", + "question_id": "BPH56", + "video_file": "Charades/videos/BPH56.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turn on the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <29.1>." + } + ] + }, + { + "video_id": "YU9BS", + "question_id": "YU9BS", + "video_file": "Charades/videos/YU9BS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns the light off'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <34.9> to <39.7>." + } + ] + }, + { + "video_id": "YU9BS", + "question_id": "YU9BS", + "video_file": "Charades/videos/YU9BS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns off the light'." + }, + { + "from": "gpt", + "value": "From <34.9> to <39.7>." + } + ] + }, + { + "video_id": "YU9BS", + "question_id": "YU9BS", + "video_file": "Charades/videos/YU9BS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person looks out the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.3> to <30.6>." + } + ] + }, + { + "video_id": "YU9BS", + "question_id": "YU9BS", + "video_file": "Charades/videos/YU9BS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person looks out the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.3> to <30.6>." + } + ] + }, + { + "video_id": "VOU8V", + "question_id": "VOU8V", + "video_file": "Charades/videos/VOU8V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.9>." + } + ] + }, + { + "video_id": "WXUZK", + "question_id": "WXUZK", + "video_file": "Charades/videos/WXUZK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person a walks into the bedroom undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.7>." + } + ] + }, + { + "video_id": "75MO8", + "question_id": "75MO8", + "video_file": "Charades/videos/75MO8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <25.7>." + } + ] + }, + { + "video_id": "75MO8", + "question_id": "75MO8", + "video_file": "Charades/videos/75MO8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs into the recreation room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.9>." + } + ] + }, + { + "video_id": "75MO8", + "question_id": "75MO8", + "video_file": "Charades/videos/75MO8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs over to a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.9>." + } + ] + }, + { + "video_id": "KAFV7", + "question_id": "KAFV7", + "video_file": "Charades/videos/KAFV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a picture of something' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.9> to <15.2>." + } + ] + }, + { + "video_id": "KAFV7", + "question_id": "KAFV7", + "video_file": "Charades/videos/KAFV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts their phone away'." + }, + { + "from": "gpt", + "value": "From <23.1> to <28.4>." + } + ] + }, + { + "video_id": "KAFV7", + "question_id": "KAFV7", + "video_file": "Charades/videos/KAFV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes a picture with their phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.9> to <15.2>." + } + ] + }, + { + "video_id": "KAFV7", + "question_id": "KAFV7", + "video_file": "Charades/videos/KAFV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person lies down on the sofa'." + }, + { + "from": "gpt", + "value": "From <26.8> to <31.0>." + } + ] + }, + { + "video_id": "XUV0Y", + "question_id": "XUV0Y", + "video_file": "Charades/videos/XUV0Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.2> to <20.4>." + } + ] + }, + { + "video_id": "XUV0Y", + "question_id": "XUV0Y", + "video_file": "Charades/videos/XUV0Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating in a living room'." + }, + { + "from": "gpt", + "value": "From <14.2> to <20.4>." + } + ] + }, + { + "video_id": "F99PM", + "question_id": "F99PM", + "video_file": "Charades/videos/F99PM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person open the cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.1> to <30.4>." + } + ] + }, + { + "video_id": "1D2OX", + "question_id": "1D2OX", + "video_file": "Charades/videos/1D2OX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is putting shoes on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.7> to <14.1>." + } + ] + }, + { + "video_id": "8BBX0", + "question_id": "8BBX0", + "video_file": "Charades/videos/8BBX0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opening the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.3> to <19.3>." + } + ] + }, + { + "video_id": "3VM0R", + "question_id": "3VM0R", + "video_file": "Charades/videos/3VM0R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person is putting a bag down on a desk' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.7> to <27.6>." + } + ] + }, + { + "video_id": "3VM0R", + "question_id": "3VM0R", + "video_file": "Charades/videos/3VM0R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is snuggling with a pillow on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <18.9>." + } + ] + }, + { + "video_id": "5BTFZ", + "question_id": "5BTFZ", + "video_file": "Charades/videos/5BTFZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person is standing by a window sneezing'." + }, + { + "from": "gpt", + "value": "From <17.3> to <30.7>." + } + ] + }, + { + "video_id": "5BTFZ", + "question_id": "5BTFZ", + "video_file": "Charades/videos/5BTFZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person is sneezing by the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.3> to <30.7>." + } + ] + }, + { + "video_id": "5BTFZ", + "question_id": "5BTFZ", + "video_file": "Charades/videos/5BTFZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person is looking out a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.5> to <32.0>." + } + ] + }, + { + "video_id": "OKYYQ", + "question_id": "OKYYQ", + "video_file": "Charades/videos/OKYYQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person closes a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <8.5>." + } + ] + }, + { + "video_id": "OKYYQ", + "question_id": "OKYYQ", + "video_file": "Charades/videos/OKYYQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closing the cabinet door'." + }, + { + "from": "gpt", + "value": "From <24.4> to <28.0>." + } + ] + }, + { + "video_id": "OKYYQ", + "question_id": "OKYYQ", + "video_file": "Charades/videos/OKYYQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person reads a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "5YA6D", + "question_id": "5YA6D", + "video_file": "Charades/videos/5YA6D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the front door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <12.8>." + } + ] + }, + { + "video_id": "5YA6D", + "question_id": "5YA6D", + "video_file": "Charades/videos/5YA6D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take off their shoes'." + }, + { + "from": "gpt", + "value": "From <14.0> to <23.2>." + } + ] + }, + { + "video_id": "5YA6D", + "question_id": "5YA6D", + "video_file": "Charades/videos/5YA6D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <23.2>." + } + ] + }, + { + "video_id": "5YA6D", + "question_id": "5YA6D", + "video_file": "Charades/videos/5YA6D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <12.8>." + } + ] + }, + { + "video_id": "5NK4L", + "question_id": "5NK4L", + "video_file": "Charades/videos/5NK4L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is tidying up the things in the cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.3> to <35.0>." + } + ] + }, + { + "video_id": "Q01UB", + "question_id": "Q01UB", + "video_file": "Charades/videos/Q01UB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts some shoes next to an open laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.9> to <32.4>." + } + ] + }, + { + "video_id": "Q01UB", + "question_id": "Q01UB", + "video_file": "Charades/videos/Q01UB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is dressing by putting on a jacket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "YQOPM", + "question_id": "YQOPM", + "video_file": "Charades/videos/YQOPM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person throws some clothes into a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.8> to <18.1>." + } + ] + }, + { + "video_id": "OAC4U", + "question_id": "OAC4U", + "video_file": "Charades/videos/OAC4U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throw the towel on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.8> to <12.5>." + } + ] + }, + { + "video_id": "5HOAV", + "question_id": "5HOAV", + "video_file": "Charades/videos/5HOAV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting the camera on the shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.6> to <20.1>." + } + ] + }, + { + "video_id": "G8O8P", + "question_id": "G8O8P", + "video_file": "Charades/videos/G8O8P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is closing a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.3> to <8.9>." + } + ] + }, + { + "video_id": "G8O8P", + "question_id": "G8O8P", + "video_file": "Charades/videos/G8O8P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <2.3> to <8.9>." + } + ] + }, + { + "video_id": "G8O8P", + "question_id": "G8O8P", + "video_file": "Charades/videos/G8O8P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.1> to <11.3>." + } + ] + }, + { + "video_id": "AJRIJ", + "question_id": "AJRIJ", + "video_file": "Charades/videos/AJRIJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person snuggling with a pillow begins watching the doorway'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.9>." + } + ] + }, + { + "video_id": "TVQO9", + "question_id": "TVQO9", + "video_file": "Charades/videos/TVQO9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking cup of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <9.0>." + } + ] + }, + { + "video_id": "TVQO9", + "question_id": "TVQO9", + "video_file": "Charades/videos/TVQO9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person drinks from the cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.3> to <9.0>." + } + ] + }, + { + "video_id": "OSCS2", + "question_id": "OSCS2", + "video_file": "Charades/videos/OSCS2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is laying in bed holding a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.7>." + } + ] + }, + { + "video_id": "OSCS2", + "question_id": "OSCS2", + "video_file": "Charades/videos/OSCS2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person look out a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.8> to <37.0>." + } + ] + }, + { + "video_id": "OSCS2", + "question_id": "OSCS2", + "video_file": "Charades/videos/OSCS2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person awoke from bed holding a blanket' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <31.7> to <37.0>." + } + ] + }, + { + "video_id": "OSCS2", + "question_id": "OSCS2", + "video_file": "Charades/videos/OSCS2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person moved to the window to look outside' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.8> to <37.0>." + } + ] + }, + { + "video_id": "OSCS2", + "question_id": "OSCS2", + "video_file": "Charades/videos/OSCS2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens in bed'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.1>." + } + ] + }, + { + "video_id": "F08QY", + "question_id": "F08QY", + "video_file": "Charades/videos/F08QY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person opens up a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "XKLW2", + "question_id": "XKLW2", + "video_file": "Charades/videos/XKLW2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person walks through the nearest doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.2> to <13.7>." + } + ] + }, + { + "video_id": "5BQMX", + "question_id": "5BQMX", + "video_file": "Charades/videos/5BQMX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is turning off a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.2> to <11.9>." + } + ] + }, + { + "video_id": "5BQMX", + "question_id": "5BQMX", + "video_file": "Charades/videos/5BQMX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <23.0>." + } + ] + }, + { + "video_id": "5BQMX", + "question_id": "5BQMX", + "video_file": "Charades/videos/5BQMX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is throwing a sofa pillow onto a sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <8.1>." + } + ] + }, + { + "video_id": "HDP39", + "question_id": "HDP39", + "video_file": "Charades/videos/HDP39.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person wearing a towel drinks water from a bottle' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.0>." + } + ] + }, + { + "video_id": "HDP39", + "question_id": "HDP39", + "video_file": "Charades/videos/HDP39.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a shirtless person drinks out of a bottle'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.0>." + } + ] + }, + { + "video_id": "ZEC1D", + "question_id": "ZEC1D", + "video_file": "Charades/videos/ZEC1D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a box out' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.8> to <26.8>." + } + ] + }, + { + "video_id": "ZEC1D", + "question_id": "ZEC1D", + "video_file": "Charades/videos/ZEC1D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a picture of the box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <26.8>." + } + ] + }, + { + "video_id": "DP8U3", + "question_id": "DP8U3", + "video_file": "Charades/videos/DP8U3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person snuggling a blanket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <11.3>." + } + ] + }, + { + "video_id": "CDT32", + "question_id": "CDT32", + "video_file": "Charades/videos/CDT32.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws it on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.4> to <21.8>." + } + ] + }, + { + "video_id": "HGSB7", + "question_id": "HGSB7", + "video_file": "Charades/videos/HGSB7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes their laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.4> to <20.2>." + } + ] + }, + { + "video_id": "HGSB7", + "question_id": "HGSB7", + "video_file": "Charades/videos/HGSB7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sitting in the chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.0> to <32.0>." + } + ] + }, + { + "video_id": "WGUW8", + "question_id": "WGUW8", + "video_file": "Charades/videos/WGUW8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person undresses in front of a mirror' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <15.5>." + } + ] + }, + { + "video_id": "WGUW8", + "question_id": "WGUW8", + "video_file": "Charades/videos/WGUW8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the clothes in a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <27.4>." + } + ] + }, + { + "video_id": "WGUW8", + "question_id": "WGUW8", + "video_file": "Charades/videos/WGUW8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is getting undressed in front of a mirror' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.1> to <15.5>." + } + ] + }, + { + "video_id": "Z9WWJ", + "question_id": "Z9WWJ", + "video_file": "Charades/videos/Z9WWJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a broom'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.9>." + } + ] + }, + { + "video_id": "Z9WWJ", + "question_id": "Z9WWJ", + "video_file": "Charades/videos/Z9WWJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person throws the broom into a corner' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.3> to <7.4>." + } + ] + }, + { + "video_id": "Z9WWJ", + "question_id": "Z9WWJ", + "video_file": "Charades/videos/Z9WWJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs into the basement'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "Z9WWJ", + "question_id": "Z9WWJ", + "video_file": "Charades/videos/Z9WWJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs down the stairs holding a broom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.9>." + } + ] + }, + { + "video_id": "CJHFD", + "question_id": "CJHFD", + "video_file": "Charades/videos/CJHFD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person holding a sandwich opens'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "CJHFD", + "question_id": "CJHFD", + "video_file": "Charades/videos/CJHFD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.6> to <29.1>." + } + ] + }, + { + "video_id": "HHEIV", + "question_id": "HHEIV", + "video_file": "Charades/videos/HHEIV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person smiling on the sofa'." + }, + { + "from": "gpt", + "value": "From <11.6> to <24.5>." + } + ] + }, + { + "video_id": "HHEIV", + "question_id": "HHEIV", + "video_file": "Charades/videos/HHEIV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person is smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <24.5>." + } + ] + }, + { + "video_id": "UCX0K", + "question_id": "UCX0K", + "video_file": "Charades/videos/UCX0K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sneezing on a picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "COPHL", + "question_id": "COPHL", + "video_file": "Charades/videos/COPHL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opened a book'." + }, + { + "from": "gpt", + "value": "From <9.3> to <14.2>." + } + ] + }, + { + "video_id": "COPHL", + "question_id": "COPHL", + "video_file": "Charades/videos/COPHL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens book'." + }, + { + "from": "gpt", + "value": "From <9.3> to <14.2>." + } + ] + }, + { + "video_id": "COPHL", + "question_id": "COPHL", + "video_file": "Charades/videos/COPHL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person is awakening from falling asleep behind desk' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <12.7>." + } + ] + }, + { + "video_id": "COPHL", + "question_id": "COPHL", + "video_file": "Charades/videos/COPHL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they open up a book on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.3> to <14.2>." + } + ] + }, + { + "video_id": "F0PNU", + "question_id": "F0PNU", + "video_file": "Charades/videos/F0PNU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <16.1>." + } + ] + }, + { + "video_id": "9C4JX", + "question_id": "9C4JX", + "video_file": "Charades/videos/9C4JX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding a bag full of items' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.4>." + } + ] + }, + { + "video_id": "BNEQL", + "question_id": "BNEQL", + "video_file": "Charades/videos/BNEQL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is dressing in the laundry room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <20.0>." + } + ] + }, + { + "video_id": "BNEQL", + "question_id": "BNEQL", + "video_file": "Charades/videos/BNEQL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a drink from a cup'." + }, + { + "from": "gpt", + "value": "From <14.3> to <26.7>." + } + ] + }, + { + "video_id": "BNEQL", + "question_id": "BNEQL", + "video_file": "Charades/videos/BNEQL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person picking up the coffee cup to take a drink'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <26.7>." + } + ] + }, + { + "video_id": "LF49A", + "question_id": "LF49A", + "video_file": "Charades/videos/LF49A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is opening a bag in the closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.6> to <16.4>." + } + ] + }, + { + "video_id": "I7MEA", + "question_id": "I7MEA", + "video_file": "Charades/videos/I7MEA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes out a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <9.6>." + } + ] + }, + { + "video_id": "I7MEA", + "question_id": "I7MEA", + "video_file": "Charades/videos/I7MEA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person continues cooking' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "I7MEA", + "question_id": "I7MEA", + "video_file": "Charades/videos/I7MEA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <7.3>." + } + ] + }, + { + "video_id": "0IFCL", + "question_id": "0IFCL", + "video_file": "Charades/videos/0IFCL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person closes the door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "0IFCL", + "question_id": "0IFCL", + "video_file": "Charades/videos/0IFCL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person closes a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "0IFCL", + "question_id": "0IFCL", + "video_file": "Charades/videos/0IFCL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.3> to <38.9>." + } + ] + }, + { + "video_id": "NUMAW", + "question_id": "NUMAW", + "video_file": "Charades/videos/NUMAW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is smiling'." + }, + { + "from": "gpt", + "value": "From <23.0> to <28.4>." + } + ] + }, + { + "video_id": "NUMAW", + "question_id": "NUMAW", + "video_file": "Charades/videos/NUMAW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person is sneezing'." + }, + { + "from": "gpt", + "value": "From <1.7> to <12.2>." + } + ] + }, + { + "video_id": "FI2TP", + "question_id": "FI2TP", + "video_file": "Charades/videos/FI2TP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they begin to have rampant sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <23.7>." + } + ] + }, + { + "video_id": "EBTD6", + "question_id": "EBTD6", + "video_file": "Charades/videos/EBTD6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.2> to <26.2>." + } + ] + }, + { + "video_id": "EBTD6", + "question_id": "EBTD6", + "video_file": "Charades/videos/EBTD6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.5> to <9.5>." + } + ] + }, + { + "video_id": "EBTD6", + "question_id": "EBTD6", + "video_file": "Charades/videos/EBTD6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.2> to <26.2>." + } + ] + }, + { + "video_id": "EBTD6", + "question_id": "EBTD6", + "video_file": "Charades/videos/EBTD6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opend the door,sneezed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.5> to <9.5>." + } + ] + }, + { + "video_id": "EBTD6", + "question_id": "EBTD6", + "video_file": "Charades/videos/EBTD6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is running in place' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.7> to <31.9>." + } + ] + }, + { + "video_id": "EBTD6", + "question_id": "EBTD6", + "video_file": "Charades/videos/EBTD6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sneezing as they open a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <9.5>." + } + ] + }, + { + "video_id": "EBTD6", + "question_id": "EBTD6", + "video_file": "Charades/videos/EBTD6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.5> to <9.4>." + } + ] + }, + { + "video_id": "4N06O", + "question_id": "4N06O", + "video_file": "Charades/videos/4N06O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws a book through a doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.3> to <26.3>." + } + ] + }, + { + "video_id": "4N06O", + "question_id": "4N06O", + "video_file": "Charades/videos/4N06O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person awakens' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <12.2>." + } + ] + }, + { + "video_id": "4N06O", + "question_id": "4N06O", + "video_file": "Charades/videos/4N06O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins to look at a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.1> to <20.7>." + } + ] + }, + { + "video_id": "4N06O", + "question_id": "4N06O", + "video_file": "Charades/videos/4N06O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.3> to <26.3>." + } + ] + }, + { + "video_id": "4N06O", + "question_id": "4N06O", + "video_file": "Charades/videos/4N06O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person awakens in a bathroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <12.2>." + } + ] + }, + { + "video_id": "BT8WU", + "question_id": "BT8WU", + "video_file": "Charades/videos/BT8WU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin eating' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.2> to <32.0>." + } + ] + }, + { + "video_id": "BT8WU", + "question_id": "BT8WU", + "video_file": "Charades/videos/BT8WU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eat something' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.2> to <32.0>." + } + ] + }, + { + "video_id": "K8RW9", + "question_id": "K8RW9", + "video_file": "Charades/videos/K8RW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person turns on a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.6> to <19.5>." + } + ] + }, + { + "video_id": "K8RW9", + "question_id": "K8RW9", + "video_file": "Charades/videos/K8RW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turning on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.6> to <19.5>." + } + ] + }, + { + "video_id": "MUE7H", + "question_id": "MUE7H", + "video_file": "Charades/videos/MUE7H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put on their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <21.0>." + } + ] + }, + { + "video_id": "MUE7H", + "question_id": "MUE7H", + "video_file": "Charades/videos/MUE7H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is awakening'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "MUE7H", + "question_id": "MUE7H", + "video_file": "Charades/videos/MUE7H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting on shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <21.0>." + } + ] + }, + { + "video_id": "MUE7H", + "question_id": "MUE7H", + "video_file": "Charades/videos/MUE7H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person awakens on a bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "MUE7H", + "question_id": "MUE7H", + "video_file": "Charades/videos/MUE7H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <23.5>." + } + ] + }, + { + "video_id": "MUE7H", + "question_id": "MUE7H", + "video_file": "Charades/videos/MUE7H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begin tidying the floor with it'." + }, + { + "from": "gpt", + "value": "From <22.8> to <43.0>." + } + ] + }, + { + "video_id": "21MI8", + "question_id": "21MI8", + "video_file": "Charades/videos/21MI8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking from a cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "21MI8", + "question_id": "21MI8", + "video_file": "Charades/videos/21MI8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is opening a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <10.1>." + } + ] + }, + { + "video_id": "21MI8", + "question_id": "21MI8", + "video_file": "Charades/videos/21MI8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting a glass down on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.7> to <13.4>." + } + ] + }, + { + "video_id": "BQ07S", + "question_id": "BQ07S", + "video_file": "Charades/videos/BQ07S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is pouring water in a glass'." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.3>." + } + ] + }, + { + "video_id": "QRSP8", + "question_id": "QRSP8", + "video_file": "Charades/videos/QRSP8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turn off the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.1> to <19.6>." + } + ] + }, + { + "video_id": "QRSP8", + "question_id": "QRSP8", + "video_file": "Charades/videos/QRSP8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.1> to <19.6>." + } + ] + }, + { + "video_id": "1R9XW", + "question_id": "1R9XW", + "video_file": "Charades/videos/1R9XW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes a closet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.2> to <9.6>." + } + ] + }, + { + "video_id": "MXBFE", + "question_id": "MXBFE", + "video_file": "Charades/videos/MXBFE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they look at a picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.8> to <13.7>." + } + ] + }, + { + "video_id": "IFFPZ", + "question_id": "IFFPZ", + "video_file": "Charades/videos/IFFPZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person awakens in a bedroom'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <8.2>." + } + ] + }, + { + "video_id": "XOMJ4", + "question_id": "XOMJ4", + "video_file": "Charades/videos/XOMJ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <30.0>." + } + ] + }, + { + "video_id": "XOMJ4", + "question_id": "XOMJ4", + "video_file": "Charades/videos/XOMJ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'finally the person sits down in the chair'." + }, + { + "from": "gpt", + "value": "From <19.5> to <30.0>." + } + ] + }, + { + "video_id": "FNM5E", + "question_id": "FNM5E", + "video_file": "Charades/videos/FNM5E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'starts laughing .the person turns' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.1> to <12.1>." + } + ] + }, + { + "video_id": "FNM5E", + "question_id": "FNM5E", + "video_file": "Charades/videos/FNM5E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <14.7>." + } + ] + }, + { + "video_id": "FNM5E", + "question_id": "FNM5E", + "video_file": "Charades/videos/FNM5E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <14.7>." + } + ] + }, + { + "video_id": "FNM5E", + "question_id": "FNM5E", + "video_file": "Charades/videos/FNM5E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they begin laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.1> to <12.1>." + } + ] + }, + { + "video_id": "H9V9I", + "question_id": "H9V9I", + "video_file": "Charades/videos/H9V9I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is standing holding an open laptop'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "H9V9I", + "question_id": "H9V9I", + "video_file": "Charades/videos/H9V9I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person puts their laptop onto a desk' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.7> to <12.7>." + } + ] + }, + { + "video_id": "OUAIS", + "question_id": "OUAIS", + "video_file": "Charades/videos/OUAIS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws the pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.1> to <29.0>." + } + ] + }, + { + "video_id": "OUAIS", + "question_id": "OUAIS", + "video_file": "Charades/videos/OUAIS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws a pillow across the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.1> to <29.0>." + } + ] + }, + { + "video_id": "OUAIS", + "question_id": "OUAIS", + "video_file": "Charades/videos/OUAIS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.2> to <36.7>." + } + ] + }, + { + "video_id": "OUAIS", + "question_id": "OUAIS", + "video_file": "Charades/videos/OUAIS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person start dressing by putting on a sweater'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.9> to <26.4>." + } + ] + }, + { + "video_id": "OUAIS", + "question_id": "OUAIS", + "video_file": "Charades/videos/OUAIS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'this person is laying on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "5AE54", + "question_id": "5AE54", + "video_file": "Charades/videos/5AE54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person is tidying the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.1> to <28.0>." + } + ] + }, + { + "video_id": "5AE54", + "question_id": "5AE54", + "video_file": "Charades/videos/5AE54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'lies on the bed when a second person walks in'." + }, + { + "from": "gpt", + "value": "From <8.4> to <16.6>." + } + ] + }, + { + "video_id": "5AE54", + "question_id": "5AE54", + "video_file": "Charades/videos/5AE54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is running to the phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.5>." + } + ] + }, + { + "video_id": "FBHBP", + "question_id": "FBHBP", + "video_file": "Charades/videos/FBHBP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits on couch' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <7.3>." + } + ] + }, + { + "video_id": "FBHBP", + "question_id": "FBHBP", + "video_file": "Charades/videos/FBHBP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sits on the sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.9> to <7.3>." + } + ] + }, + { + "video_id": "FBHBP", + "question_id": "FBHBP", + "video_file": "Charades/videos/FBHBP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats some food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <14.7>." + } + ] + }, + { + "video_id": "FBHBP", + "question_id": "FBHBP", + "video_file": "Charades/videos/FBHBP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats the food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.5> to <14.7>." + } + ] + }, + { + "video_id": "B8PQL", + "question_id": "B8PQL", + "video_file": "Charades/videos/B8PQL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts the towel on the table'." + }, + { + "from": "gpt", + "value": "From <0.7> to <5.9>." + } + ] + }, + { + "video_id": "K47J5", + "question_id": "K47J5", + "video_file": "Charades/videos/K47J5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person in a home office takes of their shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.3> to <14.2>." + } + ] + }, + { + "video_id": "K47J5", + "question_id": "K47J5", + "video_file": "Charades/videos/K47J5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes off their shoes'." + }, + { + "from": "gpt", + "value": "From <6.3> to <14.2>." + } + ] + }, + { + "video_id": "CII5H", + "question_id": "CII5H", + "video_file": "Charades/videos/CII5H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a closet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.5> to <23.9>." + } + ] + }, + { + "video_id": "CII5H", + "question_id": "CII5H", + "video_file": "Charades/videos/CII5H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person with a blanket in hand open closet door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <23.9>." + } + ] + }, + { + "video_id": "CII5H", + "question_id": "CII5H", + "video_file": "Charades/videos/CII5H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put clothing on the shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.0> to <19.8>." + } + ] + }, + { + "video_id": "CII5H", + "question_id": "CII5H", + "video_file": "Charades/videos/CII5H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <23.9>." + } + ] + }, + { + "video_id": "CPFLM", + "question_id": "CPFLM", + "video_file": "Charades/videos/CPFLM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person starts throwing clothes onto a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.7> to <14.3>." + } + ] + }, + { + "video_id": "CPFLM", + "question_id": "CPFLM", + "video_file": "Charades/videos/CPFLM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sneezing into a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "GA4DZ", + "question_id": "GA4DZ", + "video_file": "Charades/videos/GA4DZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is awakening near a light'." + }, + { + "from": "gpt", + "value": "From <1.0> to <7.5>." + } + ] + }, + { + "video_id": "6KI0S", + "question_id": "6KI0S", + "video_file": "Charades/videos/6KI0S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person snuggling with a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <27.0>." + } + ] + }, + { + "video_id": "6KI0S", + "question_id": "6KI0S", + "video_file": "Charades/videos/6KI0S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person awakening'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.0>." + } + ] + }, + { + "video_id": "HOIKF", + "question_id": "HOIKF", + "video_file": "Charades/videos/HOIKF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person is working on a laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.8> to <32.0>." + } + ] + }, + { + "video_id": "HOIKF", + "question_id": "HOIKF", + "video_file": "Charades/videos/HOIKF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person is playing with a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <32.0>." + } + ] + }, + { + "video_id": "GL38X", + "question_id": "GL38X", + "video_file": "Charades/videos/GL38X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person is watching a picture on the wall'." + }, + { + "from": "gpt", + "value": "From <9.0> to <19.8>." + } + ] + }, + { + "video_id": "JTXUA", + "question_id": "JTXUA", + "video_file": "Charades/videos/JTXUA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sitting down on a sofa'." + }, + { + "from": "gpt", + "value": "From <27.6> to <37.0>." + } + ] + }, + { + "video_id": "JTXUA", + "question_id": "JTXUA", + "video_file": "Charades/videos/JTXUA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.3> to <13.3>." + } + ] + }, + { + "video_id": "JTXUA", + "question_id": "JTXUA", + "video_file": "Charades/videos/JTXUA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person sits on a couch' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.6> to <37.0>." + } + ] + }, + { + "video_id": "JTXUA", + "question_id": "JTXUA", + "video_file": "Charades/videos/JTXUA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person look out the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.4> to <29.7>." + } + ] + }, + { + "video_id": "JTXUA", + "question_id": "JTXUA", + "video_file": "Charades/videos/JTXUA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits on the sofa'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.6> to <37.0>." + } + ] + }, + { + "video_id": "JTXUA", + "question_id": "JTXUA", + "video_file": "Charades/videos/JTXUA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person fixes light bulb'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.6> to <54.2>." + } + ] + }, + { + "video_id": "JTXUA", + "question_id": "JTXUA", + "video_file": "Charades/videos/JTXUA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <8.8>." + } + ] + }, + { + "video_id": "JTXUA", + "question_id": "JTXUA", + "video_file": "Charades/videos/JTXUA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <8.8>." + } + ] + }, + { + "video_id": "JTXUA", + "question_id": "JTXUA", + "video_file": "Charades/videos/JTXUA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person fixing a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <33.6> to <54.2>." + } + ] + }, + { + "video_id": "CPDH9", + "question_id": "CPDH9", + "video_file": "Charades/videos/CPDH9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the glass on the shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.9> to <12.1>." + } + ] + }, + { + "video_id": "7L0TN", + "question_id": "7L0TN", + "video_file": "Charades/videos/7L0TN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "XEGTL", + "question_id": "XEGTL", + "video_file": "Charades/videos/XEGTL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person laughs into a mirror'." + }, + { + "from": "gpt", + "value": "From <22.6> to <28.9>." + } + ] + }, + { + "video_id": "QDZ38", + "question_id": "QDZ38", + "video_file": "Charades/videos/QDZ38.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs into a bathroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.4> to <9.2>." + } + ] + }, + { + "video_id": "QDZ38", + "question_id": "QDZ38", + "video_file": "Charades/videos/QDZ38.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs into the room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <9.2>." + } + ] + }, + { + "video_id": "QDZ38", + "question_id": "QDZ38", + "video_file": "Charades/videos/QDZ38.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person washes the mirror with a towel' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <15.6>." + } + ] + }, + { + "video_id": "QDZ38", + "question_id": "QDZ38", + "video_file": "Charades/videos/QDZ38.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person runs into the bathroom'." + }, + { + "from": "gpt", + "value": "From <2.4> to <9.2>." + } + ] + }, + { + "video_id": "QDZ38", + "question_id": "QDZ38", + "video_file": "Charades/videos/QDZ38.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts washing the mirror'." + }, + { + "from": "gpt", + "value": "From <9.8> to <15.6>." + } + ] + }, + { + "video_id": "ZIWZH", + "question_id": "ZIWZH", + "video_file": "Charades/videos/ZIWZH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person close the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "ZIWZH", + "question_id": "ZIWZH", + "video_file": "Charades/videos/ZIWZH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting on a chair opening a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <13.5>." + } + ] + }, + { + "video_id": "9JZO3", + "question_id": "9JZO3", + "video_file": "Charades/videos/9JZO3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'awakening the person snuggling with the pillow'." + }, + { + "from": "gpt", + "value": "From <16.7> to <30.0>." + } + ] + }, + { + "video_id": "X1KKZ", + "question_id": "X1KKZ", + "video_file": "Charades/videos/X1KKZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person throws a pillow onto the sofa' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.7> to <15.2>." + } + ] + }, + { + "video_id": "X1KKZ", + "question_id": "X1KKZ", + "video_file": "Charades/videos/X1KKZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs to the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <17.7>." + } + ] + }, + { + "video_id": "DUNVR", + "question_id": "DUNVR", + "video_file": "Charades/videos/DUNVR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they run out of the room with the pantry' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.0> to <32.0>." + } + ] + }, + { + "video_id": "DUNVR", + "question_id": "DUNVR", + "video_file": "Charades/videos/DUNVR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the first person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.1> to <31.4>." + } + ] + }, + { + "video_id": "DUNVR", + "question_id": "DUNVR", + "video_file": "Charades/videos/DUNVR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person both run away'." + }, + { + "from": "gpt", + "value": "From <27.0> to <32.0>." + } + ] + }, + { + "video_id": "CO3LU", + "question_id": "CO3LU", + "video_file": "Charades/videos/CO3LU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walks over to the light switch'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.5> to <38.0>." + } + ] + }, + { + "video_id": "CO3LU", + "question_id": "CO3LU", + "video_file": "Charades/videos/CO3LU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <31.5> to <38.0>." + } + ] + }, + { + "video_id": "X8KIS", + "question_id": "X8KIS", + "video_file": "Charades/videos/X8KIS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks through a doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "Q24KP", + "question_id": "Q24KP", + "video_file": "Charades/videos/Q24KP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts a towel away'." + }, + { + "from": "gpt", + "value": "From <9.8> to <17.1>." + } + ] + }, + { + "video_id": "Q24KP", + "question_id": "Q24KP", + "video_file": "Charades/videos/Q24KP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person puts a blanket on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.5> to <18.1>." + } + ] + }, + { + "video_id": "SW64W", + "question_id": "SW64W", + "video_file": "Charades/videos/SW64W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person is opening a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <13.9>." + } + ] + }, + { + "video_id": "SW64W", + "question_id": "SW64W", + "video_file": "Charades/videos/SW64W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening up their laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <13.9>." + } + ] + }, + { + "video_id": "SW64W", + "question_id": "SW64W", + "video_file": "Charades/videos/SW64W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting at a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.4>." + } + ] + }, + { + "video_id": "SW64W", + "question_id": "SW64W", + "video_file": "Charades/videos/SW64W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person open their laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <13.9>." + } + ] + }, + { + "video_id": "JGYF6", + "question_id": "JGYF6", + "video_file": "Charades/videos/JGYF6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is lying on the sofa sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.4> to <8.0>." + } + ] + }, + { + "video_id": "JGYF6", + "question_id": "JGYF6", + "video_file": "Charades/videos/JGYF6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person talking on the phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <19.0>." + } + ] + }, + { + "video_id": "NSDFA", + "question_id": "NSDFA", + "video_file": "Charades/videos/NSDFA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the box down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.6> to <29.2>." + } + ] + }, + { + "video_id": "NSDFA", + "question_id": "NSDFA", + "video_file": "Charades/videos/NSDFA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take a picture'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.7> to <36.0>." + } + ] + }, + { + "video_id": "CM5SK", + "question_id": "CM5SK", + "video_file": "Charades/videos/CM5SK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "CM5SK", + "question_id": "CM5SK", + "video_file": "Charades/videos/CM5SK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person goes to start reading a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.4> to <17.5>." + } + ] + }, + { + "video_id": "CM5SK", + "question_id": "CM5SK", + "video_file": "Charades/videos/CM5SK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins reading a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.4> to <17.5>." + } + ] + }, + { + "video_id": "VMRAV", + "question_id": "VMRAV", + "video_file": "Charades/videos/VMRAV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on light grabs a napkin'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <17.6>." + } + ] + }, + { + "video_id": "VMRAV", + "question_id": "VMRAV", + "video_file": "Charades/videos/VMRAV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns on the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.5> to <17.6>." + } + ] + }, + { + "video_id": "34J80", + "question_id": "34J80", + "video_file": "Charades/videos/34J80.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person begins putting food into the cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.3> to <28.0>." + } + ] + }, + { + "video_id": "UY7BK", + "question_id": "UY7BK", + "video_file": "Charades/videos/UY7BK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looking out the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <29.3>." + } + ] + }, + { + "video_id": "3VWR9", + "question_id": "3VWR9", + "video_file": "Charades/videos/3VWR9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person close the door'." + }, + { + "from": "gpt", + "value": "From <10.7> to <16.9>." + } + ] + }, + { + "video_id": "3VWR9", + "question_id": "3VWR9", + "video_file": "Charades/videos/3VWR9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.7> to <16.9>." + } + ] + }, + { + "video_id": "WV8SY", + "question_id": "WV8SY", + "video_file": "Charades/videos/WV8SY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.1> to <19.0>." + } + ] + }, + { + "video_id": "WV8SY", + "question_id": "WV8SY", + "video_file": "Charades/videos/WV8SY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person running into the house' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <20.0>." + } + ] + }, + { + "video_id": "WV8SY", + "question_id": "WV8SY", + "video_file": "Charades/videos/WV8SY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes off their shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.8> to <23.0>." + } + ] + }, + { + "video_id": "3AE3T", + "question_id": "3AE3T", + "video_file": "Charades/videos/3AE3T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a second person sits on a sofa'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <19.5>." + } + ] + }, + { + "video_id": "43UGG", + "question_id": "43UGG", + "video_file": "Charades/videos/43UGG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is lying on the carpet eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.1> to <11.8>." + } + ] + }, + { + "video_id": "43UGG", + "question_id": "43UGG", + "video_file": "Charades/videos/43UGG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts on their shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <43.4> to <54.0>." + } + ] + }, + { + "video_id": "43UGG", + "question_id": "43UGG", + "video_file": "Charades/videos/43UGG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating a sandwich on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <11.8>." + } + ] + }, + { + "video_id": "4C5S6", + "question_id": "4C5S6", + "video_file": "Charades/videos/4C5S6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open the refrigerator door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <26.4>." + } + ] + }, + { + "video_id": "4C5S6", + "question_id": "4C5S6", + "video_file": "Charades/videos/4C5S6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person the open the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <26.4>." + } + ] + }, + { + "video_id": "FTETB", + "question_id": "FTETB", + "video_file": "Charades/videos/FTETB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the box down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.9> to <13.7>." + } + ] + }, + { + "video_id": "FTETB", + "question_id": "FTETB", + "video_file": "Charades/videos/FTETB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take something out of the box'." + }, + { + "from": "gpt", + "value": "From <0.2> to <11.6>." + } + ] + }, + { + "video_id": "FTETB", + "question_id": "FTETB", + "video_file": "Charades/videos/FTETB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes food out of a box'." + }, + { + "from": "gpt", + "value": "From <0.2> to <11.6>." + } + ] + }, + { + "video_id": "77UTO", + "question_id": "77UTO", + "video_file": "Charades/videos/77UTO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person washes a window with a towel' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.2> to <33.0>." + } + ] + }, + { + "video_id": "77UTO", + "question_id": "77UTO", + "video_file": "Charades/videos/77UTO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is pouring a cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.8>." + } + ] + }, + { + "video_id": "77UTO", + "question_id": "77UTO", + "video_file": "Charades/videos/77UTO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person pours a drink from one cup to another'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.8>." + } + ] + }, + { + "video_id": "BEC09", + "question_id": "BEC09", + "video_file": "Charades/videos/BEC09.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person in a kitchen is cooking food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.8> to <27.0>." + } + ] + }, + { + "video_id": "BEC09", + "question_id": "BEC09", + "video_file": "Charades/videos/BEC09.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.9> to <19.8>." + } + ] + }, + { + "video_id": "BEC09", + "question_id": "BEC09", + "video_file": "Charades/videos/BEC09.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a box on the counter' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.9> to <19.8>." + } + ] + }, + { + "video_id": "COM6M", + "question_id": "COM6M", + "video_file": "Charades/videos/COM6M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closed the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.3> to <23.0>." + } + ] + }, + { + "video_id": "COM6M", + "question_id": "COM6M", + "video_file": "Charades/videos/COM6M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.3> to <23.0>." + } + ] + }, + { + "video_id": "6RQHP", + "question_id": "6RQHP", + "video_file": "Charades/videos/6RQHP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.5> to <33.3>." + } + ] + }, + { + "video_id": "6RQHP", + "question_id": "6RQHP", + "video_file": "Charades/videos/6RQHP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is cooking at a stove in a kitchen'." + }, + { + "from": "gpt", + "value": "From <3.1> to <11.9>." + } + ] + }, + { + "video_id": "6RQHP", + "question_id": "6RQHP", + "video_file": "Charades/videos/6RQHP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is cooking on the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <11.9>." + } + ] + }, + { + "video_id": "6RQHP", + "question_id": "6RQHP", + "video_file": "Charades/videos/6RQHP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person washes some dishes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.4> to <21.1>." + } + ] + }, + { + "video_id": "I6IPP", + "question_id": "I6IPP", + "video_file": "Charades/videos/I6IPP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes a drink from a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.1> to <29.0>." + } + ] + }, + { + "video_id": "P36YW", + "question_id": "P36YW", + "video_file": "Charades/videos/P36YW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person are seen drinking a cup of coffee' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.6> to <28.0>." + } + ] + }, + { + "video_id": "P36YW", + "question_id": "P36YW", + "video_file": "Charades/videos/P36YW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the pillow down on the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.2> to <19.7>." + } + ] + }, + { + "video_id": "P36YW", + "question_id": "P36YW", + "video_file": "Charades/videos/P36YW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <20.3>." + } + ] + }, + { + "video_id": "9JIHR", + "question_id": "9JIHR", + "video_file": "Charades/videos/9JIHR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person kicking a door they stand up'." + }, + { + "from": "gpt", + "value": "From <13.2> to <20.6>." + } + ] + }, + { + "video_id": "AMLI4", + "question_id": "AMLI4", + "video_file": "Charades/videos/AMLI4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they pour a glass of water into a dish' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.5> to <26.4>." + } + ] + }, + { + "video_id": "AMLI4", + "question_id": "AMLI4", + "video_file": "Charades/videos/AMLI4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pour some water from a bottle into a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.5> to <26.4>." + } + ] + }, + { + "video_id": "MRXEW", + "question_id": "MRXEW", + "video_file": "Charades/videos/MRXEW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the box down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <27.2>." + } + ] + }, + { + "video_id": "DICXG", + "question_id": "DICXG", + "video_file": "Charades/videos/DICXG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person enters their house holding their laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.8>." + } + ] + }, + { + "video_id": "DICXG", + "question_id": "DICXG", + "video_file": "Charades/videos/DICXG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looks at the picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.9> to <25.6>." + } + ] + }, + { + "video_id": "DICXG", + "question_id": "DICXG", + "video_file": "Charades/videos/DICXG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks through a door way holding a laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.8>." + } + ] + }, + { + "video_id": "P87UK", + "question_id": "P87UK", + "video_file": "Charades/videos/P87UK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put food in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <22.3>." + } + ] + }, + { + "video_id": "8CQ9Y", + "question_id": "8CQ9Y", + "video_file": "Charades/videos/8CQ9Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the phone on the table'." + }, + { + "from": "gpt", + "value": "From <27.3> to <31.0>." + } + ] + }, + { + "video_id": "8CQ9Y", + "question_id": "8CQ9Y", + "video_file": "Charades/videos/8CQ9Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sneezing in the bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.3>." + } + ] + }, + { + "video_id": "VA7IT", + "question_id": "VA7IT", + "video_file": "Charades/videos/VA7IT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws the pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <22.0>." + } + ] + }, + { + "video_id": "VA7IT", + "question_id": "VA7IT", + "video_file": "Charades/videos/VA7IT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws the pillow down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <22.0>." + } + ] + }, + { + "video_id": "7LYA9", + "question_id": "7LYA9", + "video_file": "Charades/videos/7LYA9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws it on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <15.0>." + } + ] + }, + { + "video_id": "7LYA9", + "question_id": "7LYA9", + "video_file": "Charades/videos/7LYA9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person smiling as they throw a pillow on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <15.0>." + } + ] + }, + { + "video_id": "7LYA9", + "question_id": "7LYA9", + "video_file": "Charades/videos/7LYA9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.7> to <22.2>." + } + ] + }, + { + "video_id": "I5KUM", + "question_id": "I5KUM", + "video_file": "Charades/videos/I5KUM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person one is holding some medicine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "2J2EW", + "question_id": "2J2EW", + "video_file": "Charades/videos/2J2EW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person open a bag of groceries' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.1> to <20.7>." + } + ] + }, + { + "video_id": "22FJU", + "question_id": "22FJU", + "video_file": "Charades/videos/22FJU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks into the laundry room holding a picture' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <14.6>." + } + ] + }, + { + "video_id": "22FJU", + "question_id": "22FJU", + "video_file": "Charades/videos/22FJU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks into the room holding a picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <14.6>." + } + ] + }, + { + "video_id": "22FJU", + "question_id": "22FJU", + "video_file": "Charades/videos/22FJU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they turn on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <11.1>." + } + ] + }, + { + "video_id": "22FJU", + "question_id": "22FJU", + "video_file": "Charades/videos/22FJU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns on a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.4> to <11.1>." + } + ] + }, + { + "video_id": "RXRAS", + "question_id": "RXRAS", + "video_file": "Charades/videos/RXRAS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is tidying up a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.3>." + } + ] + }, + { + "video_id": "4ZWLA", + "question_id": "4ZWLA", + "video_file": "Charades/videos/4ZWLA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person is sneezing on a vacuum'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.6> to <23.2>." + } + ] + }, + { + "video_id": "0XV7B", + "question_id": "0XV7B", + "video_file": "Charades/videos/0XV7B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a door'." + }, + { + "from": "gpt", + "value": "From <1.2> to <9.0>." + } + ] + }, + { + "video_id": "J5TTU", + "question_id": "J5TTU", + "video_file": "Charades/videos/J5TTU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person washes their hands'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.1> to <30.4>." + } + ] + }, + { + "video_id": "G6LYD", + "question_id": "G6LYD", + "video_file": "Charades/videos/G6LYD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is throwing clothes on the floor'." + }, + { + "from": "gpt", + "value": "From <9.4> to <14.1>." + } + ] + }, + { + "video_id": "G6LYD", + "question_id": "G6LYD", + "video_file": "Charades/videos/G6LYD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they throw their clothes on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.4> to <14.1>." + } + ] + }, + { + "video_id": "ENDIE", + "question_id": "ENDIE", + "video_file": "Charades/videos/ENDIE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts down a bag of groceries' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <10.9>." + } + ] + }, + { + "video_id": "ENDIE", + "question_id": "ENDIE", + "video_file": "Charades/videos/ENDIE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinks from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.9> to <16.7>." + } + ] + }, + { + "video_id": "KUBPN", + "question_id": "KUBPN", + "video_file": "Charades/videos/KUBPN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating sandwich'." + }, + { + "from": "gpt", + "value": "From <2.3> to <9.4>." + } + ] + }, + { + "video_id": "KO34N", + "question_id": "KO34N", + "video_file": "Charades/videos/KO34N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the phone on the chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.8> to <11.9>." + } + ] + }, + { + "video_id": "S1XW9", + "question_id": "S1XW9", + "video_file": "Charades/videos/S1XW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <25.9> to <33.9>." + } + ] + }, + { + "video_id": "S1XW9", + "question_id": "S1XW9", + "video_file": "Charades/videos/S1XW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closing the doorknob' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.9> to <33.9>." + } + ] + }, + { + "video_id": "S1XW9", + "question_id": "S1XW9", + "video_file": "Charades/videos/S1XW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person walks through the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.2>." + } + ] + }, + { + "video_id": "S1XW9", + "question_id": "S1XW9", + "video_file": "Charades/videos/S1XW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person he closes the book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.7> to <29.2>." + } + ] + }, + { + "video_id": "OZ5SV", + "question_id": "OZ5SV", + "video_file": "Charades/videos/OZ5SV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the book down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <8.5>." + } + ] + }, + { + "video_id": "RLA7N", + "question_id": "RLA7N", + "video_file": "Charades/videos/RLA7N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person watching a cooking show on television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.0> to <13.5>." + } + ] + }, + { + "video_id": "RLA7N", + "question_id": "RLA7N", + "video_file": "Charades/videos/RLA7N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person started washing dishes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.4> to <39.0>." + } + ] + }, + { + "video_id": "RLA7N", + "question_id": "RLA7N", + "video_file": "Charades/videos/RLA7N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person was cooking dinner'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <13.5>." + } + ] + }, + { + "video_id": "5LETG", + "question_id": "5LETG", + "video_file": "Charades/videos/5LETG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns the light on' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.6> to <23.5>." + } + ] + }, + { + "video_id": "5LETG", + "question_id": "5LETG", + "video_file": "Charades/videos/5LETG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person turns on a light'." + }, + { + "from": "gpt", + "value": "From <16.6> to <23.5>." + } + ] + }, + { + "video_id": "JUF24", + "question_id": "JUF24", + "video_file": "Charades/videos/JUF24.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes food out of a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.4> to <26.9>." + } + ] + }, + { + "video_id": "NGWRE", + "question_id": "NGWRE", + "video_file": "Charades/videos/NGWRE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a cup'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.2>." + } + ] + }, + { + "video_id": "NGWRE", + "question_id": "NGWRE", + "video_file": "Charades/videos/NGWRE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is playing with their phone in the basement' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.7>." + } + ] + }, + { + "video_id": "HZXXN", + "question_id": "HZXXN", + "video_file": "Charades/videos/HZXXN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <11.1>." + } + ] + }, + { + "video_id": "HZXXN", + "question_id": "HZXXN", + "video_file": "Charades/videos/HZXXN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.5> to <11.1>." + } + ] + }, + { + "video_id": "HZXXN", + "question_id": "HZXXN", + "video_file": "Charades/videos/HZXXN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person was sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <8.8>." + } + ] + }, + { + "video_id": "7L4DW", + "question_id": "7L4DW", + "video_file": "Charades/videos/7L4DW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person takes a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "7L4DW", + "question_id": "7L4DW", + "video_file": "Charades/videos/7L4DW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person undressing in a basement opens a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <17.7>." + } + ] + }, + { + "video_id": "KCEJF", + "question_id": "KCEJF", + "video_file": "Charades/videos/KCEJF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looking in a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.9> to <22.0>." + } + ] + }, + { + "video_id": "KCEJF", + "question_id": "KCEJF", + "video_file": "Charades/videos/KCEJF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is walking up the stairs eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.1> to <9.3>." + } + ] + }, + { + "video_id": "KCEJF", + "question_id": "KCEJF", + "video_file": "Charades/videos/KCEJF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person watches something out a window'." + }, + { + "from": "gpt", + "value": "From <7.9> to <22.0>." + } + ] + }, + { + "video_id": "WCKTP", + "question_id": "WCKTP", + "video_file": "Charades/videos/WCKTP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a man putting something on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.0> to <9.5>." + } + ] + }, + { + "video_id": "KNDY2", + "question_id": "KNDY2", + "video_file": "Charades/videos/KNDY2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person is using a camera to take pictures'." + }, + { + "from": "gpt", + "value": "From <6.4> to <16.3>." + } + ] + }, + { + "video_id": "KNDY2", + "question_id": "KNDY2", + "video_file": "Charades/videos/KNDY2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person uses a camera to take pictures of the other'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <16.3>." + } + ] + }, + { + "video_id": "2JTRG", + "question_id": "2JTRG", + "video_file": "Charades/videos/2JTRG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person laughing as they eat a sandwich'." + }, + { + "from": "gpt", + "value": "From <15.0> to <25.1>." + } + ] + }, + { + "video_id": "Q1L24", + "question_id": "Q1L24", + "video_file": "Charades/videos/Q1L24.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is cooking at the stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "1KOZ4", + "question_id": "1KOZ4", + "video_file": "Charades/videos/1KOZ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person dresses in front of a mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.5> to <20.6>." + } + ] + }, + { + "video_id": "Z6U1X", + "question_id": "Z6U1X", + "video_file": "Charades/videos/Z6U1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a broom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "AMIHY", + "question_id": "AMIHY", + "video_file": "Charades/videos/AMIHY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes it with a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.4> to <29.5>." + } + ] + }, + { + "video_id": "AMIHY", + "question_id": "AMIHY", + "video_file": "Charades/videos/AMIHY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.8>." + } + ] + }, + { + "video_id": "J95X1", + "question_id": "J95X1", + "video_file": "Charades/videos/J95X1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "J95X1", + "question_id": "J95X1", + "video_file": "Charades/videos/J95X1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens their laptop'." + }, + { + "from": "gpt", + "value": "From <15.7> to <22.9>." + } + ] + }, + { + "video_id": "2OY8R", + "question_id": "2OY8R", + "video_file": "Charades/videos/2OY8R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they play a game on the laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <16.9>." + } + ] + }, + { + "video_id": "2OY8R", + "question_id": "2OY8R", + "video_file": "Charades/videos/2OY8R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <7.2>." + } + ] + }, + { + "video_id": "SX07Q", + "question_id": "SX07Q", + "video_file": "Charades/videos/SX07Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the blanket down'." + }, + { + "from": "gpt", + "value": "From <43.1> to <49.0>." + } + ] + }, + { + "video_id": "SX07Q", + "question_id": "SX07Q", + "video_file": "Charades/videos/SX07Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is opening a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.5> to <24.2>." + } + ] + }, + { + "video_id": "ZAO65", + "question_id": "ZAO65", + "video_file": "Charades/videos/ZAO65.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <7.2>." + } + ] + }, + { + "video_id": "ZAO65", + "question_id": "ZAO65", + "video_file": "Charades/videos/ZAO65.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks into a garage eating a cookie' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.9> to <13.8>." + } + ] + }, + { + "video_id": "ZAO65", + "question_id": "ZAO65", + "video_file": "Charades/videos/ZAO65.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person take off their shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <26.9>." + } + ] + }, + { + "video_id": "75HWR", + "question_id": "75HWR", + "video_file": "Charades/videos/75HWR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a sandwich on a plate'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.1> to <31.6>." + } + ] + }, + { + "video_id": "75HWR", + "question_id": "75HWR", + "video_file": "Charades/videos/75HWR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person runs by with a sandwich in their hands' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.1> to <31.6>." + } + ] + }, + { + "video_id": "FGBXJ", + "question_id": "FGBXJ", + "video_file": "Charades/videos/FGBXJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.4> to <27.2>." + } + ] + }, + { + "video_id": "PHH6B", + "question_id": "PHH6B", + "video_file": "Charades/videos/PHH6B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pours coffee into a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.8> to <29.2>." + } + ] + }, + { + "video_id": "PHH6B", + "question_id": "PHH6B", + "video_file": "Charades/videos/PHH6B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating food from a plate' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.1>." + } + ] + }, + { + "video_id": "PHH6B", + "question_id": "PHH6B", + "video_file": "Charades/videos/PHH6B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats some food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.1>." + } + ] + }, + { + "video_id": "PHH6B", + "question_id": "PHH6B", + "video_file": "Charades/videos/PHH6B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pouring tea into a cup on the counter' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.8> to <29.2>." + } + ] + }, + { + "video_id": "8S9AF", + "question_id": "8S9AF", + "video_file": "Charades/videos/8S9AF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts a towel on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.9> to <25.6>." + } + ] + }, + { + "video_id": "8S9AF", + "question_id": "8S9AF", + "video_file": "Charades/videos/8S9AF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the towel down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.9> to <25.6>." + } + ] + }, + { + "video_id": "028CE", + "question_id": "028CE", + "video_file": "Charades/videos/028CE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is in a garage fixing a vacuum' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.0> to <11.0>." + } + ] + }, + { + "video_id": "028CE", + "question_id": "028CE", + "video_file": "Charades/videos/028CE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person start laughing as they get up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.3> to <14.6>." + } + ] + }, + { + "video_id": "GG1X2", + "question_id": "GG1X2", + "video_file": "Charades/videos/GG1X2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes their shoes off'." + }, + { + "from": "gpt", + "value": "From <9.2> to <17.0>." + } + ] + }, + { + "video_id": "GG1X2", + "question_id": "GG1X2", + "video_file": "Charades/videos/GG1X2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closing the door behind them' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <11.1>." + } + ] + }, + { + "video_id": "F9RY4", + "question_id": "F9RY4", + "video_file": "Charades/videos/F9RY4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put the drinks on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.9> to <20.7>." + } + ] + }, + { + "video_id": "2TU40", + "question_id": "2TU40", + "video_file": "Charades/videos/2TU40.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person began cleaning the blanket he was holding'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "2TU40", + "question_id": "2TU40", + "video_file": "Charades/videos/2TU40.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is putting a bag on the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "5S587", + "question_id": "5S587", + "video_file": "Charades/videos/5S587.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person after looking out the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <13.4>." + } + ] + }, + { + "video_id": "3Q3YY", + "question_id": "3Q3YY", + "video_file": "Charades/videos/3Q3YY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person closes a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <13.8>." + } + ] + }, + { + "video_id": "FL7X4", + "question_id": "FL7X4", + "video_file": "Charades/videos/FL7X4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinks from the glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.1> to <9.7>." + } + ] + }, + { + "video_id": "YGYU5", + "question_id": "YGYU5", + "video_file": "Charades/videos/YGYU5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begin to tidy clothes on the bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.8> to <38.0>." + } + ] + }, + { + "video_id": "YGYU5", + "question_id": "YGYU5", + "video_file": "Charades/videos/YGYU5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person sitting on the bed drinking coffee,stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <23.6>." + } + ] + }, + { + "video_id": "YGYU5", + "question_id": "YGYU5", + "video_file": "Charades/videos/YGYU5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person sits down on the bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <23.0>." + } + ] + }, + { + "video_id": "DIR8Q", + "question_id": "DIR8Q", + "video_file": "Charades/videos/DIR8Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is playing with a phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "DIR8Q", + "question_id": "DIR8Q", + "video_file": "Charades/videos/DIR8Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person starts putting items on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <10.7>." + } + ] + }, + { + "video_id": "827O1", + "question_id": "827O1", + "video_file": "Charades/videos/827O1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they are smiling' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.1> to <20.3>." + } + ] + }, + { + "video_id": "0GFE8", + "question_id": "0GFE8", + "video_file": "Charades/videos/0GFE8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person runs away' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.1> to <19.0>." + } + ] + }, + { + "video_id": "0GFE8", + "question_id": "0GFE8", + "video_file": "Charades/videos/0GFE8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person run out of the room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <19.0>." + } + ] + }, + { + "video_id": "X0CKA", + "question_id": "X0CKA", + "video_file": "Charades/videos/X0CKA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a drink from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.2> to <19.0>." + } + ] + }, + { + "video_id": "X0CKA", + "question_id": "X0CKA", + "video_file": "Charades/videos/X0CKA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins drinking a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <19.0>." + } + ] + }, + { + "video_id": "S3TZ1", + "question_id": "S3TZ1", + "video_file": "Charades/videos/S3TZ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begin cooking a meal'." + }, + { + "from": "gpt", + "value": "From <31.9> to <49.9>." + } + ] + }, + { + "video_id": "S3TZ1", + "question_id": "S3TZ1", + "video_file": "Charades/videos/S3TZ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens up the refrigerator'." + }, + { + "from": "gpt", + "value": "From <0.3> to <6.3>." + } + ] + }, + { + "video_id": "S3TZ1", + "question_id": "S3TZ1", + "video_file": "Charades/videos/S3TZ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes out some food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.6> to <12.4>." + } + ] + }, + { + "video_id": "LMX0P", + "question_id": "LMX0P", + "video_file": "Charades/videos/LMX0P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is playing with a phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.9> to <33.5>." + } + ] + }, + { + "video_id": "LMX0P", + "question_id": "LMX0P", + "video_file": "Charades/videos/LMX0P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person playing with their phone'." + }, + { + "from": "gpt", + "value": "From <18.9> to <33.5>." + } + ] + }, + { + "video_id": "AVVX3", + "question_id": "AVVX3", + "video_file": "Charades/videos/AVVX3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person throws the medicine on the floor'." + }, + { + "from": "gpt", + "value": "From <10.5> to <15.4>." + } + ] + }, + { + "video_id": "5DJ3S", + "question_id": "5DJ3S", + "video_file": "Charades/videos/5DJ3S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looking out the window'." + }, + { + "from": "gpt", + "value": "From <7.3> to <18.0>." + } + ] + }, + { + "video_id": "5DJ3S", + "question_id": "5DJ3S", + "video_file": "Charades/videos/5DJ3S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person snuggles up to a blanket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <26.0>." + } + ] + }, + { + "video_id": "5DJ3S", + "question_id": "5DJ3S", + "video_file": "Charades/videos/5DJ3S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking a cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <11.2>." + } + ] + }, + { + "video_id": "5DJ3S", + "question_id": "5DJ3S", + "video_file": "Charades/videos/5DJ3S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins to snuggle with a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <26.0>." + } + ] + }, + { + "video_id": "KTL2G", + "question_id": "KTL2G", + "video_file": "Charades/videos/KTL2G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a broom'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "KTL2G", + "question_id": "KTL2G", + "video_file": "Charades/videos/KTL2G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turn off the light'." + }, + { + "from": "gpt", + "value": "From <13.7> to <19.0>." + } + ] + }, + { + "video_id": "KTL2G", + "question_id": "KTL2G", + "video_file": "Charades/videos/KTL2G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns off a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.7> to <19.0>." + } + ] + }, + { + "video_id": "KTL2G", + "question_id": "KTL2G", + "video_file": "Charades/videos/KTL2G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they close the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.8> to <17.9>." + } + ] + }, + { + "video_id": "KTL2G", + "question_id": "KTL2G", + "video_file": "Charades/videos/KTL2G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.8> to <17.9>." + } + ] + }, + { + "video_id": "GKVBS", + "question_id": "GKVBS", + "video_file": "Charades/videos/GKVBS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts a tool on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.6> to <37.6>." + } + ] + }, + { + "video_id": "GKVBS", + "question_id": "GKVBS", + "video_file": "Charades/videos/GKVBS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'starts to take the bottom off .the person starts laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.8> to <41.0>." + } + ] + }, + { + "video_id": "GKVBS", + "question_id": "GKVBS", + "video_file": "Charades/videos/GKVBS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put a tool on a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.6> to <37.6>." + } + ] + }, + { + "video_id": "G70CT", + "question_id": "G70CT", + "video_file": "Charades/videos/G70CT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person turns on a light'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "G70CT", + "question_id": "G70CT", + "video_file": "Charades/videos/G70CT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person turns on the light in a bathroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "UEP20", + "question_id": "UEP20", + "video_file": "Charades/videos/UEP20.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins to eat it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.4> to <31.4>." + } + ] + }, + { + "video_id": "UEP20", + "question_id": "UEP20", + "video_file": "Charades/videos/UEP20.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating something' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.4> to <31.4>." + } + ] + }, + { + "video_id": "UEP20", + "question_id": "UEP20", + "video_file": "Charades/videos/UEP20.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts notebook away' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.8> to <19.8>." + } + ] + }, + { + "video_id": "UEP20", + "question_id": "UEP20", + "video_file": "Charades/videos/UEP20.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins to eat a sandwich'." + }, + { + "from": "gpt", + "value": "From <18.7> to <28.7>." + } + ] + }, + { + "video_id": "UEP20", + "question_id": "UEP20", + "video_file": "Charades/videos/UEP20.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts it on the shelf next to a picture'." + }, + { + "from": "gpt", + "value": "From <11.6> to <19.2>." + } + ] + }, + { + "video_id": "UEP20", + "question_id": "UEP20", + "video_file": "Charades/videos/UEP20.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person closes the book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <15.6>." + } + ] + }, + { + "video_id": "UEP20", + "question_id": "UEP20", + "video_file": "Charades/videos/UEP20.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person putting some papers on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.6> to <19.2>." + } + ] + }, + { + "video_id": "WCIBT", + "question_id": "WCIBT", + "video_file": "Charades/videos/WCIBT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws a bag on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.2> to <6.8>." + } + ] + }, + { + "video_id": "WCIBT", + "question_id": "WCIBT", + "video_file": "Charades/videos/WCIBT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <11.7>." + } + ] + }, + { + "video_id": "WCIBT", + "question_id": "WCIBT", + "video_file": "Charades/videos/WCIBT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throwing a bag on the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <6.8>." + } + ] + }, + { + "video_id": "WCIBT", + "question_id": "WCIBT", + "video_file": "Charades/videos/WCIBT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person walks through a doorway'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "WCIBT", + "question_id": "WCIBT", + "video_file": "Charades/videos/WCIBT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.6>." + } + ] + }, + { + "video_id": "GPJ6E", + "question_id": "GPJ6E", + "video_file": "Charades/videos/GPJ6E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put blanket on chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <25.6>." + } + ] + }, + { + "video_id": "HURN7", + "question_id": "HURN7", + "video_file": "Charades/videos/HURN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they washed their hands in the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.8> to <31.0>." + } + ] + }, + { + "video_id": "HURN7", + "question_id": "HURN7", + "video_file": "Charades/videos/HURN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts cooking a meal' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <23.4>." + } + ] + }, + { + "video_id": "HURN7", + "question_id": "HURN7", + "video_file": "Charades/videos/HURN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person undresses' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <13.0>." + } + ] + }, + { + "video_id": "HURN7", + "question_id": "HURN7", + "video_file": "Charades/videos/HURN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person washes his hands' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.8> to <31.0>." + } + ] + }, + { + "video_id": "HURN7", + "question_id": "HURN7", + "video_file": "Charades/videos/HURN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person was cooking' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <23.4>." + } + ] + }, + { + "video_id": "HURN7", + "question_id": "HURN7", + "video_file": "Charades/videos/HURN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person started to undress' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.0> to <13.0>." + } + ] + }, + { + "video_id": "TI9JB", + "question_id": "TI9JB", + "video_file": "Charades/videos/TI9JB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they close the dryer door'." + }, + { + "from": "gpt", + "value": "From <19.7> to <25.8>." + } + ] + }, + { + "video_id": "TI9JB", + "question_id": "TI9JB", + "video_file": "Charades/videos/TI9JB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person closes the washing machine door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.7> to <25.8>." + } + ] + }, + { + "video_id": "MNEQ4", + "question_id": "MNEQ4", + "video_file": "Charades/videos/MNEQ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes a blanket from a nearby chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.8> to <24.6>." + } + ] + }, + { + "video_id": "9A58F", + "question_id": "9A58F", + "video_file": "Charades/videos/9A58F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is throwing down their shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.9> to <8.1>." + } + ] + }, + { + "video_id": "9A58F", + "question_id": "9A58F", + "video_file": "Charades/videos/9A58F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws shoes on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.9> to <8.1>." + } + ] + }, + { + "video_id": "9A58F", + "question_id": "9A58F", + "video_file": "Charades/videos/9A58F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening a laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <12.9>." + } + ] + }, + { + "video_id": "LJ3AI", + "question_id": "LJ3AI", + "video_file": "Charades/videos/LJ3AI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting on the couch drinking a soda' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.5> to <7.7>." + } + ] + }, + { + "video_id": "LJ3AI", + "question_id": "LJ3AI", + "video_file": "Charades/videos/LJ3AI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking for a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <13.6>." + } + ] + }, + { + "video_id": "LJ3AI", + "question_id": "LJ3AI", + "video_file": "Charades/videos/LJ3AI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person is sitting on a sofa doing homework' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <7.7>." + } + ] + }, + { + "video_id": "D3KN2", + "question_id": "D3KN2", + "video_file": "Charades/videos/D3KN2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs into the dining room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.0>." + } + ] + }, + { + "video_id": "D3KN2", + "question_id": "D3KN2", + "video_file": "Charades/videos/D3KN2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put on their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.0> to <22.0>." + } + ] + }, + { + "video_id": "D3KN2", + "question_id": "D3KN2", + "video_file": "Charades/videos/D3KN2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts on shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.0> to <22.0>." + } + ] + }, + { + "video_id": "D3KN2", + "question_id": "D3KN2", + "video_file": "Charades/videos/D3KN2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a sneezing person runs into a room'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.0>." + } + ] + }, + { + "video_id": "GM4UP", + "question_id": "GM4UP", + "video_file": "Charades/videos/GM4UP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they begin sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.5> to <29.0>." + } + ] + }, + { + "video_id": "5Q3QB", + "question_id": "5Q3QB", + "video_file": "Charades/videos/5Q3QB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes a door multiple times' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.1> to <12.3>." + } + ] + }, + { + "video_id": "5Q3QB", + "question_id": "5Q3QB", + "video_file": "Charades/videos/5Q3QB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.1> to <12.3>." + } + ] + }, + { + "video_id": "VXM6A", + "question_id": "VXM6A", + "video_file": "Charades/videos/VXM6A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'putting them under the clothes.the person picks a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <18.5>." + } + ] + }, + { + "video_id": "VXM6A", + "question_id": "VXM6A", + "video_file": "Charades/videos/VXM6A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put it in a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.9> to <23.9>." + } + ] + }, + { + "video_id": "VXM6A", + "question_id": "VXM6A", + "video_file": "Charades/videos/VXM6A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting on the floor in the closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.2> to <12.0>." + } + ] + }, + { + "video_id": "VXM6A", + "question_id": "VXM6A", + "video_file": "Charades/videos/VXM6A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the box under the closet shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <20.2>." + } + ] + }, + { + "video_id": "VXM6A", + "question_id": "VXM6A", + "video_file": "Charades/videos/VXM6A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'put stuff in a box.the person gets up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.9> to <23.9>." + } + ] + }, + { + "video_id": "MTZ0T", + "question_id": "MTZ0T", + "video_file": "Charades/videos/MTZ0T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pours coffee from a cup in to the toilet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <15.7>." + } + ] + }, + { + "video_id": "OMHFD", + "question_id": "OMHFD", + "video_file": "Charades/videos/OMHFD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes out their phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <22.5>." + } + ] + }, + { + "video_id": "OMHFD", + "question_id": "OMHFD", + "video_file": "Charades/videos/OMHFD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person laughs at something on the screen' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.9> to <29.1>." + } + ] + }, + { + "video_id": "OMHFD", + "question_id": "OMHFD", + "video_file": "Charades/videos/OMHFD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.9> to <29.1>." + } + ] + }, + { + "video_id": "WREDC", + "question_id": "WREDC", + "video_file": "Charades/videos/WREDC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting the cup on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <24.0>." + } + ] + }, + { + "video_id": "WREDC", + "question_id": "WREDC", + "video_file": "Charades/videos/WREDC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person puts down a glass on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <24.0>." + } + ] + }, + { + "video_id": "YKD6U", + "question_id": "YKD6U", + "video_file": "Charades/videos/YKD6U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <14.2>." + } + ] + }, + { + "video_id": "YKD6U", + "question_id": "YKD6U", + "video_file": "Charades/videos/YKD6U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits on the chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.9> to <21.2>." + } + ] + }, + { + "video_id": "38BDO", + "question_id": "38BDO", + "video_file": "Charades/videos/38BDO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put it on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.8> to <34.0>." + } + ] + }, + { + "video_id": "38BDO", + "question_id": "38BDO", + "video_file": "Charades/videos/38BDO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts it on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.8> to <34.0>." + } + ] + }, + { + "video_id": "38BDO", + "question_id": "38BDO", + "video_file": "Charades/videos/38BDO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put a blanket in a box'." + }, + { + "from": "gpt", + "value": "From <20.3> to <34.0>." + } + ] + }, + { + "video_id": "38BDO", + "question_id": "38BDO", + "video_file": "Charades/videos/38BDO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts a blanket in a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.3> to <34.0>." + } + ] + }, + { + "video_id": "38BDO", + "question_id": "38BDO", + "video_file": "Charades/videos/38BDO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts a folded blanket in a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.3> to <34.0>." + } + ] + }, + { + "video_id": "WT9CR", + "question_id": "WT9CR", + "video_file": "Charades/videos/WT9CR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person swinging a door open' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.8>." + } + ] + }, + { + "video_id": "406LH", + "question_id": "406LH", + "video_file": "Charades/videos/406LH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.4> to <9.0>." + } + ] + }, + { + "video_id": "406LH", + "question_id": "406LH", + "video_file": "Charades/videos/406LH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person stands in the doorway eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.9> to <17.4>." + } + ] + }, + { + "video_id": "406LH", + "question_id": "406LH", + "video_file": "Charades/videos/406LH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats something'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <17.9>." + } + ] + }, + { + "video_id": "406LH", + "question_id": "406LH", + "video_file": "Charades/videos/406LH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person stand in the doorway eating food from a dish'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <17.9>." + } + ] + }, + { + "video_id": "406LH", + "question_id": "406LH", + "video_file": "Charades/videos/406LH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person close the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <20.1>." + } + ] + }, + { + "video_id": "452FV", + "question_id": "452FV", + "video_file": "Charades/videos/452FV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is holding a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.4> to <12.5>." + } + ] + }, + { + "video_id": "LXSIG", + "question_id": "LXSIG", + "video_file": "Charades/videos/LXSIG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person looks out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.8>." + } + ] + }, + { + "video_id": "KP53W", + "question_id": "KP53W", + "video_file": "Charades/videos/KP53W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they close a laptop that's on the counter' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.8> to <24.6>." + } + ] + }, + { + "video_id": "I1OLQ", + "question_id": "I1OLQ", + "video_file": "Charades/videos/I1OLQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.8> to <12.6>." + } + ] + }, + { + "video_id": "I1OLQ", + "question_id": "I1OLQ", + "video_file": "Charades/videos/I1OLQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person both watch more television'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.9>." + } + ] + }, + { + "video_id": "I1OLQ", + "question_id": "I1OLQ", + "video_file": "Charades/videos/I1OLQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person watching television'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.9>." + } + ] + }, + { + "video_id": "I1OLQ", + "question_id": "I1OLQ", + "video_file": "Charades/videos/I1OLQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walks into a room ,closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.4> to <14.6>." + } + ] + }, + { + "video_id": "H32FR", + "question_id": "H32FR", + "video_file": "Charades/videos/H32FR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put dishes in a corner' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.5> to <18.6>." + } + ] + }, + { + "video_id": "H32FR", + "question_id": "H32FR", + "video_file": "Charades/videos/H32FR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person 2 people open a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <5.3>." + } + ] + }, + { + "video_id": "1QT62", + "question_id": "1QT62", + "video_file": "Charades/videos/1QT62.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.3> to <29.5>." + } + ] + }, + { + "video_id": "7C9RP", + "question_id": "7C9RP", + "video_file": "Charades/videos/7C9RP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person turns on the light in the bathroom'." + }, + { + "from": "gpt", + "value": "From <1.6> to <5.9>." + } + ] + }, + { + "video_id": "7C9RP", + "question_id": "7C9RP", + "video_file": "Charades/videos/7C9RP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <5.9>." + } + ] + }, + { + "video_id": "7C9RP", + "question_id": "7C9RP", + "video_file": "Charades/videos/7C9RP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begins undressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.9> to <14.3>." + } + ] + }, + { + "video_id": "VG9II", + "question_id": "VG9II", + "video_file": "Charades/videos/VG9II.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinks from a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <18.4>." + } + ] + }, + { + "video_id": "WY433", + "question_id": "WY433", + "video_file": "Charades/videos/WY433.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pours some out into a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <19.9>." + } + ] + }, + { + "video_id": "WY433", + "question_id": "WY433", + "video_file": "Charades/videos/WY433.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks into room laughing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <12.6>." + } + ] + }, + { + "video_id": "WY433", + "question_id": "WY433", + "video_file": "Charades/videos/WY433.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person laughs as they walk through the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "WY433", + "question_id": "WY433", + "video_file": "Charades/videos/WY433.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours it into cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <19.9>." + } + ] + }, + { + "video_id": "WY433", + "question_id": "WY433", + "video_file": "Charades/videos/WY433.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours something in a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <19.9>." + } + ] + }, + { + "video_id": "77JB7", + "question_id": "77JB7", + "video_file": "Charades/videos/77JB7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.2> to <41.6>." + } + ] + }, + { + "video_id": "XJU8U", + "question_id": "XJU8U", + "video_file": "Charades/videos/XJU8U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks through the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.3> to <6.7>." + } + ] + }, + { + "video_id": "XJU8U", + "question_id": "XJU8U", + "video_file": "Charades/videos/XJU8U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes their shoes from the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.4> to <16.4>." + } + ] + }, + { + "video_id": "XJU8U", + "question_id": "XJU8U", + "video_file": "Charades/videos/XJU8U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "XJU8U", + "question_id": "XJU8U", + "video_file": "Charades/videos/XJU8U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens the basement door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "XJU8U", + "question_id": "XJU8U", + "video_file": "Charades/videos/XJU8U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.8> to <23.6>." + } + ] + }, + { + "video_id": "RVLXD", + "question_id": "RVLXD", + "video_file": "Charades/videos/RVLXD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they open a cabinet'." + }, + { + "from": "gpt", + "value": "From <11.5> to <17.4>." + } + ] + }, + { + "video_id": "RVLXD", + "question_id": "RVLXD", + "video_file": "Charades/videos/RVLXD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person grab the doorknob to open the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.5> to <25.0>." + } + ] + }, + { + "video_id": "RVLXD", + "question_id": "RVLXD", + "video_file": "Charades/videos/RVLXD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a doorknob'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.9> to <25.0>." + } + ] + }, + { + "video_id": "3A400", + "question_id": "3A400", + "video_file": "Charades/videos/3A400.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person runs into garage with books in hand' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "3A400", + "question_id": "3A400", + "video_file": "Charades/videos/3A400.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is running into the garage' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "3A400", + "question_id": "3A400", + "video_file": "Charades/videos/3A400.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens the books' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.1> to <8.3>." + } + ] + }, + { + "video_id": "CN480", + "question_id": "CN480", + "video_file": "Charades/videos/CN480.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits on bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <8.0>." + } + ] + }, + { + "video_id": "XWQ3Z", + "question_id": "XWQ3Z", + "video_file": "Charades/videos/XWQ3Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws two pillows'." + }, + { + "from": "gpt", + "value": "From <6.4> to <14.5>." + } + ] + }, + { + "video_id": "IWCXM", + "question_id": "IWCXM", + "video_file": "Charades/videos/IWCXM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person runs into the room with a box'." + }, + { + "from": "gpt", + "value": "From <0.5> to <7.7>." + } + ] + }, + { + "video_id": "IWCXM", + "question_id": "IWCXM", + "video_file": "Charades/videos/IWCXM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person running through a room with a box stops' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.5> to <7.7>." + } + ] + }, + { + "video_id": "59KH0", + "question_id": "59KH0", + "video_file": "Charades/videos/59KH0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <15.4>." + } + ] + }, + { + "video_id": "59KH0", + "question_id": "59KH0", + "video_file": "Charades/videos/59KH0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person is playing games on their laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <24.8>." + } + ] + }, + { + "video_id": "59KH0", + "question_id": "59KH0", + "video_file": "Charades/videos/59KH0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person playing games on a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <24.8>." + } + ] + }, + { + "video_id": "59KH0", + "question_id": "59KH0", + "video_file": "Charades/videos/59KH0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person is undressing in front of a mirror' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <15.4>." + } + ] + }, + { + "video_id": "GV40M", + "question_id": "GV40M", + "video_file": "Charades/videos/GV40M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person is awakening in bed by window'." + }, + { + "from": "gpt", + "value": "From <5.3> to <11.8>." + } + ] + }, + { + "video_id": "GV40M", + "question_id": "GV40M", + "video_file": "Charades/videos/GV40M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is awakening'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <11.8>." + } + ] + }, + { + "video_id": "W2QG1", + "question_id": "W2QG1", + "video_file": "Charades/videos/W2QG1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws their pillow onto a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <18.6>." + } + ] + }, + { + "video_id": "W2QG1", + "question_id": "W2QG1", + "video_file": "Charades/videos/W2QG1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws a pillow on to a chair'." + }, + { + "from": "gpt", + "value": "From <12.2> to <18.6>." + } + ] + }, + { + "video_id": "W2QG1", + "question_id": "W2QG1", + "video_file": "Charades/videos/W2QG1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person awakens in bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <13.8>." + } + ] + }, + { + "video_id": "W2QG1", + "question_id": "W2QG1", + "video_file": "Charades/videos/W2QG1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws a pillow off of the bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <18.6>." + } + ] + }, + { + "video_id": "T82IE", + "question_id": "T82IE", + "video_file": "Charades/videos/T82IE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the glass on the desk' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "T82IE", + "question_id": "T82IE", + "video_file": "Charades/videos/T82IE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person walks through drinking from a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "T82IE", + "question_id": "T82IE", + "video_file": "Charades/videos/T82IE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking from a glass'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "T82IE", + "question_id": "T82IE", + "video_file": "Charades/videos/T82IE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person drinks from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "T82IE", + "question_id": "T82IE", + "video_file": "Charades/videos/T82IE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looking at a book'." + }, + { + "from": "gpt", + "value": "From <9.8> to <17.4>." + } + ] + }, + { + "video_id": "T82IE", + "question_id": "T82IE", + "video_file": "Charades/videos/T82IE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits on the table'." + }, + { + "from": "gpt", + "value": "From <4.7> to <11.5>." + } + ] + }, + { + "video_id": "T82IE", + "question_id": "T82IE", + "video_file": "Charades/videos/T82IE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the glass on the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "O495M", + "question_id": "O495M", + "video_file": "Charades/videos/O495M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person tidying up a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <17.3>." + } + ] + }, + { + "video_id": "O495M", + "question_id": "O495M", + "video_file": "Charades/videos/O495M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is holding a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <9.1>." + } + ] + }, + { + "video_id": "61KCY", + "question_id": "61KCY", + "video_file": "Charades/videos/61KCY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person starts sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.5> to <29.5>." + } + ] + }, + { + "video_id": "61KCY", + "question_id": "61KCY", + "video_file": "Charades/videos/61KCY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting on on their clothes in the bathroom'." + }, + { + "from": "gpt", + "value": "From <8.5> to <17.0>." + } + ] + }, + { + "video_id": "5IOQB", + "question_id": "5IOQB", + "video_file": "Charades/videos/5IOQB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they drink a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.8> to <25.5>." + } + ] + }, + { + "video_id": "5IOQB", + "question_id": "5IOQB", + "video_file": "Charades/videos/5IOQB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes a drink from a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.8> to <25.5>." + } + ] + }, + { + "video_id": "5IOQB", + "question_id": "5IOQB", + "video_file": "Charades/videos/5IOQB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person laughing they take a drink of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.0> to <20.0>." + } + ] + }, + { + "video_id": "MCNRZ", + "question_id": "MCNRZ", + "video_file": "Charades/videos/MCNRZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person drinks from a coffee cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.0> to <24.0>." + } + ] + }, + { + "video_id": "MCNRZ", + "question_id": "MCNRZ", + "video_file": "Charades/videos/MCNRZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is drinking a cup of coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.0> to <24.0>." + } + ] + }, + { + "video_id": "17AZ5", + "question_id": "17AZ5", + "video_file": "Charades/videos/17AZ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opens the cabinet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.5> to <19.8>." + } + ] + }, + { + "video_id": "17AZ5", + "question_id": "17AZ5", + "video_file": "Charades/videos/17AZ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.5> to <19.8>." + } + ] + }, + { + "video_id": "17AZ5", + "question_id": "17AZ5", + "video_file": "Charades/videos/17AZ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is dressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.6> to <15.3>." + } + ] + }, + { + "video_id": "P6SOI", + "question_id": "P6SOI", + "video_file": "Charades/videos/P6SOI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'that same person opens the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.3> to <31.2>." + } + ] + }, + { + "video_id": "P6SOI", + "question_id": "P6SOI", + "video_file": "Charades/videos/P6SOI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open the door'." + }, + { + "from": "gpt", + "value": "From <21.3> to <31.2>." + } + ] + }, + { + "video_id": "PW9P0", + "question_id": "PW9P0", + "video_file": "Charades/videos/PW9P0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is undressing in a home office'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "PW9P0", + "question_id": "PW9P0", + "video_file": "Charades/videos/PW9P0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting down undressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "TU1C9", + "question_id": "TU1C9", + "video_file": "Charades/videos/TU1C9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.9> to <20.7>." + } + ] + }, + { + "video_id": "TU1C9", + "question_id": "TU1C9", + "video_file": "Charades/videos/TU1C9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a sneezing person is eating a donut'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "TU1C9", + "question_id": "TU1C9", + "video_file": "Charades/videos/TU1C9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinks from a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.0> to <32.0>." + } + ] + }, + { + "video_id": "KTKWL", + "question_id": "KTKWL", + "video_file": "Charades/videos/KTKWL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.1> to <15.2>." + } + ] + }, + { + "video_id": "KTKWL", + "question_id": "KTKWL", + "video_file": "Charades/videos/KTKWL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.4> to <20.8>." + } + ] + }, + { + "video_id": "3IJGQ", + "question_id": "3IJGQ", + "video_file": "Charades/videos/3IJGQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is lying down eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.3>." + } + ] + }, + { + "video_id": "3IJGQ", + "question_id": "3IJGQ", + "video_file": "Charades/videos/3IJGQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the dirty dishes on the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <18.1>." + } + ] + }, + { + "video_id": "3IJGQ", + "question_id": "3IJGQ", + "video_file": "Charades/videos/3IJGQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is seen smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.1>." + } + ] + }, + { + "video_id": "LU82W", + "question_id": "LU82W", + "video_file": "Charades/videos/LU82W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person throws a box onto the table'." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.5>." + } + ] + }, + { + "video_id": "LU82W", + "question_id": "LU82W", + "video_file": "Charades/videos/LU82W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person throws a box onto a counter' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.5>." + } + ] + }, + { + "video_id": "QMTF2", + "question_id": "QMTF2", + "video_file": "Charades/videos/QMTF2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws blanket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <19.0>." + } + ] + }, + { + "video_id": "QMTF2", + "question_id": "QMTF2", + "video_file": "Charades/videos/QMTF2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begin throwing the blanket towards the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <19.0>." + } + ] + }, + { + "video_id": "HAA4O", + "question_id": "HAA4O", + "video_file": "Charades/videos/HAA4O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a picture out the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <30.2>." + } + ] + }, + { + "video_id": "HAA4O", + "question_id": "HAA4O", + "video_file": "Charades/videos/HAA4O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <4.6> to <13.2>." + } + ] + }, + { + "video_id": "HAA4O", + "question_id": "HAA4O", + "video_file": "Charades/videos/HAA4O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take a picture of a view outside the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.0> to <30.2>." + } + ] + }, + { + "video_id": "R77MC", + "question_id": "R77MC", + "video_file": "Charades/videos/R77MC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the original person in the room puts down a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.7> to <20.9>." + } + ] + }, + { + "video_id": "6R6RS", + "question_id": "6R6RS", + "video_file": "Charades/videos/6R6RS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put their shoes on'." + }, + { + "from": "gpt", + "value": "From <14.4> to <23.6>." + } + ] + }, + { + "video_id": "6R6RS", + "question_id": "6R6RS", + "video_file": "Charades/videos/6R6RS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts on their shoes'." + }, + { + "from": "gpt", + "value": "From <14.4> to <23.6>." + } + ] + }, + { + "video_id": "J0VXY", + "question_id": "J0VXY", + "video_file": "Charades/videos/J0VXY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person start tidying up the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.9> to <30.3>." + } + ] + }, + { + "video_id": "IM2WK", + "question_id": "IM2WK", + "video_file": "Charades/videos/IM2WK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walks in room holding clothes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <10.0>." + } + ] + }, + { + "video_id": "IM2WK", + "question_id": "IM2WK", + "video_file": "Charades/videos/IM2WK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person throws some clothes onto a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <11.4>." + } + ] + }, + { + "video_id": "IM2WK", + "question_id": "IM2WK", + "video_file": "Charades/videos/IM2WK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts clothes on chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.7> to <11.2>." + } + ] + }, + { + "video_id": "SUI1X", + "question_id": "SUI1X", + "video_file": "Charades/videos/SUI1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is opening a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.7> to <10.3>." + } + ] + }, + { + "video_id": "46A6J", + "question_id": "46A6J", + "video_file": "Charades/videos/46A6J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a broom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.2> to <22.3>." + } + ] + }, + { + "video_id": "5DPC8", + "question_id": "5DPC8", + "video_file": "Charades/videos/5DPC8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits on a chair they take out their phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.9> to <37.0>." + } + ] + }, + { + "video_id": "5DPC8", + "question_id": "5DPC8", + "video_file": "Charades/videos/5DPC8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person turns on the light in a bathroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.4> to <7.7>." + } + ] + }, + { + "video_id": "857S9", + "question_id": "857S9", + "video_file": "Charades/videos/857S9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is sitting at a dining room table smiling' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "857S9", + "question_id": "857S9", + "video_file": "Charades/videos/857S9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begins smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "G1JBI", + "question_id": "G1JBI", + "video_file": "Charades/videos/G1JBI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "G1JBI", + "question_id": "G1JBI", + "video_file": "Charades/videos/G1JBI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they close the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.1> to <35.8>." + } + ] + }, + { + "video_id": "G1JBI", + "question_id": "G1JBI", + "video_file": "Charades/videos/G1JBI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person closes the pantry doors'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.1> to <35.6>." + } + ] + }, + { + "video_id": "G1JBI", + "question_id": "G1JBI", + "video_file": "Charades/videos/G1JBI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opened the cabinet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "0VQQL", + "question_id": "0VQQL", + "video_file": "Charades/videos/0VQQL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is laying on the floor reading a book'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.7>." + } + ] + }, + { + "video_id": "0VQQL", + "question_id": "0VQQL", + "video_file": "Charades/videos/0VQQL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person open the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.8> to <26.4>." + } + ] + }, + { + "video_id": "KF42S", + "question_id": "KF42S", + "video_file": "Charades/videos/KF42S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <16.6>." + } + ] + }, + { + "video_id": "EBHC9", + "question_id": "EBHC9", + "video_file": "Charades/videos/EBHC9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person holding a pillow walks into the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.8>." + } + ] + }, + { + "video_id": "EBHC9", + "question_id": "EBHC9", + "video_file": "Charades/videos/EBHC9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person holds a throw pillow under one arm' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.8>." + } + ] + }, + { + "video_id": "B3RHF", + "question_id": "B3RHF", + "video_file": "Charades/videos/B3RHF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <21.0>." + } + ] + }, + { + "video_id": "B3RHF", + "question_id": "B3RHF", + "video_file": "Charades/videos/B3RHF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they start eating it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.2> to <21.0>." + } + ] + }, + { + "video_id": "85ABO", + "question_id": "85ABO", + "video_file": "Charades/videos/85ABO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person throws a pillow on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <4.4>." + } + ] + }, + { + "video_id": "85ABO", + "question_id": "85ABO", + "video_file": "Charades/videos/85ABO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person throws a pillow down onto the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.3> to <4.4>." + } + ] + }, + { + "video_id": "IKOAX", + "question_id": "IKOAX", + "video_file": "Charades/videos/IKOAX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they stand up' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.0> to <27.6>." + } + ] + }, + { + "video_id": "IKOAX", + "question_id": "IKOAX", + "video_file": "Charades/videos/IKOAX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a picture with a camera that they'." + }, + { + "from": "gpt", + "value": "From <20.3> to <28.7>." + } + ] + }, + { + "video_id": "IKOAX", + "question_id": "IKOAX", + "video_file": "Charades/videos/IKOAX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put on a shelf'." + }, + { + "from": "gpt", + "value": "From <26.2> to <31.0>." + } + ] + }, + { + "video_id": "IKOAX", + "question_id": "IKOAX", + "video_file": "Charades/videos/IKOAX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts a camera on a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.2> to <31.0>." + } + ] + }, + { + "video_id": "8Z9GW", + "question_id": "8Z9GW", + "video_file": "Charades/videos/8Z9GW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person gets dressed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.0>." + } + ] + }, + { + "video_id": "8Z9GW", + "question_id": "8Z9GW", + "video_file": "Charades/videos/8Z9GW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person sits on the chair'." + }, + { + "from": "gpt", + "value": "From <11.9> to <20.8>." + } + ] + }, + { + "video_id": "8Z9GW", + "question_id": "8Z9GW", + "video_file": "Charades/videos/8Z9GW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is standing in their closet dressing for work'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.0>." + } + ] + }, + { + "video_id": "3H1EA", + "question_id": "3H1EA", + "video_file": "Charades/videos/3H1EA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person throws a book onto the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <6.7>." + } + ] + }, + { + "video_id": "3H1EA", + "question_id": "3H1EA", + "video_file": "Charades/videos/3H1EA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pouring a beverage into a glass from a carton' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.7> to <16.1>." + } + ] + }, + { + "video_id": "3H1EA", + "question_id": "3H1EA", + "video_file": "Charades/videos/3H1EA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they pour a glass of milk' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.7> to <16.1>." + } + ] + }, + { + "video_id": "3H1EA", + "question_id": "3H1EA", + "video_file": "Charades/videos/3H1EA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person reads a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "NMC5N", + "question_id": "NMC5N", + "video_file": "Charades/videos/NMC5N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a cup of coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.0>." + } + ] + }, + { + "video_id": "NMC5N", + "question_id": "NMC5N", + "video_file": "Charades/videos/NMC5N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the cup of coffee down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <16.0>." + } + ] + }, + { + "video_id": "IIFVE", + "question_id": "IIFVE", + "video_file": "Charades/videos/IIFVE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is throwing clothes on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.9> to <9.7>." + } + ] + }, + { + "video_id": "X898O", + "question_id": "X898O", + "video_file": "Charades/videos/X898O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is putting papers onto a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.8> to <12.7>." + } + ] + }, + { + "video_id": "X898O", + "question_id": "X898O", + "video_file": "Charades/videos/X898O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting them on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.8> to <24.9>." + } + ] + }, + { + "video_id": "L8Y8D", + "question_id": "L8Y8D", + "video_file": "Charades/videos/L8Y8D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.6> to <36.5>." + } + ] + }, + { + "video_id": "UVP3F", + "question_id": "UVP3F", + "video_file": "Charades/videos/UVP3F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <13.0>." + } + ] + }, + { + "video_id": "FPODN", + "question_id": "FPODN", + "video_file": "Charades/videos/FPODN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits down on the couch' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.6> to <18.1>." + } + ] + }, + { + "video_id": "5VJ1X", + "question_id": "5VJ1X", + "video_file": "Charades/videos/5VJ1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <28.4>." + } + ] + }, + { + "video_id": "5VJ1X", + "question_id": "5VJ1X", + "video_file": "Charades/videos/5VJ1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a drink from a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.8> to <28.4>." + } + ] + }, + { + "video_id": "AZMVM", + "question_id": "AZMVM", + "video_file": "Charades/videos/AZMVM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <26.0>." + } + ] + }, + { + "video_id": "SHTLT", + "question_id": "SHTLT", + "video_file": "Charades/videos/SHTLT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is closing the doorway to their kitchen'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.6> to <24.0>." + } + ] + }, + { + "video_id": "SHTLT", + "question_id": "SHTLT", + "video_file": "Charades/videos/SHTLT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.6>." + } + ] + }, + { + "video_id": "SHTLT", + "question_id": "SHTLT", + "video_file": "Charades/videos/SHTLT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.6>." + } + ] + }, + { + "video_id": "0PA3N", + "question_id": "0PA3N", + "video_file": "Charades/videos/0PA3N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is throwing shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <8.8>." + } + ] + }, + { + "video_id": "0PA3N", + "question_id": "0PA3N", + "video_file": "Charades/videos/0PA3N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits on the bed'." + }, + { + "from": "gpt", + "value": "From <8.8> to <23.7>." + } + ] + }, + { + "video_id": "0PA3N", + "question_id": "0PA3N", + "video_file": "Charades/videos/0PA3N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person throws shoes on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <8.8>." + } + ] + }, + { + "video_id": "KZ7Y8", + "question_id": "KZ7Y8", + "video_file": "Charades/videos/KZ7Y8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <11.1>." + } + ] + }, + { + "video_id": "481PN", + "question_id": "481PN", + "video_file": "Charades/videos/481PN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is putting a glass down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.7> to <12.3>." + } + ] + }, + { + "video_id": "481PN", + "question_id": "481PN", + "video_file": "Charades/videos/481PN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is putting a glass down by the phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.7> to <12.3>." + } + ] + }, + { + "video_id": "QOJ09", + "question_id": "QOJ09", + "video_file": "Charades/videos/QOJ09.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <17.9>." + } + ] + }, + { + "video_id": "QOJ09", + "question_id": "QOJ09", + "video_file": "Charades/videos/QOJ09.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they open the book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.1> to <17.9>." + } + ] + }, + { + "video_id": "QOJ09", + "question_id": "QOJ09", + "video_file": "Charades/videos/QOJ09.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person laugh as they look through it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.9> to <27.5>." + } + ] + }, + { + "video_id": "QOJ09", + "question_id": "QOJ09", + "video_file": "Charades/videos/QOJ09.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take a book from the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <16.1>." + } + ] + }, + { + "video_id": "QOJ09", + "question_id": "QOJ09", + "video_file": "Charades/videos/QOJ09.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person reads whats in the book'." + }, + { + "from": "gpt", + "value": "From <13.0> to <26.4>." + } + ] + }, + { + "video_id": "GA4EI", + "question_id": "GA4EI", + "video_file": "Charades/videos/GA4EI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting in the closet holding a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.6> to <26.9>." + } + ] + }, + { + "video_id": "GA4EI", + "question_id": "GA4EI", + "video_file": "Charades/videos/GA4EI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person the bag lights up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.4> to <35.2>." + } + ] + }, + { + "video_id": "GA4EI", + "question_id": "GA4EI", + "video_file": "Charades/videos/GA4EI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the phone in a closet'." + }, + { + "from": "gpt", + "value": "From <30.3> to <36.6>." + } + ] + }, + { + "video_id": "GA4EI", + "question_id": "GA4EI", + "video_file": "Charades/videos/GA4EI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts it up on the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.0> to <29.1>." + } + ] + }, + { + "video_id": "BLB6G", + "question_id": "BLB6G", + "video_file": "Charades/videos/BLB6G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is awakening on a couch' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <14.7>." + } + ] + }, + { + "video_id": "BLB6G", + "question_id": "BLB6G", + "video_file": "Charades/videos/BLB6G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is awakening on a sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <14.7>." + } + ] + }, + { + "video_id": "A75XI", + "question_id": "A75XI", + "video_file": "Charades/videos/A75XI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take a drink of water from a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <17.6>." + } + ] + }, + { + "video_id": "MAO8N", + "question_id": "MAO8N", + "video_file": "Charades/videos/MAO8N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is putting a towel in a box'." + }, + { + "from": "gpt", + "value": "From <25.0> to <37.0>." + } + ] + }, + { + "video_id": "MAO8N", + "question_id": "MAO8N", + "video_file": "Charades/videos/MAO8N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting it in a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.0> to <37.0>." + } + ] + }, + { + "video_id": "MAO8N", + "question_id": "MAO8N", + "video_file": "Charades/videos/MAO8N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the box away on a shelf'." + }, + { + "from": "gpt", + "value": "From <25.0> to <37.0>." + } + ] + }, + { + "video_id": "MAO8N", + "question_id": "MAO8N", + "video_file": "Charades/videos/MAO8N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting the box on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.0> to <37.0>." + } + ] + }, + { + "video_id": "DXUQS", + "question_id": "DXUQS", + "video_file": "Charades/videos/DXUQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding a broom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "DXUQS", + "question_id": "DXUQS", + "video_file": "Charades/videos/DXUQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting away groceries' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <26.9>." + } + ] + }, + { + "video_id": "DXUQS", + "question_id": "DXUQS", + "video_file": "Charades/videos/DXUQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person in the kitchen is laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.8> to <16.4>." + } + ] + }, + { + "video_id": "GOCDE", + "question_id": "GOCDE", + "video_file": "Charades/videos/GOCDE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits on the couch' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.4> to <11.0>." + } + ] + }, + { + "video_id": "GOCDE", + "question_id": "GOCDE", + "video_file": "Charades/videos/GOCDE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person pours water into a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <9.0>." + } + ] + }, + { + "video_id": "XHB9U", + "question_id": "XHB9U", + "video_file": "Charades/videos/XHB9U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is in a bedroom holding a mirror'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.1>." + } + ] + }, + { + "video_id": "SN85M", + "question_id": "SN85M", + "video_file": "Charades/videos/SN85M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws a box through a doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.6> to <30.6>." + } + ] + }, + { + "video_id": "SN85M", + "question_id": "SN85M", + "video_file": "Charades/videos/SN85M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is throwing a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.6> to <30.6>." + } + ] + }, + { + "video_id": "3YGOV", + "question_id": "3YGOV", + "video_file": "Charades/videos/3YGOV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person snuggles with a pillow'." + }, + { + "from": "gpt", + "value": "From <10.3> to <20.8>." + } + ] + }, + { + "video_id": "3YGOV", + "question_id": "3YGOV", + "video_file": "Charades/videos/3YGOV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs into a laundry room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <5.9>." + } + ] + }, + { + "video_id": "3YGOV", + "question_id": "3YGOV", + "video_file": "Charades/videos/3YGOV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs into the laundry machine'." + }, + { + "from": "gpt", + "value": "From <0.5> to <5.9>." + } + ] + }, + { + "video_id": "3YGOV", + "question_id": "3YGOV", + "video_file": "Charades/videos/3YGOV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinks from a cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.6> to <12.7>." + } + ] + }, + { + "video_id": "UA2PO", + "question_id": "UA2PO", + "video_file": "Charades/videos/UA2PO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is dressing in the bathroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.8> to <33.0>." + } + ] + }, + { + "video_id": "IV8MD", + "question_id": "IV8MD", + "video_file": "Charades/videos/IV8MD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts eating' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <14.4>." + } + ] + }, + { + "video_id": "IV8MD", + "question_id": "IV8MD", + "video_file": "Charades/videos/IV8MD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eat it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.1> to <14.4>." + } + ] + }, + { + "video_id": "IV8MD", + "question_id": "IV8MD", + "video_file": "Charades/videos/IV8MD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is laughing'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "IV8MD", + "question_id": "IV8MD", + "video_file": "Charades/videos/IV8MD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a sandwich on the stairs'." + }, + { + "from": "gpt", + "value": "From <8.1> to <14.4>." + } + ] + }, + { + "video_id": "ESTFR", + "question_id": "ESTFR", + "video_file": "Charades/videos/ESTFR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes their shoes off' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.9> to <19.1>." + } + ] + }, + { + "video_id": "ESTFR", + "question_id": "ESTFR", + "video_file": "Charades/videos/ESTFR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes off their shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.9> to <19.1>." + } + ] + }, + { + "video_id": "5QXCO", + "question_id": "5QXCO", + "video_file": "Charades/videos/5QXCO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts the clothes in a washing machine'." + }, + { + "from": "gpt", + "value": "From <13.6> to <22.3>." + } + ] + }, + { + "video_id": "5QXCO", + "question_id": "5QXCO", + "video_file": "Charades/videos/5QXCO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person after throwing clothes in washer' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.6> to <22.3>." + } + ] + }, + { + "video_id": "SVPZT", + "question_id": "SVPZT", + "video_file": "Charades/videos/SVPZT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "78SO8", + "question_id": "78SO8", + "video_file": "Charades/videos/78SO8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pour some into a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.1> to <11.9>." + } + ] + }, + { + "video_id": "78SO8", + "question_id": "78SO8", + "video_file": "Charades/videos/78SO8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pouring said juice into a glass'." + }, + { + "from": "gpt", + "value": "From <6.1> to <11.9>." + } + ] + }, + { + "video_id": "78SO8", + "question_id": "78SO8", + "video_file": "Charades/videos/78SO8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the door to their pantry'." + }, + { + "from": "gpt", + "value": "From <1.4> to <10.1>." + } + ] + }, + { + "video_id": "80FWF", + "question_id": "80FWF", + "video_file": "Charades/videos/80FWF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes out a towel' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <12.4>." + } + ] + }, + { + "video_id": "80FWF", + "question_id": "80FWF", + "video_file": "Charades/videos/80FWF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <5.6>." + } + ] + }, + { + "video_id": "80FWF", + "question_id": "80FWF", + "video_file": "Charades/videos/80FWF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they turn on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <5.6>." + } + ] + }, + { + "video_id": "5FHH2", + "question_id": "5FHH2", + "video_file": "Charades/videos/5FHH2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person throws the stack of clothes onto a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.3> to <26.0>." + } + ] + }, + { + "video_id": "5F1AW", + "question_id": "5F1AW", + "video_file": "Charades/videos/5F1AW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person puts a book down onto the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "5F1AW", + "question_id": "5F1AW", + "video_file": "Charades/videos/5F1AW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts a book on the shelf'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "EHHRS", + "question_id": "EHHRS", + "video_file": "Charades/videos/EHHRS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person turns on a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.6> to <6.8>." + } + ] + }, + { + "video_id": "EHHRS", + "question_id": "EHHRS", + "video_file": "Charades/videos/EHHRS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes some medicine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.2> to <21.7>." + } + ] + }, + { + "video_id": "EHHRS", + "question_id": "EHHRS", + "video_file": "Charades/videos/EHHRS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they turn on the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.6> to <6.8>." + } + ] + }, + { + "video_id": "386XM", + "question_id": "386XM", + "video_file": "Charades/videos/386XM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "386XM", + "question_id": "386XM", + "video_file": "Charades/videos/386XM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person watching the evening news on television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.5> to <13.6>." + } + ] + }, + { + "video_id": "386XM", + "question_id": "386XM", + "video_file": "Charades/videos/386XM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person resumes watching television'." + }, + { + "from": "gpt", + "value": "From <2.5> to <13.6>." + } + ] + }, + { + "video_id": "U4W3C", + "question_id": "U4W3C", + "video_file": "Charades/videos/U4W3C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person he puts a book down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "U4W3C", + "question_id": "U4W3C", + "video_file": "Charades/videos/U4W3C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person puts down a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "U4W3C", + "question_id": "U4W3C", + "video_file": "Charades/videos/U4W3C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits down to open their book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <29.0>." + } + ] + }, + { + "video_id": "IKYVB", + "question_id": "IKYVB", + "video_file": "Charades/videos/IKYVB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person put away a pillow in the laundry room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.2>." + } + ] + }, + { + "video_id": "G4MNO", + "question_id": "G4MNO", + "video_file": "Charades/videos/G4MNO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is laughing at a television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "G4MNO", + "question_id": "G4MNO", + "video_file": "Charades/videos/G4MNO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.4> to <17.0>." + } + ] + }, + { + "video_id": "G4MNO", + "question_id": "G4MNO", + "video_file": "Charades/videos/G4MNO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.3> to <27.0>." + } + ] + }, + { + "video_id": "G4MNO", + "question_id": "G4MNO", + "video_file": "Charades/videos/G4MNO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes a book off the desk'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.8> to <13.5>." + } + ] + }, + { + "video_id": "WHRBU", + "question_id": "WHRBU", + "video_file": "Charades/videos/WHRBU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they stand up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.5> to <26.8>." + } + ] + }, + { + "video_id": "QU2WL", + "question_id": "QU2WL", + "video_file": "Charades/videos/QU2WL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they are eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <19.1>." + } + ] + }, + { + "video_id": "QU2WL", + "question_id": "QU2WL", + "video_file": "Charades/videos/QU2WL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts eating'." + }, + { + "from": "gpt", + "value": "From <12.0> to <19.1>." + } + ] + }, + { + "video_id": "F2EA8", + "question_id": "F2EA8", + "video_file": "Charades/videos/F2EA8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person both people drink from their cups' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.0> to <19.9>." + } + ] + }, + { + "video_id": "93YXJ", + "question_id": "93YXJ", + "video_file": "Charades/videos/93YXJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes some food from a bag'." + }, + { + "from": "gpt", + "value": "From <5.3> to <14.4>." + } + ] + }, + { + "video_id": "93YXJ", + "question_id": "93YXJ", + "video_file": "Charades/videos/93YXJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes out some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.3> to <14.4>." + } + ] + }, + { + "video_id": "93YXJ", + "question_id": "93YXJ", + "video_file": "Charades/videos/93YXJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <39.0> to <44.0>." + } + ] + }, + { + "video_id": "93YXJ", + "question_id": "93YXJ", + "video_file": "Charades/videos/93YXJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a bag of groceries'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <7.8>." + } + ] + }, + { + "video_id": "JJLTF", + "question_id": "JJLTF", + "video_file": "Charades/videos/JJLTF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person start sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <15.4>." + } + ] + }, + { + "video_id": "JJLTF", + "question_id": "JJLTF", + "video_file": "Charades/videos/JJLTF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is drinking from a cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.3> to <12.0>." + } + ] + }, + { + "video_id": "QEY8N", + "question_id": "QEY8N", + "video_file": "Charades/videos/QEY8N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs into the room holding a broom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.4> to <8.2>." + } + ] + }, + { + "video_id": "QEY8N", + "question_id": "QEY8N", + "video_file": "Charades/videos/QEY8N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the broom against the wall'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <10.6>." + } + ] + }, + { + "video_id": "QEY8N", + "question_id": "QEY8N", + "video_file": "Charades/videos/QEY8N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits in a chair eating the sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.9> to <24.4>." + } + ] + }, + { + "video_id": "QEY8N", + "question_id": "QEY8N", + "video_file": "Charades/videos/QEY8N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sitting in a chair eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.9> to <24.4>." + } + ] + }, + { + "video_id": "QEY8N", + "question_id": "QEY8N", + "video_file": "Charades/videos/QEY8N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walks through a doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.6> to <10.3>." + } + ] + }, + { + "video_id": "72JYK", + "question_id": "72JYK", + "video_file": "Charades/videos/72JYK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.7> to <12.0>." + } + ] + }, + { + "video_id": "QS6O1", + "question_id": "QS6O1", + "video_file": "Charades/videos/QS6O1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person pours liquid into a cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.6> to <11.0>." + } + ] + }, + { + "video_id": "QS6O1", + "question_id": "QS6O1", + "video_file": "Charades/videos/QS6O1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts food in a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <24.8>." + } + ] + }, + { + "video_id": "QS6O1", + "question_id": "QS6O1", + "video_file": "Charades/videos/QS6O1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts food from the pantry into a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <24.8>." + } + ] + }, + { + "video_id": "GV9M4", + "question_id": "GV9M4", + "video_file": "Charades/videos/GV9M4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a drink from a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <17.0>." + } + ] + }, + { + "video_id": "GV9M4", + "question_id": "GV9M4", + "video_file": "Charades/videos/GV9M4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <10.9>." + } + ] + }, + { + "video_id": "GV9M4", + "question_id": "GV9M4", + "video_file": "Charades/videos/GV9M4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.4> to <17.0>." + } + ] + }, + { + "video_id": "GV9M4", + "question_id": "GV9M4", + "video_file": "Charades/videos/GV9M4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looking out a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.4> to <26.1>." + } + ] + }, + { + "video_id": "GV9M4", + "question_id": "GV9M4", + "video_file": "Charades/videos/GV9M4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take a drink from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <17.0>." + } + ] + }, + { + "video_id": "GV9M4", + "question_id": "GV9M4", + "video_file": "Charades/videos/GV9M4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens the closet door'." + }, + { + "from": "gpt", + "value": "From <1.3> to <10.9>." + } + ] + }, + { + "video_id": "4O32W", + "question_id": "4O32W", + "video_file": "Charades/videos/4O32W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking a glass of water with the other' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "ENJ7V", + "question_id": "ENJ7V", + "video_file": "Charades/videos/ENJ7V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes off their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.9> to <34.9>." + } + ] + }, + { + "video_id": "78XZ9", + "question_id": "78XZ9", + "video_file": "Charades/videos/78XZ9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person smiling at a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.7> to <10.0>." + } + ] + }, + { + "video_id": "78XZ9", + "question_id": "78XZ9", + "video_file": "Charades/videos/78XZ9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they open the book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.3> to <20.1>." + } + ] + }, + { + "video_id": "N2TPE", + "question_id": "N2TPE", + "video_file": "Charades/videos/N2TPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the clothes in'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <29.2>." + } + ] + }, + { + "video_id": "N2TPE", + "question_id": "N2TPE", + "video_file": "Charades/videos/N2TPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes out clothes opens washing machine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.2> to <18.7>." + } + ] + }, + { + "video_id": "N2TPE", + "question_id": "N2TPE", + "video_file": "Charades/videos/N2TPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts laptop on table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <8.0>." + } + ] + }, + { + "video_id": "N2TPE", + "question_id": "N2TPE", + "video_file": "Charades/videos/N2TPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts their laptop on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <8.0>." + } + ] + }, + { + "video_id": "N2TPE", + "question_id": "N2TPE", + "video_file": "Charades/videos/N2TPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.8> to <12.7>." + } + ] + }, + { + "video_id": "HJ5TL", + "question_id": "HJ5TL", + "video_file": "Charades/videos/HJ5TL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is putting dishes away' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.6>." + } + ] + }, + { + "video_id": "6QH7V", + "question_id": "6QH7V", + "video_file": "Charades/videos/6QH7V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throw it on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.8> to <26.6>." + } + ] + }, + { + "video_id": "NWFOF", + "question_id": "NWFOF", + "video_file": "Charades/videos/NWFOF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person looking out the window for a few seconds' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.1> to <20.9>." + } + ] + }, + { + "video_id": "NWFOF", + "question_id": "NWFOF", + "video_file": "Charades/videos/NWFOF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns to look out a window in a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.1> to <20.9>." + } + ] + }, + { + "video_id": "NWFOF", + "question_id": "NWFOF", + "video_file": "Charades/videos/NWFOF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is seen smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "FPWD0", + "question_id": "FPWD0", + "video_file": "Charades/videos/FPWD0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.4> to <13.8>." + } + ] + }, + { + "video_id": "U36AB", + "question_id": "U36AB", + "video_file": "Charades/videos/U36AB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinks from cup'." + }, + { + "from": "gpt", + "value": "From <15.4> to <30.0>." + } + ] + }, + { + "video_id": "U36AB", + "question_id": "U36AB", + "video_file": "Charades/videos/U36AB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts cup on floor'." + }, + { + "from": "gpt", + "value": "From <29.4> to <33.0>." + } + ] + }, + { + "video_id": "3ND23", + "question_id": "3ND23", + "video_file": "Charades/videos/3ND23.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person snuggling under a blanket to watch television' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.0> to <31.0>." + } + ] + }, + { + "video_id": "3ND23", + "question_id": "3ND23", + "video_file": "Charades/videos/3ND23.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put a blanket around them' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <19.9>." + } + ] + }, + { + "video_id": "3ND23", + "question_id": "3ND23", + "video_file": "Charades/videos/3ND23.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they sit on a sofa' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <13.9>." + } + ] + }, + { + "video_id": "L36YT", + "question_id": "L36YT", + "video_file": "Charades/videos/L36YT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits on the sofa' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.8> to <22.0>." + } + ] + }, + { + "video_id": "HBQ33", + "question_id": "HBQ33", + "video_file": "Charades/videos/HBQ33.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a sneezing person takes medicine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.9>." + } + ] + }, + { + "video_id": "8G9A9", + "question_id": "8G9A9", + "video_file": "Charades/videos/8G9A9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person smiles as the put a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.1> to <13.4>." + } + ] + }, + { + "video_id": "TUPTT", + "question_id": "TUPTT", + "video_file": "Charades/videos/TUPTT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits down to put their shoes on'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.3> to <34.0>." + } + ] + }, + { + "video_id": "TUPTT", + "question_id": "TUPTT", + "video_file": "Charades/videos/TUPTT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person sits on a couch to put shoes on'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.3> to <34.0>." + } + ] + }, + { + "video_id": "6CGRP", + "question_id": "6CGRP", + "video_file": "Charades/videos/6CGRP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person lays on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.1> to <32.0>." + } + ] + }, + { + "video_id": "2DDGF", + "question_id": "2DDGF", + "video_file": "Charades/videos/2DDGF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts a laptop on a shelf'." + }, + { + "from": "gpt", + "value": "From <18.2> to <25.9>." + } + ] + }, + { + "video_id": "2DDGF", + "question_id": "2DDGF", + "video_file": "Charades/videos/2DDGF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a closet door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <21.9>." + } + ] + }, + { + "video_id": "2DDGF", + "question_id": "2DDGF", + "video_file": "Charades/videos/2DDGF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts their laptop on a shelf'." + }, + { + "from": "gpt", + "value": "From <18.2> to <25.9>." + } + ] + }, + { + "video_id": "2DDGF", + "question_id": "2DDGF", + "video_file": "Charades/videos/2DDGF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is drinking water from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "YQ9ML", + "question_id": "YQ9ML", + "video_file": "Charades/videos/YQ9ML.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person fixes their hair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <17.3>." + } + ] + }, + { + "video_id": "W0Y4E", + "question_id": "W0Y4E", + "video_file": "Charades/videos/W0Y4E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is throwing a blanket on the bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <8.3>." + } + ] + }, + { + "video_id": "W0Y4E", + "question_id": "W0Y4E", + "video_file": "Charades/videos/W0Y4E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.7> to <19.4>." + } + ] + }, + { + "video_id": "W0Y4E", + "question_id": "W0Y4E", + "video_file": "Charades/videos/W0Y4E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person jumps on it to eat a sandwich'." + }, + { + "from": "gpt", + "value": "From <8.7> to <19.4>." + } + ] + }, + { + "video_id": "268ET", + "question_id": "268ET", + "video_file": "Charades/videos/268ET.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a bag'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "5KOTU", + "question_id": "5KOTU", + "video_file": "Charades/videos/5KOTU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is in a hallway they open a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.1> to <8.3>." + } + ] + }, + { + "video_id": "0XB7U", + "question_id": "0XB7U", + "video_file": "Charades/videos/0XB7U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is looking through a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "0XB7U", + "question_id": "0XB7U", + "video_file": "Charades/videos/0XB7U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person washes their hands' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <31.0>." + } + ] + }, + { + "video_id": "0XB7U", + "question_id": "0XB7U", + "video_file": "Charades/videos/0XB7U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is reading a book'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "0XB7U", + "question_id": "0XB7U", + "video_file": "Charades/videos/0XB7U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sneezing into it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.8> to <17.5>." + } + ] + }, + { + "video_id": "TIPUS", + "question_id": "TIPUS", + "video_file": "Charades/videos/TIPUS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person watches something visible in the mirror' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <18.0>." + } + ] + }, + { + "video_id": "PXBN0", + "question_id": "PXBN0", + "video_file": "Charades/videos/PXBN0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is eating'." + }, + { + "from": "gpt", + "value": "From <24.6> to <30.0>." + } + ] + }, + { + "video_id": "PXBN0", + "question_id": "PXBN0", + "video_file": "Charades/videos/PXBN0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they start to eat the snack' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.6> to <30.0>." + } + ] + }, + { + "video_id": "XBXKO", + "question_id": "XBXKO", + "video_file": "Charades/videos/XBXKO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes some clothes out of the dryer' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.3> to <36.7>." + } + ] + }, + { + "video_id": "P981O", + "question_id": "P981O", + "video_file": "Charades/videos/P981O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person closes the door behind the person' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.3> to <31.1>." + } + ] + }, + { + "video_id": "DP7DG", + "question_id": "DP7DG", + "video_file": "Charades/videos/DP7DG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they are holding open a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "DP7DG", + "question_id": "DP7DG", + "video_file": "Charades/videos/DP7DG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is holding a box of dishes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <7.7>." + } + ] + }, + { + "video_id": "V45U8", + "question_id": "V45U8", + "video_file": "Charades/videos/V45U8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person fixing a desk drawer that holds shoes'." + }, + { + "from": "gpt", + "value": "From <28.4> to <33.0>." + } + ] + }, + { + "video_id": "QG338", + "question_id": "QG338", + "video_file": "Charades/videos/QG338.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.8> to <23.4>." + } + ] + }, + { + "video_id": "QG338", + "question_id": "QG338", + "video_file": "Charades/videos/QG338.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "QG338", + "question_id": "QG338", + "video_file": "Charades/videos/QG338.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looks at a picture'." + }, + { + "from": "gpt", + "value": "From <4.9> to <14.4>." + } + ] + }, + { + "video_id": "QG338", + "question_id": "QG338", + "video_file": "Charades/videos/QG338.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person open the door came inside'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "T8DK4", + "question_id": "T8DK4", + "video_file": "Charades/videos/T8DK4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person fixing hair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <23.3>." + } + ] + }, + { + "video_id": "T8DK4", + "question_id": "T8DK4", + "video_file": "Charades/videos/T8DK4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting on the floor smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "T8DK4", + "question_id": "T8DK4", + "video_file": "Charades/videos/T8DK4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person laughs'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "T8DK4", + "question_id": "T8DK4", + "video_file": "Charades/videos/T8DK4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the mirror on the shelf'." + }, + { + "from": "gpt", + "value": "From <15.7> to <26.7>." + } + ] + }, + { + "video_id": "T8DK4", + "question_id": "T8DK4", + "video_file": "Charades/videos/T8DK4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting it on the shelf'." + }, + { + "from": "gpt", + "value": "From <15.7> to <26.7>." + } + ] + }, + { + "video_id": "BWLA8", + "question_id": "BWLA8", + "video_file": "Charades/videos/BWLA8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person looks through the book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.9> to <30.4>." + } + ] + }, + { + "video_id": "NSHA7", + "question_id": "NSHA7", + "video_file": "Charades/videos/NSHA7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take a book off the bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <20.2>." + } + ] + }, + { + "video_id": "7K9XM", + "question_id": "7K9XM", + "video_file": "Charades/videos/7K9XM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person washing a mirror in a pantry' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <51.0> to <108.7>." + } + ] + }, + { + "video_id": "AJZYE", + "question_id": "AJZYE", + "video_file": "Charades/videos/AJZYE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a closet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.3> to <10.3>." + } + ] + }, + { + "video_id": "AJZYE", + "question_id": "AJZYE", + "video_file": "Charades/videos/AJZYE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they open the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <10.3>." + } + ] + }, + { + "video_id": "9RX96", + "question_id": "9RX96", + "video_file": "Charades/videos/9RX96.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looks out the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.6> to <37.5>." + } + ] + }, + { + "video_id": "1PB6T", + "question_id": "1PB6T", + "video_file": "Charades/videos/1PB6T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes a pen paper'." + }, + { + "from": "gpt", + "value": "From <8.4> to <17.3>." + } + ] + }, + { + "video_id": "1PB6T", + "question_id": "1PB6T", + "video_file": "Charades/videos/1PB6T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a box of food'." + }, + { + "from": "gpt", + "value": "From <1.1> to <9.1>." + } + ] + }, + { + "video_id": "LEPM3", + "question_id": "LEPM3", + "video_file": "Charades/videos/LEPM3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is opening a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <8.3>." + } + ] + }, + { + "video_id": "EL3N9", + "question_id": "EL3N9", + "video_file": "Charades/videos/EL3N9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <9.4>." + } + ] + }, + { + "video_id": "EL3N9", + "question_id": "EL3N9", + "video_file": "Charades/videos/EL3N9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks in holding a bag of groceries' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <19.9>." + } + ] + }, + { + "video_id": "EL3N9", + "question_id": "EL3N9", + "video_file": "Charades/videos/EL3N9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put the groceries away in the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.0> to <35.2>." + } + ] + }, + { + "video_id": "XKJAD", + "question_id": "XKJAD", + "video_file": "Charades/videos/XKJAD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits at the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.8> to <17.5>." + } + ] + }, + { + "video_id": "S1J2Q", + "question_id": "S1J2Q", + "video_file": "Charades/videos/S1J2Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person is running to the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <7.5>." + } + ] + }, + { + "video_id": "S1J2Q", + "question_id": "S1J2Q", + "video_file": "Charades/videos/S1J2Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.8> to <20.8>." + } + ] + }, + { + "video_id": "S1J2Q", + "question_id": "S1J2Q", + "video_file": "Charades/videos/S1J2Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs to another room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <7.5>." + } + ] + }, + { + "video_id": "NY0IW", + "question_id": "NY0IW", + "video_file": "Charades/videos/NY0IW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put some clothes on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.7> to <15.8>." + } + ] + }, + { + "video_id": "NY0IW", + "question_id": "NY0IW", + "video_file": "Charades/videos/NY0IW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing in the laundry room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.3> to <12.7>." + } + ] + }, + { + "video_id": "00YZL", + "question_id": "00YZL", + "video_file": "Charades/videos/00YZL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.4> to <34.0>." + } + ] + }, + { + "video_id": "L8HMR", + "question_id": "L8HMR", + "video_file": "Charades/videos/L8HMR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts a blanket into the washing machine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <14.0>." + } + ] + }, + { + "video_id": "U5CDS", + "question_id": "U5CDS", + "video_file": "Charades/videos/U5CDS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person sits in a chair'." + }, + { + "from": "gpt", + "value": "From <5.1> to <10.9>." + } + ] + }, + { + "video_id": "ZO3EG", + "question_id": "ZO3EG", + "video_file": "Charades/videos/ZO3EG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put some books from a shelf into it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.9> to <31.5>." + } + ] + }, + { + "video_id": "ZO3EG", + "question_id": "ZO3EG", + "video_file": "Charades/videos/ZO3EG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open the door'." + }, + { + "from": "gpt", + "value": "From <31.9> to <38.6>." + } + ] + }, + { + "video_id": "ZO3EG", + "question_id": "ZO3EG", + "video_file": "Charades/videos/ZO3EG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they hold a bag open'." + }, + { + "from": "gpt", + "value": "From <12.1> to <18.8>." + } + ] + }, + { + "video_id": "ZO3EG", + "question_id": "ZO3EG", + "video_file": "Charades/videos/ZO3EG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person some other clothes in a washer' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.6> to <11.5>." + } + ] + }, + { + "video_id": "ZO3EG", + "question_id": "ZO3EG", + "video_file": "Charades/videos/ZO3EG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they open a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <18.8>." + } + ] + }, + { + "video_id": "ZO3EG", + "question_id": "ZO3EG", + "video_file": "Charades/videos/ZO3EG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person undresses in the laundry room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <9.4>." + } + ] + }, + { + "video_id": "ZO3EG", + "question_id": "ZO3EG", + "video_file": "Charades/videos/ZO3EG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws the clothes into the open laundry machine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <13.4>." + } + ] + }, + { + "video_id": "X5E6E", + "question_id": "X5E6E", + "video_file": "Charades/videos/X5E6E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is drinking a cup of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <11.5>." + } + ] + }, + { + "video_id": "5ACD3", + "question_id": "5ACD3", + "video_file": "Charades/videos/5ACD3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person as they stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <41.0> to <46.0>." + } + ] + }, + { + "video_id": "5ACD3", + "question_id": "5ACD3", + "video_file": "Charades/videos/5ACD3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put it on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <41.0> to <46.0>." + } + ] + }, + { + "video_id": "U91SO", + "question_id": "U91SO", + "video_file": "Charades/videos/U91SO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person start closing a nearby cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.5> to <32.0>." + } + ] + }, + { + "video_id": "U91SO", + "question_id": "U91SO", + "video_file": "Charades/videos/U91SO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the cabinet doors' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.5> to <32.0>." + } + ] + }, + { + "video_id": "DP5S8", + "question_id": "DP5S8", + "video_file": "Charades/videos/DP5S8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person is undressing'." + }, + { + "from": "gpt", + "value": "From <4.7> to <13.0>." + } + ] + }, + { + "video_id": "W4BZX", + "question_id": "W4BZX", + "video_file": "Charades/videos/W4BZX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening up a laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.4> to <15.4>." + } + ] + }, + { + "video_id": "W4BZX", + "question_id": "W4BZX", + "video_file": "Charades/videos/W4BZX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens up the laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.4> to <15.4>." + } + ] + }, + { + "video_id": "SXDG5", + "question_id": "SXDG5", + "video_file": "Charades/videos/SXDG5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person turns on a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.0> to <14.2>." + } + ] + }, + { + "video_id": "SXDG5", + "question_id": "SXDG5", + "video_file": "Charades/videos/SXDG5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person walks into a room turns on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <14.2>." + } + ] + }, + { + "video_id": "SXDG5", + "question_id": "SXDG5", + "video_file": "Charades/videos/SXDG5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a smiling person turns on the light in their garage' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.0> to <14.2>." + } + ] + }, + { + "video_id": "DQFF2", + "question_id": "DQFF2", + "video_file": "Charades/videos/DQFF2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person he opens his window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.1> to <13.7>." + } + ] + }, + { + "video_id": "DQFF2", + "question_id": "DQFF2", + "video_file": "Charades/videos/DQFF2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is holding a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.6> to <19.3>." + } + ] + }, + { + "video_id": "DQFF2", + "question_id": "DQFF2", + "video_file": "Charades/videos/DQFF2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opens a window'." + }, + { + "from": "gpt", + "value": "From <7.1> to <13.7>." + } + ] + }, + { + "video_id": "K3IGM", + "question_id": "K3IGM", + "video_file": "Charades/videos/K3IGM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins sneezing'." + }, + { + "from": "gpt", + "value": "From <34.9> to <40.0>." + } + ] + }, + { + "video_id": "K3IGM", + "question_id": "K3IGM", + "video_file": "Charades/videos/K3IGM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts on a pair of shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.6> to <37.9>." + } + ] + }, + { + "video_id": "K3IGM", + "question_id": "K3IGM", + "video_file": "Charades/videos/K3IGM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is holding a towel' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "K3IGM", + "question_id": "K3IGM", + "video_file": "Charades/videos/K3IGM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts dressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.1> to <21.6>." + } + ] + }, + { + "video_id": "K3IGM", + "question_id": "K3IGM", + "video_file": "Charades/videos/K3IGM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting on shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.6> to <37.9>." + } + ] + }, + { + "video_id": "K3IGM", + "question_id": "K3IGM", + "video_file": "Charades/videos/K3IGM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the towel down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "VTOYP", + "question_id": "VTOYP", + "video_file": "Charades/videos/VTOYP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person start laughing'." + }, + { + "from": "gpt", + "value": "From <14.1> to <24.3>." + } + ] + }, + { + "video_id": "B7SZJ", + "question_id": "B7SZJ", + "video_file": "Charades/videos/B7SZJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the clothes in' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.9> to <17.2>." + } + ] + }, + { + "video_id": "B7SZJ", + "question_id": "B7SZJ", + "video_file": "Charades/videos/B7SZJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person walks out a door holding clothes'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.3>." + } + ] + }, + { + "video_id": "B7SZJ", + "question_id": "B7SZJ", + "video_file": "Charades/videos/B7SZJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.5> to <7.5>." + } + ] + }, + { + "video_id": "B7SZJ", + "question_id": "B7SZJ", + "video_file": "Charades/videos/B7SZJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closing the door behind' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <7.5>." + } + ] + }, + { + "video_id": "B7SZJ", + "question_id": "B7SZJ", + "video_file": "Charades/videos/B7SZJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <7.5>." + } + ] + }, + { + "video_id": "B7SZJ", + "question_id": "B7SZJ", + "video_file": "Charades/videos/B7SZJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person opens the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <10.7>." + } + ] + }, + { + "video_id": "V54TI", + "question_id": "V54TI", + "video_file": "Charades/videos/V54TI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person stand up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <21.9>." + } + ] + }, + { + "video_id": "V54TI", + "question_id": "V54TI", + "video_file": "Charades/videos/V54TI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.0> to <18.2>." + } + ] + }, + { + "video_id": "V25D1", + "question_id": "V25D1", + "video_file": "Charades/videos/V25D1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is undressing in front of a mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "V25D1", + "question_id": "V25D1", + "video_file": "Charades/videos/V25D1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin to eat the sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.6> to <14.4>." + } + ] + }, + { + "video_id": "V25D1", + "question_id": "V25D1", + "video_file": "Charades/videos/V25D1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.6> to <14.4>." + } + ] + }, + { + "video_id": "V25D1", + "question_id": "V25D1", + "video_file": "Charades/videos/V25D1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person undresses in front of a mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "98NYF", + "question_id": "98NYF", + "video_file": "Charades/videos/98NYF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person run up the stairs' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.6> to <21.0>." + } + ] + }, + { + "video_id": "98NYF", + "question_id": "98NYF", + "video_file": "Charades/videos/98NYF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person run out of the room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.6> to <21.0>." + } + ] + }, + { + "video_id": "IJ3QB", + "question_id": "IJ3QB", + "video_file": "Charades/videos/IJ3QB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits down on sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "IJ3QB", + "question_id": "IJ3QB", + "video_file": "Charades/videos/IJ3QB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.2> to <19.2>." + } + ] + }, + { + "video_id": "IJ3QB", + "question_id": "IJ3QB", + "video_file": "Charades/videos/IJ3QB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sitting on a couch opening medicine' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "IM3I0", + "question_id": "IM3I0", + "video_file": "Charades/videos/IM3I0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits down on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.7> to <32.0>." + } + ] + }, + { + "video_id": "IM3I0", + "question_id": "IM3I0", + "video_file": "Charades/videos/IM3I0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they sit on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <32.0>." + } + ] + }, + { + "video_id": "LDL8D", + "question_id": "LDL8D", + "video_file": "Charades/videos/LDL8D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is laughing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "LDL8D", + "question_id": "LDL8D", + "video_file": "Charades/videos/LDL8D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sitting at dresser looking at her phone laughing'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "ZOL0R", + "question_id": "ZOL0R", + "video_file": "Charades/videos/ZOL0R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting clothes on floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <20.5>." + } + ] + }, + { + "video_id": "ZOL0R", + "question_id": "ZOL0R", + "video_file": "Charades/videos/ZOL0R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts opening door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <10.4>." + } + ] + }, + { + "video_id": "ZOL0R", + "question_id": "ZOL0R", + "video_file": "Charades/videos/ZOL0R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opening a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.7> to <10.4>." + } + ] + }, + { + "video_id": "LRQXG", + "question_id": "LRQXG", + "video_file": "Charades/videos/LRQXG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is putting clothes on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.8> to <13.1>." + } + ] + }, + { + "video_id": "LRQXG", + "question_id": "LRQXG", + "video_file": "Charades/videos/LRQXG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is throwing some clothes into a pile' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.2>." + } + ] + }, + { + "video_id": "ZED9E", + "question_id": "ZED9E", + "video_file": "Charades/videos/ZED9E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is in the living room eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.1>." + } + ] + }, + { + "video_id": "ZED9E", + "question_id": "ZED9E", + "video_file": "Charades/videos/ZED9E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person eating a sandwich crouches next to a vacuum' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.1>." + } + ] + }, + { + "video_id": "ZED9E", + "question_id": "ZED9E", + "video_file": "Charades/videos/ZED9E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person fixes a vacuum cleaner' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <9.4>." + } + ] + }, + { + "video_id": "ZED9E", + "question_id": "ZED9E", + "video_file": "Charades/videos/ZED9E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person fixing the vacuum' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <9.4>." + } + ] + }, + { + "video_id": "JQLD3", + "question_id": "JQLD3", + "video_file": "Charades/videos/JQLD3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eat some' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.1> to <14.4>." + } + ] + }, + { + "video_id": "JQLD3", + "question_id": "JQLD3", + "video_file": "Charades/videos/JQLD3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats some food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.1> to <14.4>." + } + ] + }, + { + "video_id": "JQLD3", + "question_id": "JQLD3", + "video_file": "Charades/videos/JQLD3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is holding food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <11.9>." + } + ] + }, + { + "video_id": "3MAY9", + "question_id": "3MAY9", + "video_file": "Charades/videos/3MAY9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinks a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <19.6>." + } + ] + }, + { + "video_id": "2D5GN", + "question_id": "2D5GN", + "video_file": "Charades/videos/2D5GN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins tidying a shelf'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "XZM6Y", + "question_id": "XZM6Y", + "video_file": "Charades/videos/XZM6Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person runs out' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.5> to <18.0>." + } + ] + }, + { + "video_id": "XZM6Y", + "question_id": "XZM6Y", + "video_file": "Charades/videos/XZM6Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person with a sandwich closes a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <13.8>." + } + ] + }, + { + "video_id": "XZM6Y", + "question_id": "XZM6Y", + "video_file": "Charades/videos/XZM6Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person runs out of the room'." + }, + { + "from": "gpt", + "value": "From <12.5> to <18.0>." + } + ] + }, + { + "video_id": "3ZUVI", + "question_id": "3ZUVI", + "video_file": "Charades/videos/3ZUVI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens bag'." + }, + { + "from": "gpt", + "value": "From <1.0> to <6.8>." + } + ] + }, + { + "video_id": "3ZUVI", + "question_id": "3ZUVI", + "video_file": "Charades/videos/3ZUVI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes out a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.4>." + } + ] + }, + { + "video_id": "3ZUVI", + "question_id": "3ZUVI", + "video_file": "Charades/videos/3ZUVI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.9> to <13.2>." + } + ] + }, + { + "video_id": "3ZUVI", + "question_id": "3ZUVI", + "video_file": "Charades/videos/3ZUVI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.2> to <11.5>." + } + ] + }, + { + "video_id": "3ZUVI", + "question_id": "3ZUVI", + "video_file": "Charades/videos/3ZUVI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person throws a bag containing a book down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <7.6>." + } + ] + }, + { + "video_id": "6N4BD", + "question_id": "6N4BD", + "video_file": "Charades/videos/6N4BD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is awakening from laying down on the floor'." + }, + { + "from": "gpt", + "value": "From <1.7> to <18.0>." + } + ] + }, + { + "video_id": "6N4BD", + "question_id": "6N4BD", + "video_file": "Charades/videos/6N4BD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begin eating some unhealthy snack food from it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.3> to <34.0>." + } + ] + }, + { + "video_id": "6N4BD", + "question_id": "6N4BD", + "video_file": "Charades/videos/6N4BD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eat from a grocery bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.3> to <34.0>." + } + ] + }, + { + "video_id": "6N4BD", + "question_id": "6N4BD", + "video_file": "Charades/videos/6N4BD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats something'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.3> to <34.0>." + } + ] + }, + { + "video_id": "6N4BD", + "question_id": "6N4BD", + "video_file": "Charades/videos/6N4BD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they open the bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <25.1>." + } + ] + }, + { + "video_id": "YVK8M", + "question_id": "YVK8M", + "video_file": "Charades/videos/YVK8M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the other person is leaving the refrigerator open' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "YVK8M", + "question_id": "YVK8M", + "video_file": "Charades/videos/YVK8M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening the refrigerator taking out some products' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "F5C5M", + "question_id": "F5C5M", + "video_file": "Charades/videos/F5C5M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a picture of it'." + }, + { + "from": "gpt", + "value": "From <22.9> to <32.0>." + } + ] + }, + { + "video_id": "F5C5M", + "question_id": "F5C5M", + "video_file": "Charades/videos/F5C5M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes out a camera' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.3> to <23.7>." + } + ] + }, + { + "video_id": "F5C5M", + "question_id": "F5C5M", + "video_file": "Charades/videos/F5C5M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looks through some pictures'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.9> to <32.0>." + } + ] + }, + { + "video_id": "LIP4U", + "question_id": "LIP4U", + "video_file": "Charades/videos/LIP4U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person undressed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "2RJF6", + "question_id": "2RJF6", + "video_file": "Charades/videos/2RJF6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person fixing hair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.0> to <29.6>." + } + ] + }, + { + "video_id": "T2EIY", + "question_id": "T2EIY", + "video_file": "Charades/videos/T2EIY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they are drinking a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.4> to <23.9>." + } + ] + }, + { + "video_id": "T2EIY", + "question_id": "T2EIY", + "video_file": "Charades/videos/T2EIY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.4> to <23.9>." + } + ] + }, + { + "video_id": "18CTK", + "question_id": "18CTK", + "video_file": "Charades/videos/18CTK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person comes running through a door holding a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.6> to <7.8>." + } + ] + }, + { + "video_id": "B3HCE", + "question_id": "B3HCE", + "video_file": "Charades/videos/B3HCE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put a bag on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <12.1>." + } + ] + }, + { + "video_id": "B3HCE", + "question_id": "B3HCE", + "video_file": "Charades/videos/B3HCE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <8.4>." + } + ] + }, + { + "video_id": "B3HCE", + "question_id": "B3HCE", + "video_file": "Charades/videos/B3HCE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <8.4>." + } + ] + }, + { + "video_id": "B3HCE", + "question_id": "B3HCE", + "video_file": "Charades/videos/B3HCE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.4> to <17.7>." + } + ] + }, + { + "video_id": "B3HCE", + "question_id": "B3HCE", + "video_file": "Charades/videos/B3HCE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks in the entryway holding a bag of groceries' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <12.9>." + } + ] + }, + { + "video_id": "B3HCE", + "question_id": "B3HCE", + "video_file": "Charades/videos/B3HCE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person open a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.5> to <23.3>." + } + ] + }, + { + "video_id": "A9BDS", + "question_id": "A9BDS", + "video_file": "Charades/videos/A9BDS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a bag of groceries' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.9> to <7.3>." + } + ] + }, + { + "video_id": "QKOP0", + "question_id": "QKOP0", + "video_file": "Charades/videos/QKOP0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "1XFP6", + "question_id": "1XFP6", + "video_file": "Charades/videos/1XFP6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person lays on floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.4> to <31.0>." + } + ] + }, + { + "video_id": "PG32H", + "question_id": "PG32H", + "video_file": "Charades/videos/PG32H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person dresses' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <26.6>." + } + ] + }, + { + "video_id": "PG32H", + "question_id": "PG32H", + "video_file": "Charades/videos/PG32H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person tidily puts it on a table near a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <36.4> to <53.0>." + } + ] + }, + { + "video_id": "PG32H", + "question_id": "PG32H", + "video_file": "Charades/videos/PG32H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person dressing themselves'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <26.6>." + } + ] + }, + { + "video_id": "YLIMP", + "question_id": "YLIMP", + "video_file": "Charades/videos/YLIMP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person bending down to put their shoes on'." + }, + { + "from": "gpt", + "value": "From <11.7> to <30.4>." + } + ] + }, + { + "video_id": "WQSYC", + "question_id": "WQSYC", + "video_file": "Charades/videos/WQSYC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person reads a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <12.5>." + } + ] + }, + { + "video_id": "WQSYC", + "question_id": "WQSYC", + "video_file": "Charades/videos/WQSYC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinks some water from a bottle'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.5> to <21.0>." + } + ] + }, + { + "video_id": "WQSYC", + "question_id": "WQSYC", + "video_file": "Charades/videos/WQSYC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking from bottle of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.5> to <21.0>." + } + ] + }, + { + "video_id": "OMBR9", + "question_id": "OMBR9", + "video_file": "Charades/videos/OMBR9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.8> to <14.4>." + } + ] + }, + { + "video_id": "OMBR9", + "question_id": "OMBR9", + "video_file": "Charades/videos/OMBR9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sits down at a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.8> to <14.4>." + } + ] + }, + { + "video_id": "OMBR9", + "question_id": "OMBR9", + "video_file": "Charades/videos/OMBR9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts to eat food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.7> to <51.6>." + } + ] + }, + { + "video_id": "OMBR9", + "question_id": "OMBR9", + "video_file": "Charades/videos/OMBR9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats one'." + }, + { + "from": "gpt", + "value": "From <24.7> to <51.6>." + } + ] + }, + { + "video_id": "OMBR9", + "question_id": "OMBR9", + "video_file": "Charades/videos/OMBR9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begin eating food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.7> to <51.6>." + } + ] + }, + { + "video_id": "OMBR9", + "question_id": "OMBR9", + "video_file": "Charades/videos/OMBR9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person taking a drink from the bottle'." + }, + { + "from": "gpt", + "value": "From <48.2> to <63.0>." + } + ] + }, + { + "video_id": "83XF0", + "question_id": "83XF0", + "video_file": "Charades/videos/83XF0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.9>." + } + ] + }, + { + "video_id": "PZS0C", + "question_id": "PZS0C", + "video_file": "Charades/videos/PZS0C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <10.1>." + } + ] + }, + { + "video_id": "PZS0C", + "question_id": "PZS0C", + "video_file": "Charades/videos/PZS0C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is awakening in their bedroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "PZS0C", + "question_id": "PZS0C", + "video_file": "Charades/videos/PZS0C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person who is sitting on a bed gets up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <8.6>." + } + ] + }, + { + "video_id": "892RS", + "question_id": "892RS", + "video_file": "Charades/videos/892RS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walks through a doorway out of the room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.2> to <32.0>." + } + ] + }, + { + "video_id": "892RS", + "question_id": "892RS", + "video_file": "Charades/videos/892RS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person he opens the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.8> to <25.4>." + } + ] + }, + { + "video_id": "892RS", + "question_id": "892RS", + "video_file": "Charades/videos/892RS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts dishes inside' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.7> to <26.6>." + } + ] + }, + { + "video_id": "892RS", + "question_id": "892RS", + "video_file": "Charades/videos/892RS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts down the towel' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.5> to <20.8>." + } + ] + }, + { + "video_id": "1514T", + "question_id": "1514T", + "video_file": "Charades/videos/1514T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.7> to <23.2>." + } + ] + }, + { + "video_id": "1514T", + "question_id": "1514T", + "video_file": "Charades/videos/1514T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person laughing as they close the door in a pantry' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.7> to <23.2>." + } + ] + }, + { + "video_id": "AM4QC", + "question_id": "AM4QC", + "video_file": "Charades/videos/AM4QC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person closes the refrigerator'." + }, + { + "from": "gpt", + "value": "From <4.3> to <10.6>." + } + ] + }, + { + "video_id": "AM4QC", + "question_id": "AM4QC", + "video_file": "Charades/videos/AM4QC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "4JKLQ", + "question_id": "4JKLQ", + "video_file": "Charades/videos/4JKLQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.5> to <8.3>." + } + ] + }, + { + "video_id": "4JKLQ", + "question_id": "4JKLQ", + "video_file": "Charades/videos/4JKLQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts some food inside' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <18.4>." + } + ] + }, + { + "video_id": "PT2Z0", + "question_id": "PT2Z0", + "video_file": "Charades/videos/PT2Z0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws a plastic bag onto the floor'." + }, + { + "from": "gpt", + "value": "From <11.0> to <20.0>." + } + ] + }, + { + "video_id": "PT2Z0", + "question_id": "PT2Z0", + "video_file": "Charades/videos/PT2Z0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person throws a bag onto the ground' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.0> to <20.0>." + } + ] + }, + { + "video_id": "WRXI4", + "question_id": "WRXI4", + "video_file": "Charades/videos/WRXI4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a towel from the dryer' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.3> to <32.0>." + } + ] + }, + { + "video_id": "WRXI4", + "question_id": "WRXI4", + "video_file": "Charades/videos/WRXI4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts a towel into the washing machine' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <27.0> to <32.0>." + } + ] + }, + { + "video_id": "AGZ28", + "question_id": "AGZ28", + "video_file": "Charades/videos/AGZ28.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is closing a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "QYE21", + "question_id": "QYE21", + "video_file": "Charades/videos/QYE21.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close the window to the room they're in'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.9> to <30.7>." + } + ] + }, + { + "video_id": "QYE21", + "question_id": "QYE21", + "video_file": "Charades/videos/QYE21.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.9> to <30.7>." + } + ] + }, + { + "video_id": "2F7AQ", + "question_id": "2F7AQ", + "video_file": "Charades/videos/2F7AQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is playing on a laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.4>." + } + ] + }, + { + "video_id": "0TQZQ", + "question_id": "0TQZQ", + "video_file": "Charades/videos/0TQZQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person closes the book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <18.8>." + } + ] + }, + { + "video_id": "VGU6Y", + "question_id": "VGU6Y", + "video_file": "Charades/videos/VGU6Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.0> to <16.6>." + } + ] + }, + { + "video_id": "CDNV7", + "question_id": "CDNV7", + "video_file": "Charades/videos/CDNV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person turns on a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.4> to <23.8>." + } + ] + }, + { + "video_id": "CDNV7", + "question_id": "CDNV7", + "video_file": "Charades/videos/CDNV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns on the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.4> to <23.8>." + } + ] + }, + { + "video_id": "4JQBM", + "question_id": "4JQBM", + "video_file": "Charades/videos/4JQBM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person standing on a steps holding a blanket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.0>." + } + ] + }, + { + "video_id": "4JQBM", + "question_id": "4JQBM", + "video_file": "Charades/videos/4JQBM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is standing on the stairs holding a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.0>." + } + ] + }, + { + "video_id": "TRHB9", + "question_id": "TRHB9", + "video_file": "Charades/videos/TRHB9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the towel down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <9.1>." + } + ] + }, + { + "video_id": "TRHB9", + "question_id": "TRHB9", + "video_file": "Charades/videos/TRHB9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a towel'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.1>." + } + ] + }, + { + "video_id": "TRHB9", + "question_id": "TRHB9", + "video_file": "Charades/videos/TRHB9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put on shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.8> to <21.2>." + } + ] + }, + { + "video_id": "TRHB9", + "question_id": "TRHB9", + "video_file": "Charades/videos/TRHB9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting shoes on' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.8> to <21.2>." + } + ] + }, + { + "video_id": "TRHB9", + "question_id": "TRHB9", + "video_file": "Charades/videos/TRHB9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting a towel on a dryer' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <9.1>." + } + ] + }, + { + "video_id": "OLELC", + "question_id": "OLELC", + "video_file": "Charades/videos/OLELC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes a sip from a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <7.0>." + } + ] + }, + { + "video_id": "OLELC", + "question_id": "OLELC", + "video_file": "Charades/videos/OLELC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking out of a coffee cup'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "OLELC", + "question_id": "OLELC", + "video_file": "Charades/videos/OLELC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person cooking on the stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <10.7>." + } + ] + }, + { + "video_id": "279W6", + "question_id": "279W6", + "video_file": "Charades/videos/279W6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person snuggling up to a pillow'." + }, + { + "from": "gpt", + "value": "From <3.4> to <10.8>." + } + ] + }, + { + "video_id": "279W6", + "question_id": "279W6", + "video_file": "Charades/videos/279W6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person snuggles with a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.4> to <10.8>." + } + ] + }, + { + "video_id": "279W6", + "question_id": "279W6", + "video_file": "Charades/videos/279W6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person snuggling with a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <10.8>." + } + ] + }, + { + "video_id": "MKDCU", + "question_id": "MKDCU", + "video_file": "Charades/videos/MKDCU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person awakens in their bedroom snuggling a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "JYDCI", + "question_id": "JYDCI", + "video_file": "Charades/videos/JYDCI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is awakening in a bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <32.0>." + } + ] + }, + { + "video_id": "TK76G", + "question_id": "TK76G", + "video_file": "Charades/videos/TK76G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is eating at the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.4> to <17.0>." + } + ] + }, + { + "video_id": "QEG5O", + "question_id": "QEG5O", + "video_file": "Charades/videos/QEG5O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakens' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "QEG5O", + "question_id": "QEG5O", + "video_file": "Charades/videos/QEG5O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is awakening in bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "QEG5O", + "question_id": "QEG5O", + "video_file": "Charades/videos/QEG5O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person as they get up they begin sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.6> to <11.1>." + } + ] + }, + { + "video_id": "9LO9F", + "question_id": "9LO9F", + "video_file": "Charades/videos/9LO9F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person in bed sits up'." + }, + { + "from": "gpt", + "value": "From <27.4> to <37.0>." + } + ] + }, + { + "video_id": "9LO9F", + "question_id": "9LO9F", + "video_file": "Charades/videos/9LO9F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person is holding a medicine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.6> to <37.0>." + } + ] + }, + { + "video_id": "HFYNW", + "question_id": "HFYNW", + "video_file": "Charades/videos/HFYNW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they start sneezing'." + }, + { + "from": "gpt", + "value": "From <14.5> to <19.0>." + } + ] + }, + { + "video_id": "HFYNW", + "question_id": "HFYNW", + "video_file": "Charades/videos/HFYNW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting at a kitchen table eating'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "HFYNW", + "question_id": "HFYNW", + "video_file": "Charades/videos/HFYNW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is eating in the kitchen' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "V3LMY", + "question_id": "V3LMY", + "video_file": "Charades/videos/V3LMY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the cup down on the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.0> to <32.0>." + } + ] + }, + { + "video_id": "V3LMY", + "question_id": "V3LMY", + "video_file": "Charades/videos/V3LMY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person is sitting on a couch'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.7>." + } + ] + }, + { + "video_id": "YJV7W", + "question_id": "YJV7W", + "video_file": "Charades/videos/YJV7W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a picture of a light in the room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <12.0>." + } + ] + }, + { + "video_id": "YJV7W", + "question_id": "YJV7W", + "video_file": "Charades/videos/YJV7W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person begins dressing'." + }, + { + "from": "gpt", + "value": "From <9.6> to <16.6>." + } + ] + }, + { + "video_id": "YJV7W", + "question_id": "YJV7W", + "video_file": "Charades/videos/YJV7W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <12.0>." + } + ] + }, + { + "video_id": "YJV7W", + "question_id": "YJV7W", + "video_file": "Charades/videos/YJV7W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looks at a light'." + }, + { + "from": "gpt", + "value": "From <1.8> to <12.0>." + } + ] + }, + { + "video_id": "0XV4T", + "question_id": "0XV4T", + "video_file": "Charades/videos/0XV4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they put clothes into the washer' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.4> to <18.0>." + } + ] + }, + { + "video_id": "0XV4T", + "question_id": "0XV4T", + "video_file": "Charades/videos/0XV4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'this person pours a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "0XV4T", + "question_id": "0XV4T", + "video_file": "Charades/videos/0XV4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts clothes in the washer'." + }, + { + "from": "gpt", + "value": "From <9.4> to <18.0>." + } + ] + }, + { + "video_id": "0XV4T", + "question_id": "0XV4T", + "video_file": "Charades/videos/0XV4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person pours some liquid into a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "0XV4T", + "question_id": "0XV4T", + "video_file": "Charades/videos/0XV4T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'next the person puts a blanket in the washer' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.9> to <17.3>." + } + ] + }, + { + "video_id": "Y6P1E", + "question_id": "Y6P1E", + "video_file": "Charades/videos/Y6P1E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is opening a cabinet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.0> to <7.6>." + } + ] + }, + { + "video_id": "GK08Z", + "question_id": "GK08Z", + "video_file": "Charades/videos/GK08Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.3> to <22.7>." + } + ] + }, + { + "video_id": "GK08Z", + "question_id": "GK08Z", + "video_file": "Charades/videos/GK08Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <9.6>." + } + ] + }, + { + "video_id": "GK08Z", + "question_id": "GK08Z", + "video_file": "Charades/videos/GK08Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is opening a closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <9.6>." + } + ] + }, + { + "video_id": "GK08Z", + "question_id": "GK08Z", + "video_file": "Charades/videos/GK08Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting a box inside' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.2> to <15.3>." + } + ] + }, + { + "video_id": "RTEO8", + "question_id": "RTEO8", + "video_file": "Charades/videos/RTEO8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person throws down a pair of shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <9.1>." + } + ] + }, + { + "video_id": "RTEO8", + "question_id": "RTEO8", + "video_file": "Charades/videos/RTEO8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws some shoes on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.7> to <9.1>." + } + ] + }, + { + "video_id": "OO4FE", + "question_id": "OO4FE", + "video_file": "Charades/videos/OO4FE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <18.6>." + } + ] + }, + { + "video_id": "OO4FE", + "question_id": "OO4FE", + "video_file": "Charades/videos/OO4FE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person throws a bag that was on the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <6.6>." + } + ] + }, + { + "video_id": "71J7E", + "question_id": "71J7E", + "video_file": "Charades/videos/71J7E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put the broom in a corner' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.8> to <18.7>." + } + ] + }, + { + "video_id": "71J7E", + "question_id": "71J7E", + "video_file": "Charades/videos/71J7E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is putting a broom in a corner' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.8> to <18.7>." + } + ] + }, + { + "video_id": "LR6RZ", + "question_id": "LR6RZ", + "video_file": "Charades/videos/LR6RZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is running to the stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.8> to <11.4>." + } + ] + }, + { + "video_id": "LR6RZ", + "question_id": "LR6RZ", + "video_file": "Charades/videos/LR6RZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs into the kitchen'." + }, + { + "from": "gpt", + "value": "From <2.8> to <11.4>." + } + ] + }, + { + "video_id": "TIGIP", + "question_id": "TIGIP", + "video_file": "Charades/videos/TIGIP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts a blanket to their legs' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.7> to <21.6>." + } + ] + }, + { + "video_id": "TIGIP", + "question_id": "TIGIP", + "video_file": "Charades/videos/TIGIP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "TIGIP", + "question_id": "TIGIP", + "video_file": "Charades/videos/TIGIP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits down on a couch' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.4> to <13.5>." + } + ] + }, + { + "video_id": "TIGIP", + "question_id": "TIGIP", + "video_file": "Charades/videos/TIGIP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person sits on the sofa' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.4> to <13.5>." + } + ] + }, + { + "video_id": "TIGIP", + "question_id": "TIGIP", + "video_file": "Charades/videos/TIGIP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.5> to <9.2>." + } + ] + }, + { + "video_id": "48EEP", + "question_id": "48EEP", + "video_file": "Charades/videos/48EEP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "48EEP", + "question_id": "48EEP", + "video_file": "Charades/videos/48EEP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person laughing by the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "U79C9", + "question_id": "U79C9", + "video_file": "Charades/videos/U79C9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begin to snuggle with a blanket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.0> to <36.0>." + } + ] + }, + { + "video_id": "U79C9", + "question_id": "U79C9", + "video_file": "Charades/videos/U79C9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <17.5>." + } + ] + }, + { + "video_id": "01KM1", + "question_id": "01KM1", + "video_file": "Charades/videos/01KM1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they are eating some food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <26.3>." + } + ] + }, + { + "video_id": "WD6WB", + "question_id": "WD6WB", + "video_file": "Charades/videos/WD6WB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is fixing a doorknob' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.4> to <20.0>." + } + ] + }, + { + "video_id": "WD6WB", + "question_id": "WD6WB", + "video_file": "Charades/videos/WD6WB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person is fixing a squeaky doorknob' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.4> to <20.0>." + } + ] + }, + { + "video_id": "LZ8Z0", + "question_id": "LZ8Z0", + "video_file": "Charades/videos/LZ8Z0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the broom back to where it was found'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.8> to <34.4>." + } + ] + }, + { + "video_id": "LZ8Z0", + "question_id": "LZ8Z0", + "video_file": "Charades/videos/LZ8Z0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a broom'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <9.2>." + } + ] + }, + { + "video_id": "S1UFD", + "question_id": "S1UFD", + "video_file": "Charades/videos/S1UFD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.6> to <30.5>." + } + ] + }, + { + "video_id": "X1EZQ", + "question_id": "X1EZQ", + "video_file": "Charades/videos/X1EZQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <11.3>." + } + ] + }, + { + "video_id": "X1EZQ", + "question_id": "X1EZQ", + "video_file": "Charades/videos/X1EZQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating a snack' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.0> to <11.3>." + } + ] + }, + { + "video_id": "1LZ53", + "question_id": "1LZ53", + "video_file": "Charades/videos/1LZ53.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person sitting on a bed stands up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <13.0>." + } + ] + }, + { + "video_id": "1LZ53", + "question_id": "1LZ53", + "video_file": "Charades/videos/1LZ53.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks over to the glass window to close it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.2> to <31.4>." + } + ] + }, + { + "video_id": "1LZ53", + "question_id": "1LZ53", + "video_file": "Charades/videos/1LZ53.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sitting on bed watching tv' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.7> to <13.0>." + } + ] + }, + { + "video_id": "5A3FT", + "question_id": "5A3FT", + "video_file": "Charades/videos/5A3FT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person begins undressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.9> to <22.1>." + } + ] + }, + { + "video_id": "EWBOG", + "question_id": "EWBOG", + "video_file": "Charades/videos/EWBOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating from a plate' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <21.3>." + } + ] + }, + { + "video_id": "EWBOG", + "question_id": "EWBOG", + "video_file": "Charades/videos/EWBOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <22.3>." + } + ] + }, + { + "video_id": "EWBOG", + "question_id": "EWBOG", + "video_file": "Charades/videos/EWBOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they eat the sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.5> to <22.3>." + } + ] + }, + { + "video_id": "EWBOG", + "question_id": "EWBOG", + "video_file": "Charades/videos/EWBOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person in sneezing on a doorknob' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <11.6>." + } + ] + }, + { + "video_id": "R6TI2", + "question_id": "R6TI2", + "video_file": "Charades/videos/R6TI2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.9> to <29.2>." + } + ] + }, + { + "video_id": "R6TI2", + "question_id": "R6TI2", + "video_file": "Charades/videos/R6TI2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the door'." + }, + { + "from": "gpt", + "value": "From <22.9> to <29.2>." + } + ] + }, + { + "video_id": "R6TI2", + "question_id": "R6TI2", + "video_file": "Charades/videos/R6TI2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person stand up'." + }, + { + "from": "gpt", + "value": "From <23.1> to <28.5>." + } + ] + }, + { + "video_id": "6J8NE", + "question_id": "6J8NE", + "video_file": "Charades/videos/6J8NE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "0H04V", + "question_id": "0H04V", + "video_file": "Charades/videos/0H04V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'laughing at another person pouring something down the sink'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.5> to <26.8>." + } + ] + }, + { + "video_id": "0H04V", + "question_id": "0H04V", + "video_file": "Charades/videos/0H04V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person walks into his home holding a bag'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "3CA1X", + "question_id": "3CA1X", + "video_file": "Charades/videos/3CA1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eat it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.2> to <32.0>." + } + ] + }, + { + "video_id": "3CA1X", + "question_id": "3CA1X", + "video_file": "Charades/videos/3CA1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person start eating' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.2> to <32.0>." + } + ] + }, + { + "video_id": "81R3C", + "question_id": "81R3C", + "video_file": "Charades/videos/81R3C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they look out the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.5> to <17.5>." + } + ] + }, + { + "video_id": "81R3C", + "question_id": "81R3C", + "video_file": "Charades/videos/81R3C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person looks out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.5> to <17.5>." + } + ] + }, + { + "video_id": "81R3C", + "question_id": "81R3C", + "video_file": "Charades/videos/81R3C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts it back down on the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "KAU5T", + "question_id": "KAU5T", + "video_file": "Charades/videos/KAU5T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person runs out the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.5> to <33.0>." + } + ] + }, + { + "video_id": "KAU5T", + "question_id": "KAU5T", + "video_file": "Charades/videos/KAU5T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person drinks water from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.4>." + } + ] + }, + { + "video_id": "KAU5T", + "question_id": "KAU5T", + "video_file": "Charades/videos/KAU5T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs out of the room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.5> to <33.0>." + } + ] + }, + { + "video_id": "MKBH8", + "question_id": "MKBH8", + "video_file": "Charades/videos/MKBH8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person walks in the room smiling' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.5> to <31.0>." + } + ] + }, + { + "video_id": "MKBH8", + "question_id": "MKBH8", + "video_file": "Charades/videos/MKBH8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person walks into the kitchen smiling holding dishes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.5> to <31.0>." + } + ] + }, + { + "video_id": "PWC0F", + "question_id": "PWC0F", + "video_file": "Charades/videos/PWC0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts running to grab some medicine' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.6> to <11.7>." + } + ] + }, + { + "video_id": "PWC0F", + "question_id": "PWC0F", + "video_file": "Charades/videos/PWC0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person runs to the closet'." + }, + { + "from": "gpt", + "value": "From <5.6> to <11.7>." + } + ] + }, + { + "video_id": "PWC0F", + "question_id": "PWC0F", + "video_file": "Charades/videos/PWC0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person awakens' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.5>." + } + ] + }, + { + "video_id": "RVWK8", + "question_id": "RVWK8", + "video_file": "Charades/videos/RVWK8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.4> to <29.6>." + } + ] + }, + { + "video_id": "RVWK8", + "question_id": "RVWK8", + "video_file": "Charades/videos/RVWK8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person person takes some medicine from the shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.7> to <24.0>." + } + ] + }, + { + "video_id": "RVWK8", + "question_id": "RVWK8", + "video_file": "Charades/videos/RVWK8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <17.4>." + } + ] + }, + { + "video_id": "RQVJK", + "question_id": "RQVJK", + "video_file": "Charades/videos/RQVJK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sitting in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "RQVJK", + "question_id": "RQVJK", + "video_file": "Charades/videos/RQVJK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they begin sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "SL9ZD", + "question_id": "SL9ZD", + "video_file": "Charades/videos/SL9ZD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts a box on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "ZZN85", + "question_id": "ZZN85", + "video_file": "Charades/videos/ZZN85.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is putting dishes away in a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <9.7>." + } + ] + }, + { + "video_id": "ZZN85", + "question_id": "ZZN85", + "video_file": "Charades/videos/ZZN85.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is putting dishes in a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <9.7>." + } + ] + }, + { + "video_id": "3CAPI", + "question_id": "3CAPI", + "video_file": "Charades/videos/3CAPI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is throwing clothes everywhere' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <10.5>." + } + ] + }, + { + "video_id": "3CAPI", + "question_id": "3CAPI", + "video_file": "Charades/videos/3CAPI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person throws come clothes on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <10.5>." + } + ] + }, + { + "video_id": "QIQUA", + "question_id": "QIQUA", + "video_file": "Charades/videos/QIQUA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person washing the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <16.2>." + } + ] + }, + { + "video_id": "LVGX3", + "question_id": "LVGX3", + "video_file": "Charades/videos/LVGX3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking on a stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "LVGX3", + "question_id": "LVGX3", + "video_file": "Charades/videos/LVGX3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they stop undressing when they step on a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.9> to <15.2>." + } + ] + }, + { + "video_id": "LVGX3", + "question_id": "LVGX3", + "video_file": "Charades/videos/LVGX3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is cooking' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "LVGX3", + "question_id": "LVGX3", + "video_file": "Charades/videos/LVGX3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person beginning to undress' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.9> to <15.2>." + } + ] + }, + { + "video_id": "LVGX3", + "question_id": "LVGX3", + "video_file": "Charades/videos/LVGX3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begin undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.9> to <15.2>." + } + ] + }, + { + "video_id": "7U9EW", + "question_id": "7U9EW", + "video_file": "Charades/videos/7U9EW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person another walks through the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <18.3>." + } + ] + }, + { + "video_id": "7U9EW", + "question_id": "7U9EW", + "video_file": "Charades/videos/7U9EW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person eats in a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.3> to <28.4>." + } + ] + }, + { + "video_id": "7U9EW", + "question_id": "7U9EW", + "video_file": "Charades/videos/7U9EW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person is eating near the doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.3> to <28.4>." + } + ] + }, + { + "video_id": "2X7F6", + "question_id": "2X7F6", + "video_file": "Charades/videos/2X7F6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins eating'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <25.5>." + } + ] + }, + { + "video_id": "HOGVD", + "question_id": "HOGVD", + "video_file": "Charades/videos/HOGVD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins eating it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.5> to <25.4>." + } + ] + }, + { + "video_id": "HOGVD", + "question_id": "HOGVD", + "video_file": "Charades/videos/HOGVD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats one' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.5> to <25.4>." + } + ] + }, + { + "video_id": "CGNBJ", + "question_id": "CGNBJ", + "video_file": "Charades/videos/CGNBJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they throw them on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.3> to <23.4>." + } + ] + }, + { + "video_id": "CGNBJ", + "question_id": "CGNBJ", + "video_file": "Charades/videos/CGNBJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats an apple' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.8> to <39.0>." + } + ] + }, + { + "video_id": "CGNBJ", + "question_id": "CGNBJ", + "video_file": "Charades/videos/CGNBJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.2> to <21.7>." + } + ] + }, + { + "video_id": "CGNBJ", + "question_id": "CGNBJ", + "video_file": "Charades/videos/CGNBJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating an apple' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.8> to <39.0>." + } + ] + }, + { + "video_id": "CGNBJ", + "question_id": "CGNBJ", + "video_file": "Charades/videos/CGNBJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is in their closet tidying up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.9> to <30.4>." + } + ] + }, + { + "video_id": "CGNBJ", + "question_id": "CGNBJ", + "video_file": "Charades/videos/CGNBJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begin throwing clothes onto the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.3> to <23.4>." + } + ] + }, + { + "video_id": "CGNBJ", + "question_id": "CGNBJ", + "video_file": "Charades/videos/CGNBJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws clothes on the floor'." + }, + { + "from": "gpt", + "value": "From <17.3> to <23.4>." + } + ] + }, + { + "video_id": "CGNBJ", + "question_id": "CGNBJ", + "video_file": "Charades/videos/CGNBJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person open a closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.2> to <21.7>." + } + ] + }, + { + "video_id": "Q071E", + "question_id": "Q071E", + "video_file": "Charades/videos/Q071E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person a towel they open the bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.4> to <10.7>." + } + ] + }, + { + "video_id": "08Y62", + "question_id": "08Y62", + "video_file": "Charades/videos/08Y62.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person playing with the light switch'." + }, + { + "from": "gpt", + "value": "From <11.2> to <15.5>." + } + ] + }, + { + "video_id": "08Y62", + "question_id": "08Y62", + "video_file": "Charades/videos/08Y62.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <15.2>." + } + ] + }, + { + "video_id": "08Y62", + "question_id": "08Y62", + "video_file": "Charades/videos/08Y62.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns the light back on' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.2> to <15.5>." + } + ] + }, + { + "video_id": "GH3D1", + "question_id": "GH3D1", + "video_file": "Charades/videos/GH3D1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is cooking food on a stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <11.3>." + } + ] + }, + { + "video_id": "GH3D1", + "question_id": "GH3D1", + "video_file": "Charades/videos/GH3D1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person cooking on the stove is eating food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.9> to <13.3>." + } + ] + }, + { + "video_id": "WSB7W", + "question_id": "WSB7W", + "video_file": "Charades/videos/WSB7W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opens a laptop'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "WSB7W", + "question_id": "WSB7W", + "video_file": "Charades/videos/WSB7W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is opening a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "U37WC", + "question_id": "U37WC", + "video_file": "Charades/videos/U37WC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person finally eating a sandwich in a bedroom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.9> to <29.2>." + } + ] + }, + { + "video_id": "U37WC", + "question_id": "U37WC", + "video_file": "Charades/videos/U37WC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sneezing into a towel'." + }, + { + "from": "gpt", + "value": "From <18.1> to <24.2>." + } + ] + }, + { + "video_id": "1VUDV", + "question_id": "1VUDV", + "video_file": "Charades/videos/1VUDV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes a glass of water from the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <23.5>." + } + ] + }, + { + "video_id": "1VUDV", + "question_id": "1VUDV", + "video_file": "Charades/videos/1VUDV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person laughs as they cook food on the stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "WGZ29", + "question_id": "WGZ29", + "video_file": "Charades/videos/WGZ29.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person is sitting on the bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <12.5>." + } + ] + }, + { + "video_id": "WGZ29", + "question_id": "WGZ29", + "video_file": "Charades/videos/WGZ29.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sitting on the bed watching tv' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.8> to <12.5>." + } + ] + }, + { + "video_id": "4ZY8C", + "question_id": "4ZY8C", + "video_file": "Charades/videos/4ZY8C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts a book in a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.1> to <31.1>." + } + ] + }, + { + "video_id": "JQKR5", + "question_id": "JQKR5", + "video_file": "Charades/videos/JQKR5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating it'." + }, + { + "from": "gpt", + "value": "From <26.4> to <41.9>." + } + ] + }, + { + "video_id": "JQKR5", + "question_id": "JQKR5", + "video_file": "Charades/videos/JQKR5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <26.4> to <32.7>." + } + ] + }, + { + "video_id": "JQKR5", + "question_id": "JQKR5", + "video_file": "Charades/videos/JQKR5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person sits on a couch' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.8>." + } + ] + }, + { + "video_id": "0KISF", + "question_id": "0KISF", + "video_file": "Charades/videos/0KISF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person begins laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.2> to <31.0>." + } + ] + }, + { + "video_id": "MWAGL", + "question_id": "MWAGL", + "video_file": "Charades/videos/MWAGL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take out some food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.4> to <21.5>." + } + ] + }, + { + "video_id": "MWAGL", + "question_id": "MWAGL", + "video_file": "Charades/videos/MWAGL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.0> to <15.9>." + } + ] + }, + { + "video_id": "MWAGL", + "question_id": "MWAGL", + "video_file": "Charades/videos/MWAGL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes food from a refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.4> to <21.5>." + } + ] + }, + { + "video_id": "2BS0V", + "question_id": "2BS0V", + "video_file": "Charades/videos/2BS0V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.9> to <31.3>." + } + ] + }, + { + "video_id": "2BS0V", + "question_id": "2BS0V", + "video_file": "Charades/videos/2BS0V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closing the door behind them' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.3> to <35.0>." + } + ] + }, + { + "video_id": "AOBWQ", + "question_id": "AOBWQ", + "video_file": "Charades/videos/AOBWQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin throwing clothes on it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.7> to <30.3>." + } + ] + }, + { + "video_id": "Z9THH", + "question_id": "Z9THH", + "video_file": "Charades/videos/Z9THH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a bag'." + }, + { + "from": "gpt", + "value": "From <16.8> to <23.0>." + } + ] + }, + { + "video_id": "Z9THH", + "question_id": "Z9THH", + "video_file": "Charades/videos/Z9THH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes a backpack/bag from a bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <23.0>." + } + ] + }, + { + "video_id": "E780X", + "question_id": "E780X", + "video_file": "Charades/videos/E780X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts shoes on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "E780X", + "question_id": "E780X", + "video_file": "Charades/videos/E780X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinks from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <10.9>." + } + ] + }, + { + "video_id": "E780X", + "question_id": "E780X", + "video_file": "Charades/videos/E780X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking out of a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.7> to <10.9>." + } + ] + }, + { + "video_id": "O4P08", + "question_id": "O4P08", + "video_file": "Charades/videos/O4P08.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts the vacuum on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.0> to <16.4>." + } + ] + }, + { + "video_id": "ODFF6", + "question_id": "ODFF6", + "video_file": "Charades/videos/ODFF6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person two people are laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <32.0>." + } + ] + }, + { + "video_id": "ODFF6", + "question_id": "ODFF6", + "video_file": "Charades/videos/ODFF6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throwing a bag back'." + }, + { + "from": "gpt", + "value": "From <4.5> to <9.9>." + } + ] + }, + { + "video_id": "OJVOV", + "question_id": "OJVOV", + "video_file": "Charades/videos/OJVOV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a towel' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.1> to <31.0>." + } + ] + }, + { + "video_id": "OJVOV", + "question_id": "OJVOV", + "video_file": "Charades/videos/OJVOV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person holding a mirror is sitting on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.1> to <31.0>." + } + ] + }, + { + "video_id": "5EDX4", + "question_id": "5EDX4", + "video_file": "Charades/videos/5EDX4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begin to put the laptop in a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.3> to <33.0>." + } + ] + }, + { + "video_id": "5EDX4", + "question_id": "5EDX4", + "video_file": "Charades/videos/5EDX4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put the laptop in a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.3> to <33.0>." + } + ] + }, + { + "video_id": "5EDX4", + "question_id": "5EDX4", + "video_file": "Charades/videos/5EDX4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the laptop inside' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.3> to <33.0>." + } + ] + }, + { + "video_id": "5EDX4", + "question_id": "5EDX4", + "video_file": "Charades/videos/5EDX4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.3> to <33.0>." + } + ] + }, + { + "video_id": "5EDX4", + "question_id": "5EDX4", + "video_file": "Charades/videos/5EDX4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.7> to <25.4>." + } + ] + }, + { + "video_id": "EDVAM", + "question_id": "EDVAM", + "video_file": "Charades/videos/EDVAM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins closing the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.2> to <16.0>." + } + ] + }, + { + "video_id": "EDVAM", + "question_id": "EDVAM", + "video_file": "Charades/videos/EDVAM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is holding a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.1>." + } + ] + }, + { + "video_id": "EDVAM", + "question_id": "EDVAM", + "video_file": "Charades/videos/EDVAM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sitting on a couch' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "0V4B3", + "question_id": "0V4B3", + "video_file": "Charades/videos/0V4B3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person is drinking from glass cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <6.6>." + } + ] + }, + { + "video_id": "LU3SU", + "question_id": "LU3SU", + "video_file": "Charades/videos/LU3SU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take clothes out of the box'." + }, + { + "from": "gpt", + "value": "From <5.0> to <12.0>." + } + ] + }, + { + "video_id": "LU3SU", + "question_id": "LU3SU", + "video_file": "Charades/videos/LU3SU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "QCBIG", + "question_id": "QCBIG", + "video_file": "Charades/videos/QCBIG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is in a garage snuggling a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.4>." + } + ] + }, + { + "video_id": "QCBIG", + "question_id": "QCBIG", + "video_file": "Charades/videos/QCBIG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person start to undress by the doorway'." + }, + { + "from": "gpt", + "value": "From <13.7> to <22.6>." + } + ] + }, + { + "video_id": "56JT8", + "question_id": "56JT8", + "video_file": "Charades/videos/56JT8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is eating food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "56JT8", + "question_id": "56JT8", + "video_file": "Charades/videos/56JT8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is eating food off a plate'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "29JK6", + "question_id": "29JK6", + "video_file": "Charades/videos/29JK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.6> to <30.0>." + } + ] + }, + { + "video_id": "29JK6", + "question_id": "29JK6", + "video_file": "Charades/videos/29JK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting groceries away' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.9>." + } + ] + }, + { + "video_id": "29JK6", + "question_id": "29JK6", + "video_file": "Charades/videos/29JK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.4> to <35.9>." + } + ] + }, + { + "video_id": "29JK6", + "question_id": "29JK6", + "video_file": "Charades/videos/29JK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts groceries on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.9>." + } + ] + }, + { + "video_id": "29JK6", + "question_id": "29JK6", + "video_file": "Charades/videos/29JK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting some groceries into the right place' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.9>." + } + ] + }, + { + "video_id": "29JK6", + "question_id": "29JK6", + "video_file": "Charades/videos/29JK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they begin undressing by taking their jacket off'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <22.0>." + } + ] + }, + { + "video_id": "29JK6", + "question_id": "29JK6", + "video_file": "Charades/videos/29JK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the door behind them'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.4> to <35.9>." + } + ] + }, + { + "video_id": "A1VTR", + "question_id": "A1VTR", + "video_file": "Charades/videos/A1VTR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.2> to <29.7>." + } + ] + }, + { + "video_id": "A1VTR", + "question_id": "A1VTR", + "video_file": "Charades/videos/A1VTR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they start undressing by taking their jacket off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <29.7>." + } + ] + }, + { + "video_id": "EQ4GN", + "question_id": "EQ4GN", + "video_file": "Charades/videos/EQ4GN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes some medicine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "EQ4GN", + "question_id": "EQ4GN", + "video_file": "Charades/videos/EQ4GN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinks water from a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.8> to <24.5>." + } + ] + }, + { + "video_id": "ZP6LO", + "question_id": "ZP6LO", + "video_file": "Charades/videos/ZP6LO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking from a cup'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "2ECXI", + "question_id": "2ECXI", + "video_file": "Charades/videos/2ECXI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <13.2>." + } + ] + }, + { + "video_id": "2ECXI", + "question_id": "2ECXI", + "video_file": "Charades/videos/2ECXI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening their laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.7> to <13.2>." + } + ] + }, + { + "video_id": "2ECXI", + "question_id": "2ECXI", + "video_file": "Charades/videos/2ECXI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person lies on bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.1> to <8.1>." + } + ] + }, + { + "video_id": "1W2NR", + "question_id": "1W2NR", + "video_file": "Charades/videos/1W2NR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walks into room though doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <7.4>." + } + ] + }, + { + "video_id": "1W2NR", + "question_id": "1W2NR", + "video_file": "Charades/videos/1W2NR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person goes to sink to wash their hand' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.0> to <29.1>." + } + ] + }, + { + "video_id": "1W2NR", + "question_id": "1W2NR", + "video_file": "Charades/videos/1W2NR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person wash their hands at a sink' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.0> to <29.1>." + } + ] + }, + { + "video_id": "1W2NR", + "question_id": "1W2NR", + "video_file": "Charades/videos/1W2NR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person is walking around room wrapped in blanket sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.9> to <16.5>." + } + ] + }, + { + "video_id": "VL20E", + "question_id": "VL20E", + "video_file": "Charades/videos/VL20E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs away' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <23.0>." + } + ] + }, + { + "video_id": "VL20E", + "question_id": "VL20E", + "video_file": "Charades/videos/VL20E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person run out of the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.8> to <23.0>." + } + ] + }, + { + "video_id": "VL20E", + "question_id": "VL20E", + "video_file": "Charades/videos/VL20E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they begin sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.2> to <23.0>." + } + ] + }, + { + "video_id": "CHQ41", + "question_id": "CHQ41", + "video_file": "Charades/videos/CHQ41.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they begin drinking from a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.1> to <23.5>." + } + ] + }, + { + "video_id": "DJMZC", + "question_id": "DJMZC", + "video_file": "Charades/videos/DJMZC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakens in bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.2> to <14.2>." + } + ] + }, + { + "video_id": "65RS3", + "question_id": "65RS3", + "video_file": "Charades/videos/65RS3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holds up a broom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.4> to <28.2>." + } + ] + }, + { + "video_id": "E653K", + "question_id": "E653K", + "video_file": "Charades/videos/E653K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walks out closing door behind' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.0> to <27.0>." + } + ] + }, + { + "video_id": "E653K", + "question_id": "E653K", + "video_file": "Charades/videos/E653K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person open the front door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.0> to <24.0>." + } + ] + }, + { + "video_id": "E653K", + "question_id": "E653K", + "video_file": "Charades/videos/E653K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.0> to <24.0>." + } + ] + }, + { + "video_id": "BOOKV", + "question_id": "BOOKV", + "video_file": "Charades/videos/BOOKV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sitting on the couch' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.3> to <32.0>." + } + ] + }, + { + "video_id": "HA8MN", + "question_id": "HA8MN", + "video_file": "Charades/videos/HA8MN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person with the coffee takes the other person's book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.1> to <20.4>." + } + ] + }, + { + "video_id": "HA8MN", + "question_id": "HA8MN", + "video_file": "Charades/videos/HA8MN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes the book away' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.1> to <20.4>." + } + ] + }, + { + "video_id": "J0JD1", + "question_id": "J0JD1", + "video_file": "Charades/videos/J0JD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.5> to <7.1>." + } + ] + }, + { + "video_id": "J0JD1", + "question_id": "J0JD1", + "video_file": "Charades/videos/J0JD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person looks through the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.5> to <14.3>." + } + ] + }, + { + "video_id": "J0JD1", + "question_id": "J0JD1", + "video_file": "Charades/videos/J0JD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is running through the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "J0JD1", + "question_id": "J0JD1", + "video_file": "Charades/videos/J0JD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they look out the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.5> to <14.3>." + } + ] + }, + { + "video_id": "J0JD1", + "question_id": "J0JD1", + "video_file": "Charades/videos/J0JD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs into the entryway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "M1FNW", + "question_id": "M1FNW", + "video_file": "Charades/videos/M1FNW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a laptop off a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.8> to <6.9>." + } + ] + }, + { + "video_id": "KXAIZ", + "question_id": "KXAIZ", + "video_file": "Charades/videos/KXAIZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <14.3>." + } + ] + }, + { + "video_id": "KXAIZ", + "question_id": "KXAIZ", + "video_file": "Charades/videos/KXAIZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <14.3>." + } + ] + }, + { + "video_id": "KXAIZ", + "question_id": "KXAIZ", + "video_file": "Charades/videos/KXAIZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a cabinet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <14.3>." + } + ] + }, + { + "video_id": "KXAIZ", + "question_id": "KXAIZ", + "video_file": "Charades/videos/KXAIZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks through a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <5.7>." + } + ] + }, + { + "video_id": "N5U8Z", + "question_id": "N5U8Z", + "video_file": "Charades/videos/N5U8Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looks at a picture on the wall' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.0> to <37.4>." + } + ] + }, + { + "video_id": "TBXFO", + "question_id": "TBXFO", + "video_file": "Charades/videos/TBXFO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes off shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.5> to <24.1>." + } + ] + }, + { + "video_id": "TBXFO", + "question_id": "TBXFO", + "video_file": "Charades/videos/TBXFO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes off some shoes next to a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.5> to <24.1>." + } + ] + }, + { + "video_id": "TBXFO", + "question_id": "TBXFO", + "video_file": "Charades/videos/TBXFO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person someone walking in a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.1> to <16.7>." + } + ] + }, + { + "video_id": "TBXFO", + "question_id": "TBXFO", + "video_file": "Charades/videos/TBXFO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person turns on the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.8> to <41.1>." + } + ] + }, + { + "video_id": "TBXFO", + "question_id": "TBXFO", + "video_file": "Charades/videos/TBXFO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person pours a drink into a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <40.7> to <48.4>." + } + ] + }, + { + "video_id": "TCM46", + "question_id": "TCM46", + "video_file": "Charades/videos/TCM46.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns the light off'." + }, + { + "from": "gpt", + "value": "From <19.0> to <24.0>." + } + ] + }, + { + "video_id": "TCM46", + "question_id": "TCM46", + "video_file": "Charades/videos/TCM46.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person finally turning off a light on the stairs'." + }, + { + "from": "gpt", + "value": "From <19.0> to <24.0>." + } + ] + }, + { + "video_id": "7J5AA", + "question_id": "7J5AA", + "video_file": "Charades/videos/7J5AA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person turns on a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.9> to <17.9>." + } + ] + }, + { + "video_id": "7J5AA", + "question_id": "7J5AA", + "video_file": "Charades/videos/7J5AA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits down on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.2> to <24.0>." + } + ] + }, + { + "video_id": "7J5AA", + "question_id": "7J5AA", + "video_file": "Charades/videos/7J5AA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person turns on the light in the laundry room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.9> to <17.9>." + } + ] + }, + { + "video_id": "7J5AA", + "question_id": "7J5AA", + "video_file": "Charades/videos/7J5AA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they drink a cup of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <19.1>." + } + ] + }, + { + "video_id": "7J5AA", + "question_id": "7J5AA", + "video_file": "Charades/videos/7J5AA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.2> to <24.0>." + } + ] + }, + { + "video_id": "SCOML", + "question_id": "SCOML", + "video_file": "Charades/videos/SCOML.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person throws a pillow across the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.1> to <26.2>." + } + ] + }, + { + "video_id": "SCOML", + "question_id": "SCOML", + "video_file": "Charades/videos/SCOML.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws the pillow somewhere' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <26.2>." + } + ] + }, + { + "video_id": "KLP7V", + "question_id": "KLP7V", + "video_file": "Charades/videos/KLP7V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting items on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.6> to <10.1>." + } + ] + }, + { + "video_id": "KLP7V", + "question_id": "KLP7V", + "video_file": "Charades/videos/KLP7V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts some items on the shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.6> to <10.1>." + } + ] + }, + { + "video_id": "VIN8S", + "question_id": "VIN8S", + "video_file": "Charades/videos/VIN8S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is throwing a blanket on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.4> to <15.2>." + } + ] + }, + { + "video_id": "VIN8S", + "question_id": "VIN8S", + "video_file": "Charades/videos/VIN8S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts blanket on floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.8> to <12.3>." + } + ] + }, + { + "video_id": "VIN8S", + "question_id": "VIN8S", + "video_file": "Charades/videos/VIN8S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take picture with camera'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.3> to <29.5>." + } + ] + }, + { + "video_id": "58Z9R", + "question_id": "58Z9R", + "video_file": "Charades/videos/58Z9R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "58Z9R", + "question_id": "58Z9R", + "video_file": "Charades/videos/58Z9R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is opening a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "58Z9R", + "question_id": "58Z9R", + "video_file": "Charades/videos/58Z9R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sneezing looking at a camera' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.6> to <32.0>." + } + ] + }, + { + "video_id": "3O474", + "question_id": "3O474", + "video_file": "Charades/videos/3O474.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is undressing in their wardrobe closet'." + }, + { + "from": "gpt", + "value": "From <9.2> to <19.6>." + } + ] + }, + { + "video_id": "3O474", + "question_id": "3O474", + "video_file": "Charades/videos/3O474.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person undresses from their clothes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.2> to <19.6>." + } + ] + }, + { + "video_id": "PHN7Z", + "question_id": "PHN7Z", + "video_file": "Charades/videos/PHN7Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closing a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.0> to <27.0>." + } + ] + }, + { + "video_id": "PHN7Z", + "question_id": "PHN7Z", + "video_file": "Charades/videos/PHN7Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person finally closing a door in a living room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.0> to <27.0>." + } + ] + }, + { + "video_id": "PHN7Z", + "question_id": "PHN7Z", + "video_file": "Charades/videos/PHN7Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person dressing in clothes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <18.1>." + } + ] + }, + { + "video_id": "UQ93G", + "question_id": "UQ93G", + "video_file": "Charades/videos/UQ93G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.2> to <30.0>." + } + ] + }, + { + "video_id": "250SN", + "question_id": "250SN", + "video_file": "Charades/videos/250SN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is walking through a doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <6.7>." + } + ] + }, + { + "video_id": "250SN", + "question_id": "250SN", + "video_file": "Charades/videos/250SN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the door behind them'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.4> to <21.0>." + } + ] + }, + { + "video_id": "250SN", + "question_id": "250SN", + "video_file": "Charades/videos/250SN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs into the room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <17.2>." + } + ] + }, + { + "video_id": "250SN", + "question_id": "250SN", + "video_file": "Charades/videos/250SN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person runs back out'." + }, + { + "from": "gpt", + "value": "From <11.6> to <17.2>." + } + ] + }, + { + "video_id": "250SN", + "question_id": "250SN", + "video_file": "Charades/videos/250SN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens the door'." + }, + { + "from": "gpt", + "value": "From <1.5> to <6.2>." + } + ] + }, + { + "video_id": "SCTNB", + "question_id": "SCTNB", + "video_file": "Charades/videos/SCTNB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person runs to open a door'." + }, + { + "from": "gpt", + "value": "From <0.5> to <6.6>." + } + ] + }, + { + "video_id": "SCTNB", + "question_id": "SCTNB", + "video_file": "Charades/videos/SCTNB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs through the doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "SCTNB", + "question_id": "SCTNB", + "video_file": "Charades/videos/SCTNB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person runs back out' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "SCTNB", + "question_id": "SCTNB", + "video_file": "Charades/videos/SCTNB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.5> to <6.6>." + } + ] + }, + { + "video_id": "Z1EEF", + "question_id": "Z1EEF", + "video_file": "Charades/videos/Z1EEF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person eating the chips cuddles the bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "Z1EEF", + "question_id": "Z1EEF", + "video_file": "Charades/videos/Z1EEF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person is sitting on the toilet eating potato chips' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "Z1EEF", + "question_id": "Z1EEF", + "video_file": "Charades/videos/Z1EEF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting on a toilet eating chips' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "IPH08", + "question_id": "IPH08", + "video_file": "Charades/videos/IPH08.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person is holding shoes in hands' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.6>." + } + ] + }, + { + "video_id": "IPH08", + "question_id": "IPH08", + "video_file": "Charades/videos/IPH08.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they put their shoes down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.3> to <7.9>." + } + ] + }, + { + "video_id": "RSG1C", + "question_id": "RSG1C", + "video_file": "Charades/videos/RSG1C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the cabinet'." + }, + { + "from": "gpt", + "value": "From <4.9> to <12.7>." + } + ] + }, + { + "video_id": "RSG1C", + "question_id": "RSG1C", + "video_file": "Charades/videos/RSG1C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person puts a bag in a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <16.5>." + } + ] + }, + { + "video_id": "BZ1TH", + "question_id": "BZ1TH", + "video_file": "Charades/videos/BZ1TH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is drinking from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <11.3>." + } + ] + }, + { + "video_id": "BZ1TH", + "question_id": "BZ1TH", + "video_file": "Charades/videos/BZ1TH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <7.9>." + } + ] + }, + { + "video_id": "1AA6Z", + "question_id": "1AA6Z", + "video_file": "Charades/videos/1AA6Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is undressing in their recreation room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "1AA6Z", + "question_id": "1AA6Z", + "video_file": "Charades/videos/1AA6Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats some food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <19.3>." + } + ] + }, + { + "video_id": "D4AJ5", + "question_id": "D4AJ5", + "video_file": "Charades/videos/D4AJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'takes a couple of drinks from a cup.the person turns'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <9.4>." + } + ] + }, + { + "video_id": "D4AJ5", + "question_id": "D4AJ5", + "video_file": "Charades/videos/D4AJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person smiling to themselves in front of a stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <13.6>." + } + ] + }, + { + "video_id": "SP1QS", + "question_id": "SP1QS", + "video_file": "Charades/videos/SP1QS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours coffee into a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <20.9>." + } + ] + }, + { + "video_id": "SP1QS", + "question_id": "SP1QS", + "video_file": "Charades/videos/SP1QS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "SP1QS", + "question_id": "SP1QS", + "video_file": "Charades/videos/SP1QS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pour some coffee in a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <20.9>." + } + ] + }, + { + "video_id": "SP1QS", + "question_id": "SP1QS", + "video_file": "Charades/videos/SP1QS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sitting on a chair watching coffee brew' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "5MF9Y", + "question_id": "5MF9Y", + "video_file": "Charades/videos/5MF9Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person can run with a broom'." + }, + { + "from": "gpt", + "value": "From <0.6> to <6.5>." + } + ] + }, + { + "video_id": "5MF9Y", + "question_id": "5MF9Y", + "video_file": "Charades/videos/5MF9Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts to run out of the room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.6> to <6.5>." + } + ] + }, + { + "video_id": "5MF9Y", + "question_id": "5MF9Y", + "video_file": "Charades/videos/5MF9Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person runs in the room holding a broom'." + }, + { + "from": "gpt", + "value": "From <0.6> to <6.5>." + } + ] + }, + { + "video_id": "2P2ZC", + "question_id": "2P2ZC", + "video_file": "Charades/videos/2P2ZC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns the light off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.7> to <29.0>." + } + ] + }, + { + "video_id": "2P2ZC", + "question_id": "2P2ZC", + "video_file": "Charades/videos/2P2ZC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is trying to turn on a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.1> to <27.8>." + } + ] + }, + { + "video_id": "2P2ZC", + "question_id": "2P2ZC", + "video_file": "Charades/videos/2P2ZC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person was fixing the cabinet light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.1> to <27.8>." + } + ] + }, + { + "video_id": "AIYFU", + "question_id": "AIYFU", + "video_file": "Charades/videos/AIYFU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.5> to <18.8>." + } + ] + }, + { + "video_id": "LURLC", + "question_id": "LURLC", + "video_file": "Charades/videos/LURLC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they are eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.6> to <27.4>." + } + ] + }, + { + "video_id": "SRQPI", + "question_id": "SRQPI", + "video_file": "Charades/videos/SRQPI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person the other closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.7> to <20.0>." + } + ] + }, + { + "video_id": "SRQPI", + "question_id": "SRQPI", + "video_file": "Charades/videos/SRQPI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.7> to <20.0>." + } + ] + }, + { + "video_id": "SRQPI", + "question_id": "SRQPI", + "video_file": "Charades/videos/SRQPI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the second person turns on the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <23.0>." + } + ] + }, + { + "video_id": "SRQPI", + "question_id": "SRQPI", + "video_file": "Charades/videos/SRQPI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'eats the sandwich the other person turns the lights on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <23.0>." + } + ] + }, + { + "video_id": "SRQPI", + "question_id": "SRQPI", + "video_file": "Charades/videos/SRQPI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <23.0>." + } + ] + }, + { + "video_id": "FR4K2", + "question_id": "FR4K2", + "video_file": "Charades/videos/FR4K2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens their laptop'." + }, + { + "from": "gpt", + "value": "From <23.4> to <30.1>." + } + ] + }, + { + "video_id": "FR4K2", + "question_id": "FR4K2", + "video_file": "Charades/videos/FR4K2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person washes their hands in the sink'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.1> to <22.2>." + } + ] + }, + { + "video_id": "EMYSA", + "question_id": "EMYSA", + "video_file": "Charades/videos/EMYSA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes a bag from the refrigerator'." + }, + { + "from": "gpt", + "value": "From <5.0> to <12.3>." + } + ] + }, + { + "video_id": "BUXV4", + "question_id": "BUXV4", + "video_file": "Charades/videos/BUXV4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is putting shoes inside of a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.2> to <8.1>." + } + ] + }, + { + "video_id": "BUXV4", + "question_id": "BUXV4", + "video_file": "Charades/videos/BUXV4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person smiling at their phone'." + }, + { + "from": "gpt", + "value": "From <16.2> to <23.4>." + } + ] + }, + { + "video_id": "BUXV4", + "question_id": "BUXV4", + "video_file": "Charades/videos/BUXV4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holds a phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.0> to <31.9>." + } + ] + }, + { + "video_id": "5AFC1", + "question_id": "5AFC1", + "video_file": "Charades/videos/5AFC1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.8> to <34.0>." + } + ] + }, + { + "video_id": "5AFC1", + "question_id": "5AFC1", + "video_file": "Charades/videos/5AFC1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walking over to a table to eat something' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.8> to <34.0>." + } + ] + }, + { + "video_id": "5AFC1", + "question_id": "5AFC1", + "video_file": "Charades/videos/5AFC1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes a sandwich off the desk' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.6> to <27.6>." + } + ] + }, + { + "video_id": "HIOY0", + "question_id": "HIOY0", + "video_file": "Charades/videos/HIOY0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws it on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.2> to <20.0>." + } + ] + }, + { + "video_id": "HIOY0", + "question_id": "HIOY0", + "video_file": "Charades/videos/HIOY0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person laughs at what's on it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.5> to <37.0>." + } + ] + }, + { + "video_id": "JAOPD", + "question_id": "JAOPD", + "video_file": "Charades/videos/JAOPD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting a bag on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <9.1>." + } + ] + }, + { + "video_id": "NB6F5", + "question_id": "NB6F5", + "video_file": "Charades/videos/NB6F5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens the bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "NB6F5", + "question_id": "NB6F5", + "video_file": "Charades/videos/NB6F5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turn the light off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.4> to <16.0>." + } + ] + }, + { + "video_id": "NB6F5", + "question_id": "NB6F5", + "video_file": "Charades/videos/NB6F5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.4> to <16.0>." + } + ] + }, + { + "video_id": "Z2MBK", + "question_id": "Z2MBK", + "video_file": "Charades/videos/Z2MBK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.6> to <19.0>." + } + ] + }, + { + "video_id": "Z2MBK", + "question_id": "Z2MBK", + "video_file": "Charades/videos/Z2MBK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a sandwich turns' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.6> to <19.0>." + } + ] + }, + { + "video_id": "Z2MBK", + "question_id": "Z2MBK", + "video_file": "Charades/videos/Z2MBK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking a glass of coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.9> to <14.8>." + } + ] + }, + { + "video_id": "Z2MBK", + "question_id": "Z2MBK", + "video_file": "Charades/videos/Z2MBK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is drinking something from a cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.9> to <14.8>." + } + ] + }, + { + "video_id": "NZ6PU", + "question_id": "NZ6PU", + "video_file": "Charades/videos/NZ6PU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns off the light in the room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.2> to <30.0>." + } + ] + }, + { + "video_id": "L2LF6", + "question_id": "L2LF6", + "video_file": "Charades/videos/L2LF6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door behind them' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <14.9>." + } + ] + }, + { + "video_id": "YFI1M", + "question_id": "YFI1M", + "video_file": "Charades/videos/YFI1M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person begins undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.9> to <13.7>." + } + ] + }, + { + "video_id": "V4Q75", + "question_id": "V4Q75", + "video_file": "Charades/videos/V4Q75.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a plate of food'." + }, + { + "from": "gpt", + "value": "From <2.1> to <14.3>." + } + ] + }, + { + "video_id": "V4Q75", + "question_id": "V4Q75", + "video_file": "Charades/videos/V4Q75.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating from it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.1> to <14.3>." + } + ] + }, + { + "video_id": "V4Q75", + "question_id": "V4Q75", + "video_file": "Charades/videos/V4Q75.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is eating food out of a dish' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.1> to <14.3>." + } + ] + }, + { + "video_id": "V4Q75", + "question_id": "V4Q75", + "video_file": "Charades/videos/V4Q75.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they drink some water out of a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.5> to <23.3>." + } + ] + }, + { + "video_id": "H3MG3", + "question_id": "H3MG3", + "video_file": "Charades/videos/H3MG3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put the shoes in a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.5> to <20.3>." + } + ] + }, + { + "video_id": "H3MG3", + "question_id": "H3MG3", + "video_file": "Charades/videos/H3MG3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.3> to <32.0>." + } + ] + }, + { + "video_id": "KRF68", + "question_id": "KRF68", + "video_file": "Charades/videos/KRF68.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs in' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "KRF68", + "question_id": "KRF68", + "video_file": "Charades/videos/KRF68.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person starts dressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.3> to <25.0>." + } + ] + }, + { + "video_id": "KRF68", + "question_id": "KRF68", + "video_file": "Charades/videos/KRF68.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs into their laundry room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "KRF68", + "question_id": "KRF68", + "video_file": "Charades/videos/KRF68.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person close the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.2> to <10.9>." + } + ] + }, + { + "video_id": "8R5DD", + "question_id": "8R5DD", + "video_file": "Charades/videos/8R5DD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is putting a blanket onto another person' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <17.7>." + } + ] + }, + { + "video_id": "W5RC5", + "question_id": "W5RC5", + "video_file": "Charades/videos/W5RC5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes medicine followed by a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <10.4>." + } + ] + }, + { + "video_id": "73XDD", + "question_id": "73XDD", + "video_file": "Charades/videos/73XDD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits on the couch' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "73XDD", + "question_id": "73XDD", + "video_file": "Charades/videos/73XDD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sitting on the couch' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "73XDD", + "question_id": "73XDD", + "video_file": "Charades/videos/73XDD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they begin undressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <22.8>." + } + ] + }, + { + "video_id": "16KTS", + "question_id": "16KTS", + "video_file": "Charades/videos/16KTS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person opens up the refrigerator to look for something'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.7> to <31.0>." + } + ] + }, + { + "video_id": "POU1Z", + "question_id": "POU1Z", + "video_file": "Charades/videos/POU1Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns off the light in the closet'." + }, + { + "from": "gpt", + "value": "From <32.0> to <36.6>." + } + ] + }, + { + "video_id": "PPKP0", + "question_id": "PPKP0", + "video_file": "Charades/videos/PPKP0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person holds a plate with some food on it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.1>." + } + ] + }, + { + "video_id": "Q8TB3", + "question_id": "Q8TB3", + "video_file": "Charades/videos/Q8TB3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <15.1>." + } + ] + }, + { + "video_id": "Q8TB3", + "question_id": "Q8TB3", + "video_file": "Charades/videos/Q8TB3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <15.1>." + } + ] + }, + { + "video_id": "UWPIZ", + "question_id": "UWPIZ", + "video_file": "Charades/videos/UWPIZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.1> to <14.8>." + } + ] + }, + { + "video_id": "QDVXO", + "question_id": "QDVXO", + "video_file": "Charades/videos/QDVXO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is tidying their entryway with a broom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.2>." + } + ] + }, + { + "video_id": "DE8T8", + "question_id": "DE8T8", + "video_file": "Charades/videos/DE8T8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person is standing by shelf smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.5> to <27.9>." + } + ] + }, + { + "video_id": "FLM1T", + "question_id": "FLM1T", + "video_file": "Charades/videos/FLM1T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person start eating it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.5> to <23.0>." + } + ] + }, + { + "video_id": "FLM1T", + "question_id": "FLM1T", + "video_file": "Charades/videos/FLM1T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they eat some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.5> to <23.0>." + } + ] + }, + { + "video_id": "NNZON", + "question_id": "NNZON", + "video_file": "Charades/videos/NNZON.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person b sit down on the couch'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "1375X", + "question_id": "1375X", + "video_file": "Charades/videos/1375X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put it on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.4> to <28.1>." + } + ] + }, + { + "video_id": "1375X", + "question_id": "1375X", + "video_file": "Charades/videos/1375X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person smiling in a pantry'." + }, + { + "from": "gpt", + "value": "From <10.0> to <22.8>." + } + ] + }, + { + "video_id": "1375X", + "question_id": "1375X", + "video_file": "Charades/videos/1375X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the bag in a garbage' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.3> to <11.8>." + } + ] + }, + { + "video_id": "1375X", + "question_id": "1375X", + "video_file": "Charades/videos/1375X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sneezing into a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "DOQ9Y", + "question_id": "DOQ9Y", + "video_file": "Charades/videos/DOQ9Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they close a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <34.0> to <40.9>." + } + ] + }, + { + "video_id": "DOQ9Y", + "question_id": "DOQ9Y", + "video_file": "Charades/videos/DOQ9Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <34.0> to <40.9>." + } + ] + }, + { + "video_id": "DOQ9Y", + "question_id": "DOQ9Y", + "video_file": "Charades/videos/DOQ9Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person undressing by removing their jacket in the pantry'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.1>." + } + ] + }, + { + "video_id": "OEYA3", + "question_id": "OEYA3", + "video_file": "Charades/videos/OEYA3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sits in a chair in the closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.2>." + } + ] + }, + { + "video_id": "OEYA3", + "question_id": "OEYA3", + "video_file": "Charades/videos/OEYA3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person sits down in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.2>." + } + ] + }, + { + "video_id": "OEYA3", + "question_id": "OEYA3", + "video_file": "Charades/videos/OEYA3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person sits in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.2>." + } + ] + }, + { + "video_id": "U6KQ7", + "question_id": "U6KQ7", + "video_file": "Charades/videos/U6KQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person the other one closes the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.9> to <15.0>." + } + ] + }, + { + "video_id": "U6KQ7", + "question_id": "U6KQ7", + "video_file": "Charades/videos/U6KQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.9> to <15.0>." + } + ] + }, + { + "video_id": "U6KQ7", + "question_id": "U6KQ7", + "video_file": "Charades/videos/U6KQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person the last on is eating chips' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "U6KQ7", + "question_id": "U6KQ7", + "video_file": "Charades/videos/U6KQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the door behind them' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.9> to <15.0>." + } + ] + }, + { + "video_id": "35C4E", + "question_id": "35C4E", + "video_file": "Charades/videos/35C4E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opened the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.6> to <30.8>." + } + ] + }, + { + "video_id": "35C4E", + "question_id": "35C4E", + "video_file": "Charades/videos/35C4E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.6> to <30.8>." + } + ] + }, + { + "video_id": "35C4E", + "question_id": "35C4E", + "video_file": "Charades/videos/35C4E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they open the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <30.8>." + } + ] + }, + { + "video_id": "6W2MO", + "question_id": "6W2MO", + "video_file": "Charades/videos/6W2MO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begin sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.7> to <16.6>." + } + ] + }, + { + "video_id": "XB2NM", + "question_id": "XB2NM", + "video_file": "Charades/videos/XB2NM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits on the floor'." + }, + { + "from": "gpt", + "value": "From <16.1> to <25.0>." + } + ] + }, + { + "video_id": "XB2NM", + "question_id": "XB2NM", + "video_file": "Charades/videos/XB2NM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sitting on the floor'." + }, + { + "from": "gpt", + "value": "From <16.1> to <25.0>." + } + ] + }, + { + "video_id": "WB17C", + "question_id": "WB17C", + "video_file": "Charades/videos/WB17C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person than began sneezing alot' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.2> to <35.0>." + } + ] + }, + { + "video_id": "WB17C", + "question_id": "WB17C", + "video_file": "Charades/videos/WB17C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <7.2>." + } + ] + }, + { + "video_id": "KFZ55", + "question_id": "KFZ55", + "video_file": "Charades/videos/KFZ55.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking from their morning cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <7.2>." + } + ] + }, + { + "video_id": "SSKK6", + "question_id": "SSKK6", + "video_file": "Charades/videos/SSKK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts it in a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.7> to <28.4>." + } + ] + }, + { + "video_id": "SSKK6", + "question_id": "SSKK6", + "video_file": "Charades/videos/SSKK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throw the rest in a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.2> to <27.3>." + } + ] + }, + { + "video_id": "I2DHG", + "question_id": "I2DHG", + "video_file": "Charades/videos/I2DHG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking from a cup'." + }, + { + "from": "gpt", + "value": "From <13.1> to <23.4>." + } + ] + }, + { + "video_id": "I2DHG", + "question_id": "I2DHG", + "video_file": "Charades/videos/I2DHG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person was undressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.8> to <11.4>." + } + ] + }, + { + "video_id": "I2DHG", + "question_id": "I2DHG", + "video_file": "Charades/videos/I2DHG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting their clothes on the bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "I2DHG", + "question_id": "I2DHG", + "video_file": "Charades/videos/I2DHG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person another was smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.7> to <32.0>." + } + ] + }, + { + "video_id": "I2DHG", + "question_id": "I2DHG", + "video_file": "Charades/videos/I2DHG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks through a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "FM6KD", + "question_id": "FM6KD", + "video_file": "Charades/videos/FM6KD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is reading a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.1> to <26.2>." + } + ] + }, + { + "video_id": "FM6KD", + "question_id": "FM6KD", + "video_file": "Charades/videos/FM6KD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person laughing at something' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.1> to <10.7>." + } + ] + }, + { + "video_id": "FM6KD", + "question_id": "FM6KD", + "video_file": "Charades/videos/FM6KD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person reading a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.1> to <26.2>." + } + ] + }, + { + "video_id": "XQC9F", + "question_id": "XQC9F", + "video_file": "Charades/videos/XQC9F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put the clothes into it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.7> to <23.4>." + } + ] + }, + { + "video_id": "XV8CH", + "question_id": "XV8CH", + "video_file": "Charades/videos/XV8CH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts their keys on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <26.0>." + } + ] + }, + { + "video_id": "XV8CH", + "question_id": "XV8CH", + "video_file": "Charades/videos/XV8CH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking food at a stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "XV8CH", + "question_id": "XV8CH", + "video_file": "Charades/videos/XV8CH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking at the stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "UN1PD", + "question_id": "UN1PD", + "video_file": "Charades/videos/UN1PD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they wash a dish in the sink'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.9> to <28.0>." + } + ] + }, + { + "video_id": "UN1PD", + "question_id": "UN1PD", + "video_file": "Charades/videos/UN1PD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts to wash dishes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.9> to <28.0>." + } + ] + }, + { + "video_id": "VS2RI", + "question_id": "VS2RI", + "video_file": "Charades/videos/VS2RI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person running up the stairs' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.6> to <16.0>." + } + ] + }, + { + "video_id": "L071S", + "question_id": "L071S", + "video_file": "Charades/videos/L071S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.1> to <23.6>." + } + ] + }, + { + "video_id": "L071S", + "question_id": "L071S", + "video_file": "Charades/videos/L071S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.1> to <23.6>." + } + ] + }, + { + "video_id": "UB2EJ", + "question_id": "UB2EJ", + "video_file": "Charades/videos/UB2EJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking a glass of iced coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.8> to <9.1>." + } + ] + }, + { + "video_id": "UB2EJ", + "question_id": "UB2EJ", + "video_file": "Charades/videos/UB2EJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is standing in their hallway dressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "GWXD9", + "question_id": "GWXD9", + "video_file": "Charades/videos/GWXD9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is pouring a cup of coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "GWXD9", + "question_id": "GWXD9", + "video_file": "Charades/videos/GWXD9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pour it in to a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "JIDHM", + "question_id": "JIDHM", + "video_file": "Charades/videos/JIDHM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looking out the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <16.5>." + } + ] + }, + { + "video_id": "JIDHM", + "question_id": "JIDHM", + "video_file": "Charades/videos/JIDHM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is watching out the window'." + }, + { + "from": "gpt", + "value": "From <0.0> to <16.5>." + } + ] + }, + { + "video_id": "XI9LI", + "question_id": "XI9LI", + "video_file": "Charades/videos/XI9LI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.6> to <26.3>." + } + ] + }, + { + "video_id": "NE37I", + "question_id": "NE37I", + "video_file": "Charades/videos/NE37I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take out a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.1> to <22.2>." + } + ] + }, + { + "video_id": "XC8DF", + "question_id": "XC8DF", + "video_file": "Charades/videos/XC8DF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.2> to <16.8>." + } + ] + }, + { + "video_id": "J6WE2", + "question_id": "J6WE2", + "video_file": "Charades/videos/J6WE2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the bag back on the chair'." + }, + { + "from": "gpt", + "value": "From <22.9> to <30.5>." + } + ] + }, + { + "video_id": "J6WE2", + "question_id": "J6WE2", + "video_file": "Charades/videos/J6WE2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws the pillow on a bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.9> to <34.9>." + } + ] + }, + { + "video_id": "J6WE2", + "question_id": "J6WE2", + "video_file": "Charades/videos/J6WE2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes a pillow from a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <15.6>." + } + ] + }, + { + "video_id": "J6WE2", + "question_id": "J6WE2", + "video_file": "Charades/videos/J6WE2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes a pillow out of a bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.0> to <15.6>." + } + ] + }, + { + "video_id": "J6WE2", + "question_id": "J6WE2", + "video_file": "Charades/videos/J6WE2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.2> to <12.7>." + } + ] + }, + { + "video_id": "CIIWP", + "question_id": "CIIWP", + "video_file": "Charades/videos/CIIWP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <23.2>." + } + ] + }, + { + "video_id": "CIIWP", + "question_id": "CIIWP", + "video_file": "Charades/videos/CIIWP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the door behind them' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.5> to <23.2>." + } + ] + }, + { + "video_id": "CIIWP", + "question_id": "CIIWP", + "video_file": "Charades/videos/CIIWP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws a bag onto the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <8.3>." + } + ] + }, + { + "video_id": "EOKJ0", + "question_id": "EOKJ0", + "video_file": "Charades/videos/EOKJ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person just puts the clothes on the floor'." + }, + { + "from": "gpt", + "value": "From <9.0> to <19.0>." + } + ] + }, + { + "video_id": "EOKJ0", + "question_id": "EOKJ0", + "video_file": "Charades/videos/EOKJ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person is holding book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "YA379", + "question_id": "YA379", + "video_file": "Charades/videos/YA379.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks into the hallway drinking from a cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.7> to <8.7>." + } + ] + }, + { + "video_id": "R2DA8", + "question_id": "R2DA8", + "video_file": "Charades/videos/R2DA8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person working on a laptop on a table sneezes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <17.7>." + } + ] + }, + { + "video_id": "R2DA8", + "question_id": "R2DA8", + "video_file": "Charades/videos/R2DA8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open their refrigerator'." + }, + { + "from": "gpt", + "value": "From <28.1> to <33.9>." + } + ] + }, + { + "video_id": "R2DA8", + "question_id": "R2DA8", + "video_file": "Charades/videos/R2DA8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins throwing food on a plate from the fridge'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.9> to <43.1>." + } + ] + }, + { + "video_id": "R2DA8", + "question_id": "R2DA8", + "video_file": "Charades/videos/R2DA8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person sitting on a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <20.0>." + } + ] + }, + { + "video_id": "R2DA8", + "question_id": "R2DA8", + "video_file": "Charades/videos/R2DA8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is seen sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <17.3>." + } + ] + }, + { + "video_id": "1G2QC", + "question_id": "1G2QC", + "video_file": "Charades/videos/1G2QC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they begin drinking from a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <28.8>." + } + ] + }, + { + "video_id": "1G2QC", + "question_id": "1G2QC", + "video_file": "Charades/videos/1G2QC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take a towel' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.9> to <35.0>." + } + ] + }, + { + "video_id": "OHZG5", + "question_id": "OHZG5", + "video_file": "Charades/videos/OHZG5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they laugh as they read the paper' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <14.0>." + } + ] + }, + { + "video_id": "OHZG5", + "question_id": "OHZG5", + "video_file": "Charades/videos/OHZG5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking a cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "B71ZJ", + "question_id": "B71ZJ", + "video_file": "Charades/videos/B71ZJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts a bottle of medicine on the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.0> to <8.2>." + } + ] + }, + { + "video_id": "B71ZJ", + "question_id": "B71ZJ", + "video_file": "Charades/videos/B71ZJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <15.3>." + } + ] + }, + { + "video_id": "B71ZJ", + "question_id": "B71ZJ", + "video_file": "Charades/videos/B71ZJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes their shoes out of a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <19.2>." + } + ] + }, + { + "video_id": "CXSYS", + "question_id": "CXSYS", + "video_file": "Charades/videos/CXSYS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person cooks on a stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <24.0>." + } + ] + }, + { + "video_id": "CXSYS", + "question_id": "CXSYS", + "video_file": "Charades/videos/CXSYS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person start cooking at the stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.0> to <24.0>." + } + ] + }, + { + "video_id": "ICL1M", + "question_id": "ICL1M", + "video_file": "Charades/videos/ICL1M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they stand up' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.8> to <29.2>." + } + ] + }, + { + "video_id": "ICL1M", + "question_id": "ICL1M", + "video_file": "Charades/videos/ICL1M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts down the glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <31.0>." + } + ] + }, + { + "video_id": "F2NAW", + "question_id": "F2NAW", + "video_file": "Charades/videos/F2NAW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is just finishing dressing in the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.8>." + } + ] + }, + { + "video_id": "F2NAW", + "question_id": "F2NAW", + "video_file": "Charades/videos/F2NAW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they turn off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.9> to <21.8>." + } + ] + }, + { + "video_id": "F2NAW", + "question_id": "F2NAW", + "video_file": "Charades/videos/F2NAW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person shutting off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.9> to <21.8>." + } + ] + }, + { + "video_id": "F2NAW", + "question_id": "F2NAW", + "video_file": "Charades/videos/F2NAW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person close the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.4> to <26.7>." + } + ] + }, + { + "video_id": "S4D1O", + "question_id": "S4D1O", + "video_file": "Charades/videos/S4D1O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person looks out the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.1> to <24.0>." + } + ] + }, + { + "video_id": "GQKIU", + "question_id": "GQKIU", + "video_file": "Charades/videos/GQKIU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person proceed to take their coffee cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <16.0>." + } + ] + }, + { + "video_id": "LWBPP", + "question_id": "LWBPP", + "video_file": "Charades/videos/LWBPP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person turns on a light in their closet'." + }, + { + "from": "gpt", + "value": "From <1.8> to <6.3>." + } + ] + }, + { + "video_id": "LWBPP", + "question_id": "LWBPP", + "video_file": "Charades/videos/LWBPP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turning the light on' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.8> to <6.3>." + } + ] + }, + { + "video_id": "LWBPP", + "question_id": "LWBPP", + "video_file": "Charades/videos/LWBPP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person turns on the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.8> to <6.3>." + } + ] + }, + { + "video_id": "LDFM5", + "question_id": "LDFM5", + "video_file": "Charades/videos/LDFM5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.7> to <32.0>." + } + ] + }, + { + "video_id": "LDFM5", + "question_id": "LDFM5", + "video_file": "Charades/videos/LDFM5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <7.7>." + } + ] + }, + { + "video_id": "2XVR0", + "question_id": "2XVR0", + "video_file": "Charades/videos/2XVR0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts dishes away' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.3> to <21.3>." + } + ] + }, + { + "video_id": "2XVR0", + "question_id": "2XVR0", + "video_file": "Charades/videos/2XVR0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put them on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <33.1>." + } + ] + }, + { + "video_id": "2XVR0", + "question_id": "2XVR0", + "video_file": "Charades/videos/2XVR0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begin to put dishes away on a shelf'." + }, + { + "from": "gpt", + "value": "From <20.1> to <33.1>." + } + ] + }, + { + "video_id": "9N0BP", + "question_id": "9N0BP", + "video_file": "Charades/videos/9N0BP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person laughing on the phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <22.2>." + } + ] + }, + { + "video_id": "PIEHJ", + "question_id": "PIEHJ", + "video_file": "Charades/videos/PIEHJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they put down the phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.6> to <33.5>." + } + ] + }, + { + "video_id": "PIEHJ", + "question_id": "PIEHJ", + "video_file": "Charades/videos/PIEHJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.0> to <38.0>." + } + ] + }, + { + "video_id": "PIEHJ", + "question_id": "PIEHJ", + "video_file": "Charades/videos/PIEHJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person start eating it'." + }, + { + "from": "gpt", + "value": "From <30.9> to <38.0>." + } + ] + }, + { + "video_id": "OB7UK", + "question_id": "OB7UK", + "video_file": "Charades/videos/OB7UK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts a pillow behind his back'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <26.0>." + } + ] + }, + { + "video_id": "LA0DB", + "question_id": "LA0DB", + "video_file": "Charades/videos/LA0DB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is dressing next to a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <9.7>." + } + ] + }, + { + "video_id": "LA0DB", + "question_id": "LA0DB", + "video_file": "Charades/videos/LA0DB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is throwing a pillow at the sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "LA0DB", + "question_id": "LA0DB", + "video_file": "Charades/videos/LA0DB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is throwing pillows on a sofa' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "LA0DB", + "question_id": "LA0DB", + "video_file": "Charades/videos/LA0DB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turning on a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <9.7>." + } + ] + }, + { + "video_id": "TCV1W", + "question_id": "TCV1W", + "video_file": "Charades/videos/TCV1W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns the light off in the other room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <18.2>." + } + ] + }, + { + "video_id": "TCV1W", + "question_id": "TCV1W", + "video_file": "Charades/videos/TCV1W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they turn off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <18.2>." + } + ] + }, + { + "video_id": "A0MZ9", + "question_id": "A0MZ9", + "video_file": "Charades/videos/A0MZ9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person closes a window'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.2>." + } + ] + }, + { + "video_id": "A0MZ9", + "question_id": "A0MZ9", + "video_file": "Charades/videos/A0MZ9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person closes a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.2>." + } + ] + }, + { + "video_id": "TX122", + "question_id": "TX122", + "video_file": "Charades/videos/TX122.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person awakens in their bedroom under a blanket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <11.5>." + } + ] + }, + { + "video_id": "TX122", + "question_id": "TX122", + "video_file": "Charades/videos/TX122.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person is awakening in bed'." + }, + { + "from": "gpt", + "value": "From <2.5> to <11.5>." + } + ] + }, + { + "video_id": "Q4932", + "question_id": "Q4932", + "video_file": "Charades/videos/Q4932.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws blanket on bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "Q4932", + "question_id": "Q4932", + "video_file": "Charades/videos/Q4932.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person throws a blanket on the bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "Q4932", + "question_id": "Q4932", + "video_file": "Charades/videos/Q4932.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is throwing a blanket on the bed'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "Q4932", + "question_id": "Q4932", + "video_file": "Charades/videos/Q4932.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts eating & laughing'." + }, + { + "from": "gpt", + "value": "From <7.4> to <14.8>." + } + ] + }, + { + "video_id": "EI0F6", + "question_id": "EI0F6", + "video_file": "Charades/videos/EI0F6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.9> to <32.8>." + } + ] + }, + { + "video_id": "EI0F6", + "question_id": "EI0F6", + "video_file": "Charades/videos/EI0F6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.4> to <16.1>." + } + ] + }, + { + "video_id": "EI0F6", + "question_id": "EI0F6", + "video_file": "Charades/videos/EI0F6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the medicine cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <16.1>." + } + ] + }, + { + "video_id": "4UW2K", + "question_id": "4UW2K", + "video_file": "Charades/videos/4UW2K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.9> to <18.0>." + } + ] + }, + { + "video_id": "ZRBO7", + "question_id": "ZRBO7", + "video_file": "Charades/videos/ZRBO7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person sits in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <17.0>." + } + ] + }, + { + "video_id": "ZRBO7", + "question_id": "ZRBO7", + "video_file": "Charades/videos/ZRBO7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <23.0>." + } + ] + }, + { + "video_id": "ZRBO7", + "question_id": "ZRBO7", + "video_file": "Charades/videos/ZRBO7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'watches television moments later the person laughs'." + }, + { + "from": "gpt", + "value": "From <20.0> to <31.0>." + } + ] + }, + { + "video_id": "ZRBO7", + "question_id": "ZRBO7", + "video_file": "Charades/videos/ZRBO7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits down on chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <17.0>." + } + ] + }, + { + "video_id": "03AA8", + "question_id": "03AA8", + "video_file": "Charades/videos/03AA8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person is sitting on sofa'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <17.0>." + } + ] + }, + { + "video_id": "8TDNF", + "question_id": "8TDNF", + "video_file": "Charades/videos/8TDNF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting different clothes on' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "LQO5R", + "question_id": "LQO5R", + "video_file": "Charades/videos/LQO5R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sitting on the sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.0> to <34.0>." + } + ] + }, + { + "video_id": "XVHGO", + "question_id": "XVHGO", + "video_file": "Charades/videos/XVHGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the camera down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <34.0> to <39.0>." + } + ] + }, + { + "video_id": "XVHGO", + "question_id": "XVHGO", + "video_file": "Charades/videos/XVHGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes a camera off of a bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.1> to <10.7>." + } + ] + }, + { + "video_id": "KPAS0", + "question_id": "KPAS0", + "video_file": "Charades/videos/KPAS0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a book off a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.5> to <18.8>." + } + ] + }, + { + "video_id": "KPAS0", + "question_id": "KPAS0", + "video_file": "Charades/videos/KPAS0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes off their shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.6> to <15.9>." + } + ] + }, + { + "video_id": "KPAS0", + "question_id": "KPAS0", + "video_file": "Charades/videos/KPAS0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening a book in the entryway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.0> to <20.0>." + } + ] + }, + { + "video_id": "3XM7Q", + "question_id": "3XM7Q", + "video_file": "Charades/videos/3XM7Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person throws a blanket on the bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.9> to <25.5>." + } + ] + }, + { + "video_id": "3XM7Q", + "question_id": "3XM7Q", + "video_file": "Charades/videos/3XM7Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person is putting a blanket on the bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <25.7>." + } + ] + }, + { + "video_id": "3XM7Q", + "question_id": "3XM7Q", + "video_file": "Charades/videos/3XM7Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting a drink on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <17.1>." + } + ] + }, + { + "video_id": "4JW8U", + "question_id": "4JW8U", + "video_file": "Charades/videos/4JW8U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <12.6>." + } + ] + }, + { + "video_id": "4JW8U", + "question_id": "4JW8U", + "video_file": "Charades/videos/4JW8U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put the towel down on the workbench' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <32.0>." + } + ] + }, + { + "video_id": "4JW8U", + "question_id": "4JW8U", + "video_file": "Charades/videos/4JW8U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.5> to <12.6>." + } + ] + }, + { + "video_id": "C1LVJ", + "question_id": "C1LVJ", + "video_file": "Charades/videos/C1LVJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the pillow down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.8> to <18.6>." + } + ] + }, + { + "video_id": "Y6ROG", + "question_id": "Y6ROG", + "video_file": "Charades/videos/Y6ROG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throw it on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <17.2>." + } + ] + }, + { + "video_id": "B57IW", + "question_id": "B57IW", + "video_file": "Charades/videos/B57IW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a phone out of his pocket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.1> to <17.1>." + } + ] + }, + { + "video_id": "B57IW", + "question_id": "B57IW", + "video_file": "Charades/videos/B57IW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they are opening a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <14.1>." + } + ] + }, + { + "video_id": "B57IW", + "question_id": "B57IW", + "video_file": "Charades/videos/B57IW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person turns on a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.1>." + } + ] + }, + { + "video_id": "B57IW", + "question_id": "B57IW", + "video_file": "Charades/videos/B57IW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding a vacuum' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.0> to <31.0>." + } + ] + }, + { + "video_id": "B57IW", + "question_id": "B57IW", + "video_file": "Charades/videos/B57IW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.2> to <14.1>." + } + ] + }, + { + "video_id": "B57IW", + "question_id": "B57IW", + "video_file": "Charades/videos/B57IW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person turns on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.1>." + } + ] + }, + { + "video_id": "DBT6E", + "question_id": "DBT6E", + "video_file": "Charades/videos/DBT6E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts their laptop bag down on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.6> to <15.1>." + } + ] + }, + { + "video_id": "EQQ2T", + "question_id": "EQQ2T", + "video_file": "Charades/videos/EQQ2T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person put the dryer sheets back on the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <21.8>." + } + ] + }, + { + "video_id": "EQQ2T", + "question_id": "EQQ2T", + "video_file": "Charades/videos/EQQ2T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person started smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.6> to <32.7>." + } + ] + }, + { + "video_id": "IK4Q8", + "question_id": "IK4Q8", + "video_file": "Charades/videos/IK4Q8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person on an old sofa opens a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.8> to <23.8>." + } + ] + }, + { + "video_id": "9OJXP", + "question_id": "9OJXP", + "video_file": "Charades/videos/9OJXP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <17.4>." + } + ] + }, + { + "video_id": "9OJXP", + "question_id": "9OJXP", + "video_file": "Charades/videos/9OJXP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens the door'." + }, + { + "from": "gpt", + "value": "From <5.6> to <17.4>." + } + ] + }, + { + "video_id": "NZY1U", + "question_id": "NZY1U", + "video_file": "Charades/videos/NZY1U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinks from the glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.9> to <28.0>." + } + ] + }, + { + "video_id": "55AH5", + "question_id": "55AH5", + "video_file": "Charades/videos/55AH5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks opens a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "55AH5", + "question_id": "55AH5", + "video_file": "Charades/videos/55AH5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "55AH5", + "question_id": "55AH5", + "video_file": "Charades/videos/55AH5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats some food off a plate' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.1> to <12.0>." + } + ] + }, + { + "video_id": "CS8O0", + "question_id": "CS8O0", + "video_file": "Charades/videos/CS8O0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.8>." + } + ] + }, + { + "video_id": "CS8O0", + "question_id": "CS8O0", + "video_file": "Charades/videos/CS8O0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put a laptop on the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <27.7> to <31.0>." + } + ] + }, + { + "video_id": "CS8O0", + "question_id": "CS8O0", + "video_file": "Charades/videos/CS8O0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the laptop down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.7> to <31.0>." + } + ] + }, + { + "video_id": "9J166", + "question_id": "9J166", + "video_file": "Charades/videos/9J166.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is opening a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.9> to <17.5>." + } + ] + }, + { + "video_id": "9J166", + "question_id": "9J166", + "video_file": "Charades/videos/9J166.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opening a window'." + }, + { + "from": "gpt", + "value": "From <9.9> to <17.5>." + } + ] + }, + { + "video_id": "CXF35", + "question_id": "CXF35", + "video_file": "Charades/videos/CXF35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is seen watching videos on their laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <16.4>." + } + ] + }, + { + "video_id": "CXF35", + "question_id": "CXF35", + "video_file": "Charades/videos/CXF35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person gets up opens a door to get more lighting' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.7> to <20.6>." + } + ] + }, + { + "video_id": "CXF35", + "question_id": "CXF35", + "video_file": "Charades/videos/CXF35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person he sits on his bed'." + }, + { + "from": "gpt", + "value": "From <2.6> to <17.2>." + } + ] + }, + { + "video_id": "CXF35", + "question_id": "CXF35", + "video_file": "Charades/videos/CXF35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person is watching a program on his laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <16.4>." + } + ] + }, + { + "video_id": "73E7V", + "question_id": "73E7V", + "video_file": "Charades/videos/73E7V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they begin eating a bag of chips' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.7> to <28.6>." + } + ] + }, + { + "video_id": "73E7V", + "question_id": "73E7V", + "video_file": "Charades/videos/73E7V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they eat some chips' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.7> to <28.6>." + } + ] + }, + { + "video_id": "L218P", + "question_id": "L218P", + "video_file": "Charades/videos/L218P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person throws a bag into a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.0> to <8.0>." + } + ] + }, + { + "video_id": "L218P", + "question_id": "L218P", + "video_file": "Charades/videos/L218P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person throws a bag into the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <8.0>." + } + ] + }, + { + "video_id": "L218P", + "question_id": "L218P", + "video_file": "Charades/videos/L218P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throwing a bag on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <8.0>." + } + ] + }, + { + "video_id": "L218P", + "question_id": "L218P", + "video_file": "Charades/videos/L218P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person laughs'." + }, + { + "from": "gpt", + "value": "From <10.0> to <17.0>." + } + ] + }, + { + "video_id": "5FYKE", + "question_id": "5FYKE", + "video_file": "Charades/videos/5FYKE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person puts something on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.9> to <8.1>." + } + ] + }, + { + "video_id": "XCUOD", + "question_id": "XCUOD", + "video_file": "Charades/videos/XCUOD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person in a towel is dressing for work'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.5> to <31.8>." + } + ] + }, + { + "video_id": "X3FAP", + "question_id": "X3FAP", + "video_file": "Charades/videos/X3FAP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes a laptop off of a desk'." + }, + { + "from": "gpt", + "value": "From <10.5> to <17.3>." + } + ] + }, + { + "video_id": "PPCWE", + "question_id": "PPCWE", + "video_file": "Charades/videos/PPCWE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is running up the steps' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "O5TU5", + "question_id": "O5TU5", + "video_file": "Charades/videos/O5TU5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'an awakening person goes to the sink' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.9> to <18.4>." + } + ] + }, + { + "video_id": "O5TU5", + "question_id": "O5TU5", + "video_file": "Charades/videos/O5TU5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person lays on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <20.8>." + } + ] + }, + { + "video_id": "O5TU5", + "question_id": "O5TU5", + "video_file": "Charades/videos/O5TU5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person lays down on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.0> to <20.8>." + } + ] + }, + { + "video_id": "KS5VD", + "question_id": "KS5VD", + "video_file": "Charades/videos/KS5VD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looks out a window'." + }, + { + "from": "gpt", + "value": "From <14.0> to <23.6>." + } + ] + }, + { + "video_id": "KS5VD", + "question_id": "KS5VD", + "video_file": "Charades/videos/KS5VD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person look out a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <23.6>." + } + ] + }, + { + "video_id": "36AJ1", + "question_id": "36AJ1", + "video_file": "Charades/videos/36AJ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is undressing in front of a mirror'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.8> to <32.0>." + } + ] + }, + { + "video_id": "36AJ1", + "question_id": "36AJ1", + "video_file": "Charades/videos/36AJ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opened the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.5> to <30.5>." + } + ] + }, + { + "video_id": "IV1U5", + "question_id": "IV1U5", + "video_file": "Charades/videos/IV1U5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person closes the door to the room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <6.5>." + } + ] + }, + { + "video_id": "IV1U5", + "question_id": "IV1U5", + "video_file": "Charades/videos/IV1U5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.3> to <6.5>." + } + ] + }, + { + "video_id": "IV1U5", + "question_id": "IV1U5", + "video_file": "Charades/videos/IV1U5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holds a camera to take a picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <34.2> to <40.0>." + } + ] + }, + { + "video_id": "VG0GK", + "question_id": "VG0GK", + "video_file": "Charades/videos/VG0GK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating breakfast'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "IVP8Q", + "question_id": "IVP8Q", + "video_file": "Charades/videos/IVP8Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person puts papers on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.1> to <32.0>." + } + ] + }, + { + "video_id": "9VZOY", + "question_id": "9VZOY", + "video_file": "Charades/videos/9VZOY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person lying on a pillow awakens in bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.1>." + } + ] + }, + { + "video_id": "9VZOY", + "question_id": "9VZOY", + "video_file": "Charades/videos/9VZOY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drink a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.7> to <30.8>." + } + ] + }, + { + "video_id": "GPUFH", + "question_id": "GPUFH", + "video_file": "Charades/videos/GPUFH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they are tidying up the sink with a towel' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.6> to <19.9>." + } + ] + }, + { + "video_id": "9ITF3", + "question_id": "9ITF3", + "video_file": "Charades/videos/9ITF3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turning the light off behind them' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.3> to <29.3>." + } + ] + }, + { + "video_id": "9ITF3", + "question_id": "9ITF3", + "video_file": "Charades/videos/9ITF3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns off the light'." + }, + { + "from": "gpt", + "value": "From <23.3> to <29.3>." + } + ] + }, + { + "video_id": "9ITF3", + "question_id": "9ITF3", + "video_file": "Charades/videos/9ITF3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts a bag on top of the cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "TD2DR", + "question_id": "TD2DR", + "video_file": "Charades/videos/TD2DR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person someone is sitting in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "TD2DR", + "question_id": "TD2DR", + "video_file": "Charades/videos/TD2DR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take pictures out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.9> to <25.2>." + } + ] + }, + { + "video_id": "TD2DR", + "question_id": "TD2DR", + "video_file": "Charades/videos/TD2DR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take pictures directed towards the windows' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.9> to <25.2>." + } + ] + }, + { + "video_id": "I34L6", + "question_id": "I34L6", + "video_file": "Charades/videos/I34L6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <22.2>." + } + ] + }, + { + "video_id": "7OHQP", + "question_id": "7OHQP", + "video_file": "Charades/videos/7OHQP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person fixing their hair in the mirror right' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <14.2>." + } + ] + }, + { + "video_id": "7OHQP", + "question_id": "7OHQP", + "video_file": "Charades/videos/7OHQP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a phone call'." + }, + { + "from": "gpt", + "value": "From <23.9> to <31.2>." + } + ] + }, + { + "video_id": "SSX6D", + "question_id": "SSX6D", + "video_file": "Charades/videos/SSX6D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person playing on his phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.5> to <26.7>." + } + ] + }, + { + "video_id": "APTS7", + "question_id": "APTS7", + "video_file": "Charades/videos/APTS7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person puts clothes in a washer' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "APTS7", + "question_id": "APTS7", + "video_file": "Charades/videos/APTS7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person putting clothes into a washer' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "APTS7", + "question_id": "APTS7", + "video_file": "Charades/videos/APTS7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a laptop'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.1>." + } + ] + }, + { + "video_id": "APTS7", + "question_id": "APTS7", + "video_file": "Charades/videos/APTS7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins looking at the laptop smiling' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.1> to <15.9>." + } + ] + }, + { + "video_id": "T8O8Y", + "question_id": "T8O8Y", + "video_file": "Charades/videos/T8O8Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes a camera off the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.9> to <26.1>." + } + ] + }, + { + "video_id": "T8O8Y", + "question_id": "T8O8Y", + "video_file": "Charades/videos/T8O8Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting on the bed holding their phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.6>." + } + ] + }, + { + "video_id": "T8O8Y", + "question_id": "T8O8Y", + "video_file": "Charades/videos/T8O8Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person is sitting on the bed tapping his feet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.6>." + } + ] + }, + { + "video_id": "MPX3X", + "question_id": "MPX3X", + "video_file": "Charades/videos/MPX3X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.7>." + } + ] + }, + { + "video_id": "MPX3X", + "question_id": "MPX3X", + "video_file": "Charades/videos/MPX3X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pouring a glass of beer' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.0> to <24.6>." + } + ] + }, + { + "video_id": "MPX3X", + "question_id": "MPX3X", + "video_file": "Charades/videos/MPX3X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person is eating some food'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.7>." + } + ] + }, + { + "video_id": "MPX3X", + "question_id": "MPX3X", + "video_file": "Charades/videos/MPX3X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person finally turning off a light in a garage' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.0> to <31.0>." + } + ] + }, + { + "video_id": "MPX3X", + "question_id": "MPX3X", + "video_file": "Charades/videos/MPX3X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.7>." + } + ] + }, + { + "video_id": "MPX3X", + "question_id": "MPX3X", + "video_file": "Charades/videos/MPX3X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pours some drink in a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.0> to <24.6>." + } + ] + }, + { + "video_id": "WKU7H", + "question_id": "WKU7H", + "video_file": "Charades/videos/WKU7H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts something on a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <18.1>." + } + ] + }, + { + "video_id": "WKU7H", + "question_id": "WKU7H", + "video_file": "Charades/videos/WKU7H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person drinks from a cup of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.6> to <33.5>." + } + ] + }, + { + "video_id": "WKU7H", + "question_id": "WKU7H", + "video_file": "Charades/videos/WKU7H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is closing a cabitnet door'." + }, + { + "from": "gpt", + "value": "From <4.8> to <11.7>." + } + ] + }, + { + "video_id": "1IPYC", + "question_id": "1IPYC", + "video_file": "Charades/videos/1IPYC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person reading a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "TCQXI", + "question_id": "TCQXI", + "video_file": "Charades/videos/TCQXI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "92FFB", + "question_id": "92FFB", + "video_file": "Charades/videos/92FFB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closing the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <16.4>." + } + ] + }, + { + "video_id": "92FFB", + "question_id": "92FFB", + "video_file": "Charades/videos/92FFB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open a cabinet door'." + }, + { + "from": "gpt", + "value": "From <2.0> to <10.8>." + } + ] + }, + { + "video_id": "TIS3Z", + "question_id": "TIS3Z", + "video_file": "Charades/videos/TIS3Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <28.5> to <34.0>." + } + ] + }, + { + "video_id": "P4FXX", + "question_id": "P4FXX", + "video_file": "Charades/videos/P4FXX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turn off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.2> to <34.3>." + } + ] + }, + { + "video_id": "P4FXX", + "question_id": "P4FXX", + "video_file": "Charades/videos/P4FXX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <15.8>." + } + ] + }, + { + "video_id": "SFZ6Q", + "question_id": "SFZ6Q", + "video_file": "Charades/videos/SFZ6Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is dressing next to a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.0>." + } + ] + }, + { + "video_id": "SRHJV", + "question_id": "SRHJV", + "video_file": "Charades/videos/SRHJV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person one is smiling'." + }, + { + "from": "gpt", + "value": "From <27.0> to <33.0>." + } + ] + }, + { + "video_id": "SRHJV", + "question_id": "SRHJV", + "video_file": "Charades/videos/SRHJV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person two people are sitting on a sofa watching television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "SRHJV", + "question_id": "SRHJV", + "video_file": "Charades/videos/SRHJV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person go to the dining room to eat chocolate' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.0> to <33.0>." + } + ] + }, + { + "video_id": "SRHJV", + "question_id": "SRHJV", + "video_file": "Charades/videos/SRHJV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person start eating them' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.0> to <33.0>." + } + ] + }, + { + "video_id": "11DD3", + "question_id": "11DD3", + "video_file": "Charades/videos/11DD3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looking out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "11DD3", + "question_id": "11DD3", + "video_file": "Charades/videos/11DD3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person sits back down on the bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "11DD3", + "question_id": "11DD3", + "video_file": "Charades/videos/11DD3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is closing a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.1> to <11.4>." + } + ] + }, + { + "video_id": "11DD3", + "question_id": "11DD3", + "video_file": "Charades/videos/11DD3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting on a bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "11DD3", + "question_id": "11DD3", + "video_file": "Charades/videos/11DD3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person looks at the picture' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <22.9>." + } + ] + }, + { + "video_id": "4EDGW", + "question_id": "4EDGW", + "video_file": "Charades/videos/4EDGW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person talking on the phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <21.8>." + } + ] + }, + { + "video_id": "E891N", + "question_id": "E891N", + "video_file": "Charades/videos/E891N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a closet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.8> to <27.0>." + } + ] + }, + { + "video_id": "AEHBG", + "question_id": "AEHBG", + "video_file": "Charades/videos/AEHBG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eat a chocolate' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.7> to <37.8>." + } + ] + }, + { + "video_id": "O07KU", + "question_id": "O07KU", + "video_file": "Charades/videos/O07KU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the towel down on the table'." + }, + { + "from": "gpt", + "value": "From <14.4> to <20.5>." + } + ] + }, + { + "video_id": "O07KU", + "question_id": "O07KU", + "video_file": "Charades/videos/O07KU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person uses the towel to tidy up the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.2> to <13.0>." + } + ] + }, + { + "video_id": "O07KU", + "question_id": "O07KU", + "video_file": "Charades/videos/O07KU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the towel on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.4> to <20.5>." + } + ] + }, + { + "video_id": "J10KY", + "question_id": "J10KY", + "video_file": "Charades/videos/J10KY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws it on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.8> to <31.9>." + } + ] + }, + { + "video_id": "J10KY", + "question_id": "J10KY", + "video_file": "Charades/videos/J10KY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throwing their homework on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.8> to <31.9>." + } + ] + }, + { + "video_id": "0PPPL", + "question_id": "0PPPL", + "video_file": "Charades/videos/0PPPL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is walking through a doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.6>." + } + ] + }, + { + "video_id": "0PPPL", + "question_id": "0PPPL", + "video_file": "Charades/videos/0PPPL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <14.1>." + } + ] + }, + { + "video_id": "0PPPL", + "question_id": "0PPPL", + "video_file": "Charades/videos/0PPPL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking a glass of water with one hand' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.8> to <14.1>." + } + ] + }, + { + "video_id": "M2DD1", + "question_id": "M2DD1", + "video_file": "Charades/videos/M2DD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person starts laughing'." + }, + { + "from": "gpt", + "value": "From <25.0> to <31.3>." + } + ] + }, + { + "video_id": "M2DD1", + "question_id": "M2DD1", + "video_file": "Charades/videos/M2DD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.0> to <31.3>." + } + ] + }, + { + "video_id": "M2DD1", + "question_id": "M2DD1", + "video_file": "Charades/videos/M2DD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.0> to <27.4>." + } + ] + }, + { + "video_id": "V9RT3", + "question_id": "V9RT3", + "video_file": "Charades/videos/V9RT3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts a phone on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <21.5>." + } + ] + }, + { + "video_id": "V9RT3", + "question_id": "V9RT3", + "video_file": "Charades/videos/V9RT3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they talk on their phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.9>." + } + ] + }, + { + "video_id": "CVPFH", + "question_id": "CVPFH", + "video_file": "Charades/videos/CVPFH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person turns on a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "CVPFH", + "question_id": "CVPFH", + "video_file": "Charades/videos/CVPFH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts undressing'." + }, + { + "from": "gpt", + "value": "From <18.6> to <34.3>." + } + ] + }, + { + "video_id": "CVPFH", + "question_id": "CVPFH", + "video_file": "Charades/videos/CVPFH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is working on the light in a bathroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "CVPFH", + "question_id": "CVPFH", + "video_file": "Charades/videos/CVPFH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person looking into the mirror,adjust the lights above the mirror' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "CVPFH", + "question_id": "CVPFH", + "video_file": "Charades/videos/CVPFH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting their clothes on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.0> to <34.4>." + } + ] + }, + { + "video_id": "YPTDT", + "question_id": "YPTDT", + "video_file": "Charades/videos/YPTDT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is undressing in a bedroom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <13.0>." + } + ] + }, + { + "video_id": "5XKVP", + "question_id": "5XKVP", + "video_file": "Charades/videos/5XKVP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "5XKVP", + "question_id": "5XKVP", + "video_file": "Charades/videos/5XKVP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person stand up' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.6>." + } + ] + }, + { + "video_id": "5XKVP", + "question_id": "5XKVP", + "video_file": "Charades/videos/5XKVP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting in a chair at a desk'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "5XKVP", + "question_id": "5XKVP", + "video_file": "Charades/videos/5XKVP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throw the shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.7> to <11.5>." + } + ] + }, + { + "video_id": "0JJIY", + "question_id": "0JJIY", + "video_file": "Charades/videos/0JJIY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pours something into a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.8> to <13.6>." + } + ] + }, + { + "video_id": "0JJIY", + "question_id": "0JJIY", + "video_file": "Charades/videos/0JJIY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pours a pitcher of water into a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <13.6>." + } + ] + }, + { + "video_id": "OWS57", + "question_id": "OWS57", + "video_file": "Charades/videos/OWS57.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.2> to <21.1>." + } + ] + }, + { + "video_id": "OWS57", + "question_id": "OWS57", + "video_file": "Charades/videos/OWS57.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person was opening the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.2> to <21.1>." + } + ] + }, + { + "video_id": "OWS57", + "question_id": "OWS57", + "video_file": "Charades/videos/OWS57.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person the other was sneezing right into the mirror'." + }, + { + "from": "gpt", + "value": "From <21.1> to <27.7>." + } + ] + }, + { + "video_id": "XRVQL", + "question_id": "XRVQL", + "video_file": "Charades/videos/XRVQL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws it on the floor shaking their head' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.3> to <18.8>." + } + ] + }, + { + "video_id": "XRVQL", + "question_id": "XRVQL", + "video_file": "Charades/videos/XRVQL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting on the toilet eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.7> to <11.7>." + } + ] + }, + { + "video_id": "XRVQL", + "question_id": "XRVQL", + "video_file": "Charades/videos/XRVQL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throw the sandwich on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.3> to <18.8>." + } + ] + }, + { + "video_id": "XRVQL", + "question_id": "XRVQL", + "video_file": "Charades/videos/XRVQL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws food on the floor'." + }, + { + "from": "gpt", + "value": "From <13.3> to <18.8>." + } + ] + }, + { + "video_id": "ZPAM8", + "question_id": "ZPAM8", + "video_file": "Charades/videos/ZPAM8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes several drinks from the cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.4> to <18.1>." + } + ] + }, + { + "video_id": "JLGLU", + "question_id": "JLGLU", + "video_file": "Charades/videos/JLGLU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a towel'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.3> to <31.0>." + } + ] + }, + { + "video_id": "JLGLU", + "question_id": "JLGLU", + "video_file": "Charades/videos/JLGLU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens their bedroom door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "CVK5Z", + "question_id": "CVK5Z", + "video_file": "Charades/videos/CVK5Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins to undress under a bright light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.7> to <26.8>." + } + ] + }, + { + "video_id": "CVK5Z", + "question_id": "CVK5Z", + "video_file": "Charades/videos/CVK5Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.3> to <14.8>." + } + ] + }, + { + "video_id": "CVK5Z", + "question_id": "CVK5Z", + "video_file": "Charades/videos/CVK5Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person finally they stand up again' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "3UZ88", + "question_id": "3UZ88", + "video_file": "Charades/videos/3UZ88.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person turns off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.5> to <28.0>." + } + ] + }, + { + "video_id": "H4IT6", + "question_id": "H4IT6", + "video_file": "Charades/videos/H4IT6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person walks through a doorway holding a phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "H4IT6", + "question_id": "H4IT6", + "video_file": "Charades/videos/H4IT6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person someone holding a phone'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "00NN7", + "question_id": "00NN7", + "video_file": "Charades/videos/00NN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person puts some groceries away onto the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <10.1>." + } + ] + }, + { + "video_id": "00NN7", + "question_id": "00NN7", + "video_file": "Charades/videos/00NN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <21.0> to <30.1>." + } + ] + }, + { + "video_id": "00NN7", + "question_id": "00NN7", + "video_file": "Charades/videos/00NN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits down in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.3> to <26.4>." + } + ] + }, + { + "video_id": "00NN7", + "question_id": "00NN7", + "video_file": "Charades/videos/00NN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person begins playing with their laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <24.4>." + } + ] + }, + { + "video_id": "00NN7", + "question_id": "00NN7", + "video_file": "Charades/videos/00NN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person plays on a laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <24.4>." + } + ] + }, + { + "video_id": "00NN7", + "question_id": "00NN7", + "video_file": "Charades/videos/00NN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person finally closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.0> to <30.1>." + } + ] + }, + { + "video_id": "2STCL", + "question_id": "2STCL", + "video_file": "Charades/videos/2STCL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <17.7>." + } + ] + }, + { + "video_id": "JF36Q", + "question_id": "JF36Q", + "video_file": "Charades/videos/JF36Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is looking out the window they'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "JF36Q", + "question_id": "JF36Q", + "video_file": "Charades/videos/JF36Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person proceed to close the laptop on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.4> to <31.0>." + } + ] + }, + { + "video_id": "JF36Q", + "question_id": "JF36Q", + "video_file": "Charades/videos/JF36Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.4> to <31.0>." + } + ] + }, + { + "video_id": "2J4MA", + "question_id": "2J4MA", + "video_file": "Charades/videos/2J4MA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they finish eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "2J4MA", + "question_id": "2J4MA", + "video_file": "Charades/videos/2J4MA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begin opening the refrigerator to find more food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <20.6>." + } + ] + }, + { + "video_id": "A6CF6", + "question_id": "A6CF6", + "video_file": "Charades/videos/A6CF6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person throws a box of groceries onto the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "A6CF6", + "question_id": "A6CF6", + "video_file": "Charades/videos/A6CF6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is throwing a box onto a desk'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "A6CF6", + "question_id": "A6CF6", + "video_file": "Charades/videos/A6CF6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <14.0>." + } + ] + }, + { + "video_id": "EOA7I", + "question_id": "EOA7I", + "video_file": "Charades/videos/EOA7I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.1> to <25.7>." + } + ] + }, + { + "video_id": "MI76A", + "question_id": "MI76A", + "video_file": "Charades/videos/MI76A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person close the book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <4.2>." + } + ] + }, + { + "video_id": "MI76A", + "question_id": "MI76A", + "video_file": "Charades/videos/MI76A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person reading a book closes the book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.1> to <4.2>." + } + ] + }, + { + "video_id": "SPG5Q", + "question_id": "SPG5Q", + "video_file": "Charades/videos/SPG5Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes off shoes in the laundry room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.7> to <15.1>." + } + ] + }, + { + "video_id": "1B2P0", + "question_id": "1B2P0", + "video_file": "Charades/videos/1B2P0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts the blanket'." + }, + { + "from": "gpt", + "value": "From <15.4> to <21.0>." + } + ] + }, + { + "video_id": "2OWA8", + "question_id": "2OWA8", + "video_file": "Charades/videos/2OWA8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put it away in a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.9>." + } + ] + }, + { + "video_id": "2OWA8", + "question_id": "2OWA8", + "video_file": "Charades/videos/2OWA8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the bag on one shoulder' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.9>." + } + ] + }, + { + "video_id": "2OWA8", + "question_id": "2OWA8", + "video_file": "Charades/videos/2OWA8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts a laptop into a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <15.2>." + } + ] + }, + { + "video_id": "2ZBL5", + "question_id": "2ZBL5", + "video_file": "Charades/videos/2ZBL5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take a phone from their pocket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.8> to <25.6>." + } + ] + }, + { + "video_id": "0RNRI", + "question_id": "0RNRI", + "video_file": "Charades/videos/0RNRI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person start undressing by taking their jacket off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <30.6>." + } + ] + }, + { + "video_id": "0RNRI", + "question_id": "0RNRI", + "video_file": "Charades/videos/0RNRI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drink a glass of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <23.6>." + } + ] + }, + { + "video_id": "0RNRI", + "question_id": "0RNRI", + "video_file": "Charades/videos/0RNRI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begin undressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <30.6>." + } + ] + }, + { + "video_id": "BZMLV", + "question_id": "BZMLV", + "video_file": "Charades/videos/BZMLV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes some medicine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.6> to <21.5>." + } + ] + }, + { + "video_id": "BZMLV", + "question_id": "BZMLV", + "video_file": "Charades/videos/BZMLV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns on a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.6>." + } + ] + }, + { + "video_id": "BZMLV", + "question_id": "BZMLV", + "video_file": "Charades/videos/BZMLV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person on the phone turns on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.6>." + } + ] + }, + { + "video_id": "SFPJ7", + "question_id": "SFPJ7", + "video_file": "Charades/videos/SFPJ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open a bag with some food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.8> to <20.2>." + } + ] + }, + { + "video_id": "SFPJ7", + "question_id": "SFPJ7", + "video_file": "Charades/videos/SFPJ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put them on a coffee table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.9> to <23.2>." + } + ] + }, + { + "video_id": "SFPJ7", + "question_id": "SFPJ7", + "video_file": "Charades/videos/SFPJ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put some groceries on the coffee table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.9> to <23.2>." + } + ] + }, + { + "video_id": "SFPJ7", + "question_id": "SFPJ7", + "video_file": "Charades/videos/SFPJ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put the blanket aside' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.2> to <13.5>." + } + ] + }, + { + "video_id": "SPDLO", + "question_id": "SPDLO", + "video_file": "Charades/videos/SPDLO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts their dishes in the cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.4> to <42.0>." + } + ] + }, + { + "video_id": "SPDLO", + "question_id": "SPDLO", + "video_file": "Charades/videos/SPDLO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting it into a bag'." + }, + { + "from": "gpt", + "value": "From <21.5> to <27.7>." + } + ] + }, + { + "video_id": "Y2NAO", + "question_id": "Y2NAO", + "video_file": "Charades/videos/Y2NAO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person lying on a blanket on the floor awakens' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "Y2NAO", + "question_id": "Y2NAO", + "video_file": "Charades/videos/Y2NAO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a picture'." + }, + { + "from": "gpt", + "value": "From <15.1> to <23.0>." + } + ] + }, + { + "video_id": "Y2NAO", + "question_id": "Y2NAO", + "video_file": "Charades/videos/Y2NAO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person fixes their hair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <15.6>." + } + ] + }, + { + "video_id": "Y2NAO", + "question_id": "Y2NAO", + "video_file": "Charades/videos/Y2NAO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person fixes their hair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.8> to <15.6>." + } + ] + }, + { + "video_id": "Y2NAO", + "question_id": "Y2NAO", + "video_file": "Charades/videos/Y2NAO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a picture of themselves' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.1> to <23.0>." + } + ] + }, + { + "video_id": "W4MEA", + "question_id": "W4MEA", + "video_file": "Charades/videos/W4MEA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person turns on a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <5.3>." + } + ] + }, + { + "video_id": "W4MEA", + "question_id": "W4MEA", + "video_file": "Charades/videos/W4MEA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they the turn on a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.3> to <5.3>." + } + ] + }, + { + "video_id": "W4MEA", + "question_id": "W4MEA", + "video_file": "Charades/videos/W4MEA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drink a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <23.7>." + } + ] + }, + { + "video_id": "W4MEA", + "question_id": "W4MEA", + "video_file": "Charades/videos/W4MEA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is turning on a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.3> to <5.3>." + } + ] + }, + { + "video_id": "T7T0S", + "question_id": "T7T0S", + "video_file": "Charades/videos/T7T0S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they walk over to a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.8> to <24.0>." + } + ] + }, + { + "video_id": "T7T0S", + "question_id": "T7T0S", + "video_file": "Charades/videos/T7T0S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person continues on to turn on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.8> to <24.0>." + } + ] + }, + { + "video_id": "T7T0S", + "question_id": "T7T0S", + "video_file": "Charades/videos/T7T0S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person cooking food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.3>." + } + ] + }, + { + "video_id": "T7T0S", + "question_id": "T7T0S", + "video_file": "Charades/videos/T7T0S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person turns on an additional light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.8> to <24.0>." + } + ] + }, + { + "video_id": "DNWCR", + "question_id": "DNWCR", + "video_file": "Charades/videos/DNWCR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs into the living room'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "DNWCR", + "question_id": "DNWCR", + "video_file": "Charades/videos/DNWCR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins putting on shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "DNWCR", + "question_id": "DNWCR", + "video_file": "Charades/videos/DNWCR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs through a living room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "DNWCR", + "question_id": "DNWCR", + "video_file": "Charades/videos/DNWCR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts on the shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "DHPNN", + "question_id": "DHPNN", + "video_file": "Charades/videos/DHPNN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person is throwing food at the first person'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <15.5>." + } + ] + }, + { + "video_id": "S6RYI", + "question_id": "S6RYI", + "video_file": "Charades/videos/S6RYI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinks a glass of water'." + }, + { + "from": "gpt", + "value": "From <11.7> to <21.9>." + } + ] + }, + { + "video_id": "IJZI1", + "question_id": "IJZI1", + "video_file": "Charades/videos/IJZI1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks through the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.4>." + } + ] + }, + { + "video_id": "IJZI1", + "question_id": "IJZI1", + "video_file": "Charades/videos/IJZI1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person quickly takes a tissue from the tissue box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.1> to <15.3>." + } + ] + }, + { + "video_id": "HEIW9", + "question_id": "HEIW9", + "video_file": "Charades/videos/HEIW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts a book on the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <10.2>." + } + ] + }, + { + "video_id": "HEIW9", + "question_id": "HEIW9", + "video_file": "Charades/videos/HEIW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person enters their kitchen holding a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "HEIW9", + "question_id": "HEIW9", + "video_file": "Charades/videos/HEIW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the book on top of the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <10.2>." + } + ] + }, + { + "video_id": "1SUIC", + "question_id": "1SUIC", + "video_file": "Charades/videos/1SUIC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <12.2>." + } + ] + }, + { + "video_id": "DWEHS", + "question_id": "DWEHS", + "video_file": "Charades/videos/DWEHS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.1> to <19.7>." + } + ] + }, + { + "video_id": "DWEHS", + "question_id": "DWEHS", + "video_file": "Charades/videos/DWEHS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is smiling into a mirror' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.0>." + } + ] + }, + { + "video_id": "DWEHS", + "question_id": "DWEHS", + "video_file": "Charades/videos/DWEHS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening a book up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.1> to <19.7>." + } + ] + }, + { + "video_id": "VQOBM", + "question_id": "VQOBM", + "video_file": "Charades/videos/VQOBM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opened the cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.3> to <7.0>." + } + ] + }, + { + "video_id": "VQOBM", + "question_id": "VQOBM", + "video_file": "Charades/videos/VQOBM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens cabinet'." + }, + { + "from": "gpt", + "value": "From <0.3> to <7.0>." + } + ] + }, + { + "video_id": "6N7G6", + "question_id": "6N7G6", + "video_file": "Charades/videos/6N7G6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts their dish on the table'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "1L2G1", + "question_id": "1L2G1", + "video_file": "Charades/videos/1L2G1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "JWVW9", + "question_id": "JWVW9", + "video_file": "Charades/videos/JWVW9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a sneezing person awakens in their entryway holding a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.4> to <9.9>." + } + ] + }, + { + "video_id": "55FRY", + "question_id": "55FRY", + "video_file": "Charades/videos/55FRY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person throws a book onto a table'." + }, + { + "from": "gpt", + "value": "From <3.9> to <8.4>." + } + ] + }, + { + "video_id": "55FRY", + "question_id": "55FRY", + "video_file": "Charades/videos/55FRY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person someone throwing a book on a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <8.4>." + } + ] + }, + { + "video_id": "55FRY", + "question_id": "55FRY", + "video_file": "Charades/videos/55FRY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person throws a book onto the dining room table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <8.4>." + } + ] + }, + { + "video_id": "E6PSM", + "question_id": "E6PSM", + "video_file": "Charades/videos/E6PSM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.9> to <9.6>." + } + ] + }, + { + "video_id": "E6PSM", + "question_id": "E6PSM", + "video_file": "Charades/videos/E6PSM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open up the book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.2> to <38.2>." + } + ] + }, + { + "video_id": "E6PSM", + "question_id": "E6PSM", + "video_file": "Charades/videos/E6PSM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.5> to <39.0>." + } + ] + }, + { + "video_id": "E6PSM", + "question_id": "E6PSM", + "video_file": "Charades/videos/E6PSM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person laugh at it'." + }, + { + "from": "gpt", + "value": "From <28.5> to <39.0>." + } + ] + }, + { + "video_id": "E6PSM", + "question_id": "E6PSM", + "video_file": "Charades/videos/E6PSM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they open the box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <27.2>." + } + ] + }, + { + "video_id": "52CKM", + "question_id": "52CKM", + "video_file": "Charades/videos/52CKM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person open a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.9> to <31.6>." + } + ] + }, + { + "video_id": "5DEHA", + "question_id": "5DEHA", + "video_file": "Charades/videos/5DEHA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'other person wakes up on floor sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <13.7>." + } + ] + }, + { + "video_id": "5DEHA", + "question_id": "5DEHA", + "video_file": "Charades/videos/5DEHA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'awakening another person who starts sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.6> to <15.1>." + } + ] + }, + { + "video_id": "AWCQN", + "question_id": "AWCQN", + "video_file": "Charades/videos/AWCQN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person closes the door to their house's entryway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.9> to <32.0>." + } + ] + }, + { + "video_id": "S1W8H", + "question_id": "S1W8H", + "video_file": "Charades/videos/S1W8H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a cabinet in their kitchen'." + }, + { + "from": "gpt", + "value": "From <1.2> to <6.6>." + } + ] + }, + { + "video_id": "ZZXQF", + "question_id": "ZZXQF", + "video_file": "Charades/videos/ZZXQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person eats some medicine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.4> to <16.8>." + } + ] + }, + { + "video_id": "ZZXQF", + "question_id": "ZZXQF", + "video_file": "Charades/videos/ZZXQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes the medicine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <18.2>." + } + ] + }, + { + "video_id": "B0LK0", + "question_id": "B0LK0", + "video_file": "Charades/videos/B0LK0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting away groceries on the pantry shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.2>." + } + ] + }, + { + "video_id": "B0LK0", + "question_id": "B0LK0", + "video_file": "Charades/videos/B0LK0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they begin undressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.0> to <43.9>." + } + ] + }, + { + "video_id": "B0LK0", + "question_id": "B0LK0", + "video_file": "Charades/videos/B0LK0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person get dressed'." + }, + { + "from": "gpt", + "value": "From <25.0> to <43.9>." + } + ] + }, + { + "video_id": "B0LK0", + "question_id": "B0LK0", + "video_file": "Charades/videos/B0LK0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is putting away groceries' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.2>." + } + ] + }, + { + "video_id": "75GKF", + "question_id": "75GKF", + "video_file": "Charades/videos/75GKF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is cooking on the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <20.2>." + } + ] + }, + { + "video_id": "2ZXJ5", + "question_id": "2ZXJ5", + "video_file": "Charades/videos/2ZXJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the water down to sneeze into a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.4> to <18.0>." + } + ] + }, + { + "video_id": "2ZXJ5", + "question_id": "2ZXJ5", + "video_file": "Charades/videos/2ZXJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting on their sofa drinking water'." + }, + { + "from": "gpt", + "value": "From <7.5> to <12.7>." + } + ] + }, + { + "video_id": "2ZXJ5", + "question_id": "2ZXJ5", + "video_file": "Charades/videos/2ZXJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sitting on a sofa drinking water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.5> to <12.7>." + } + ] + }, + { + "video_id": "1UI6I", + "question_id": "1UI6I", + "video_file": "Charades/videos/1UI6I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person getting dressed as they look in the mirror'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.8>." + } + ] + }, + { + "video_id": "1UI6I", + "question_id": "1UI6I", + "video_file": "Charades/videos/1UI6I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is dressing into some clothes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.8>." + } + ] + }, + { + "video_id": "1FGXJ", + "question_id": "1FGXJ", + "video_file": "Charades/videos/1FGXJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a man eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.7>." + } + ] + }, + { + "video_id": "V31YX", + "question_id": "V31YX", + "video_file": "Charades/videos/V31YX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating snack'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <16.9>." + } + ] + }, + { + "video_id": "V31YX", + "question_id": "V31YX", + "video_file": "Charades/videos/V31YX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats some'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <16.9>." + } + ] + }, + { + "video_id": "V31YX", + "question_id": "V31YX", + "video_file": "Charades/videos/V31YX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <7.0> to <16.9>." + } + ] + }, + { + "video_id": "V31YX", + "question_id": "V31YX", + "video_file": "Charades/videos/V31YX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes some food from a plate' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.2> to <10.8>." + } + ] + }, + { + "video_id": "QGQGZ", + "question_id": "QGQGZ", + "video_file": "Charades/videos/QGQGZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens the door into the kitchen' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.4>." + } + ] + }, + { + "video_id": "QGQGZ", + "question_id": "QGQGZ", + "video_file": "Charades/videos/QGQGZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns the light on in'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <15.0>." + } + ] + }, + { + "video_id": "QGQGZ", + "question_id": "QGQGZ", + "video_file": "Charades/videos/QGQGZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns on a light'." + }, + { + "from": "gpt", + "value": "From <10.4> to <15.0>." + } + ] + }, + { + "video_id": "QGQGZ", + "question_id": "QGQGZ", + "video_file": "Charades/videos/QGQGZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person walks through a doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.0> to <9.6>." + } + ] + }, + { + "video_id": "2M2O6", + "question_id": "2M2O6", + "video_file": "Charades/videos/2M2O6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person drinks from a glass'." + }, + { + "from": "gpt", + "value": "From <6.3> to <13.1>." + } + ] + }, + { + "video_id": "EMWAP", + "question_id": "EMWAP", + "video_file": "Charades/videos/EMWAP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person quickly eats a piece of food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.2> to <8.4>." + } + ] + }, + { + "video_id": "EMWAP", + "question_id": "EMWAP", + "video_file": "Charades/videos/EMWAP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <8.4>." + } + ] + }, + { + "video_id": "KV287", + "question_id": "KV287", + "video_file": "Charades/videos/KV287.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a glass of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <11.7>." + } + ] + }, + { + "video_id": "KV287", + "question_id": "KV287", + "video_file": "Charades/videos/KV287.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks through a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <7.9>." + } + ] + }, + { + "video_id": "LUYWY", + "question_id": "LUYWY", + "video_file": "Charades/videos/LUYWY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person reading a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "LUYWY", + "question_id": "LUYWY", + "video_file": "Charades/videos/LUYWY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person someone is sitting on a chair reading a book'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "7P0JX", + "question_id": "7P0JX", + "video_file": "Charades/videos/7P0JX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they take the medicine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.7> to <29.7>." + } + ] + }, + { + "video_id": "7P0JX", + "question_id": "7P0JX", + "video_file": "Charades/videos/7P0JX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person turns on a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.1> to <11.0>." + } + ] + }, + { + "video_id": "7P0JX", + "question_id": "7P0JX", + "video_file": "Charades/videos/7P0JX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.1> to <11.0>." + } + ] + }, + { + "video_id": "7P0JX", + "question_id": "7P0JX", + "video_file": "Charades/videos/7P0JX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closing the door behind them' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.3> to <35.0>." + } + ] + }, + { + "video_id": "7P0JX", + "question_id": "7P0JX", + "video_file": "Charades/videos/7P0JX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person leave as they close the door'." + }, + { + "from": "gpt", + "value": "From <29.3> to <35.0>." + } + ] + }, + { + "video_id": "144JW", + "question_id": "144JW", + "video_file": "Charades/videos/144JW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sneezing into a towel'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "144JW", + "question_id": "144JW", + "video_file": "Charades/videos/144JW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sneezing into a towel repeatedly' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "CRH69", + "question_id": "CRH69", + "video_file": "Charades/videos/CRH69.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throw the clothes on the floor'." + }, + { + "from": "gpt", + "value": "From <20.9> to <25.0>." + } + ] + }, + { + "video_id": "CRH69", + "question_id": "CRH69", + "video_file": "Charades/videos/CRH69.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person the open a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <24.8>." + } + ] + }, + { + "video_id": "CRH69", + "question_id": "CRH69", + "video_file": "Charades/videos/CRH69.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throwing them on the floor of another room'." + }, + { + "from": "gpt", + "value": "From <20.9> to <25.0>." + } + ] + }, + { + "video_id": "Q6ZD9", + "question_id": "Q6ZD9", + "video_file": "Charades/videos/Q6ZD9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is tidying the clothes from the closet doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.6> to <17.8>." + } + ] + }, + { + "video_id": "Q6ZD9", + "question_id": "Q6ZD9", + "video_file": "Charades/videos/Q6ZD9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins sneezing'." + }, + { + "from": "gpt", + "value": "From <16.3> to <23.2>." + } + ] + }, + { + "video_id": "VIKQG", + "question_id": "VIKQG", + "video_file": "Charades/videos/VIKQG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a pair of shoes from under a desk'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.8> to <20.0>." + } + ] + }, + { + "video_id": "5HJ2H", + "question_id": "5HJ2H", + "video_file": "Charades/videos/5HJ2H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "3GFIC", + "question_id": "3GFIC", + "video_file": "Charades/videos/3GFIC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.5> to <21.6>." + } + ] + }, + { + "video_id": "3GFIC", + "question_id": "3GFIC", + "video_file": "Charades/videos/3GFIC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looks out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.7> to <29.0>." + } + ] + }, + { + "video_id": "XBB9P", + "question_id": "XBB9P", + "video_file": "Charades/videos/XBB9P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "9QYHQ", + "question_id": "9QYHQ", + "video_file": "Charades/videos/9QYHQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person walking through the doorway of a room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.1> to <11.7>." + } + ] + }, + { + "video_id": "9QYHQ", + "question_id": "9QYHQ", + "video_file": "Charades/videos/9QYHQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person walks through a doorway'." + }, + { + "from": "gpt", + "value": "From <4.1> to <11.7>." + } + ] + }, + { + "video_id": "QM7Q3", + "question_id": "QM7Q3", + "video_file": "Charades/videos/QM7Q3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a picture of themselves in the mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.6> to <8.2>." + } + ] + }, + { + "video_id": "XKH7W", + "question_id": "XKH7W", + "video_file": "Charades/videos/XKH7W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating some food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.1> to <17.3>." + } + ] + }, + { + "video_id": "XKH7W", + "question_id": "XKH7W", + "video_file": "Charades/videos/XKH7W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they start drinking from a glass of water'." + }, + { + "from": "gpt", + "value": "From <8.1> to <13.6>." + } + ] + }, + { + "video_id": "XKH7W", + "question_id": "XKH7W", + "video_file": "Charades/videos/XKH7W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats food'." + }, + { + "from": "gpt", + "value": "From <12.1> to <17.3>." + } + ] + }, + { + "video_id": "IJNBL", + "question_id": "IJNBL", + "video_file": "Charades/videos/IJNBL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking food on a stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <15.6>." + } + ] + }, + { + "video_id": "IJNBL", + "question_id": "IJNBL", + "video_file": "Charades/videos/IJNBL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they cooked'." + }, + { + "from": "gpt", + "value": "From <0.3> to <15.6>." + } + ] + }, + { + "video_id": "IJNBL", + "question_id": "IJNBL", + "video_file": "Charades/videos/IJNBL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <34.8> to <41.3>." + } + ] + }, + { + "video_id": "IJNBL", + "question_id": "IJNBL", + "video_file": "Charades/videos/IJNBL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person starts cooking on the stove'." + }, + { + "from": "gpt", + "value": "From <0.3> to <15.6>." + } + ] + }, + { + "video_id": "5GG5W", + "question_id": "5GG5W", + "video_file": "Charades/videos/5GG5W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a door to a wardrobe' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "5GG5W", + "question_id": "5GG5W", + "video_file": "Charades/videos/5GG5W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.2>." + } + ] + }, + { + "video_id": "0PCAF", + "question_id": "0PCAF", + "video_file": "Charades/videos/0PCAF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts smiling'." + }, + { + "from": "gpt", + "value": "From <0.8> to <11.5>." + } + ] + }, + { + "video_id": "80ADD", + "question_id": "80ADD", + "video_file": "Charades/videos/80ADD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they begin laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <20.7>." + } + ] + }, + { + "video_id": "5U1IT", + "question_id": "5U1IT", + "video_file": "Charades/videos/5U1IT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.0> to <10.0>." + } + ] + }, + { + "video_id": "5U1IT", + "question_id": "5U1IT", + "video_file": "Charades/videos/5U1IT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person start sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.0> to <10.0>." + } + ] + }, + { + "video_id": "DRT60", + "question_id": "DRT60", + "video_file": "Charades/videos/DRT60.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person washes their hands in the sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <15.4>." + } + ] + }, + { + "video_id": "DRT60", + "question_id": "DRT60", + "video_file": "Charades/videos/DRT60.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person washing their hands at the sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <15.4>." + } + ] + }, + { + "video_id": "VAXUU", + "question_id": "VAXUU", + "video_file": "Charades/videos/VAXUU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closing the door behind themselves'." + }, + { + "from": "gpt", + "value": "From <13.7> to <22.0>." + } + ] + }, + { + "video_id": "VAXUU", + "question_id": "VAXUU", + "video_file": "Charades/videos/VAXUU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person close a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.7> to <22.0>." + } + ] + }, + { + "video_id": "VAXUU", + "question_id": "VAXUU", + "video_file": "Charades/videos/VAXUU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begin closing the closet door'." + }, + { + "from": "gpt", + "value": "From <13.7> to <22.0>." + } + ] + }, + { + "video_id": "USM36", + "question_id": "USM36", + "video_file": "Charades/videos/USM36.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.6>." + } + ] + }, + { + "video_id": "USM36", + "question_id": "USM36", + "video_file": "Charades/videos/USM36.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is opening a laptop'." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.6>." + } + ] + }, + { + "video_id": "AVH53", + "question_id": "AVH53", + "video_file": "Charades/videos/AVH53.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person fixes their hair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.4> to <26.0>." + } + ] + }, + { + "video_id": "9XW1K", + "question_id": "9XW1K", + "video_file": "Charades/videos/9XW1K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they grasp a drinking glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.5> to <28.6>." + } + ] + }, + { + "video_id": "9HNRY", + "question_id": "9HNRY", + "video_file": "Charades/videos/9HNRY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sneezing at the dusty contents' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <9.5>." + } + ] + }, + { + "video_id": "EED7D", + "question_id": "EED7D", + "video_file": "Charades/videos/EED7D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a picture' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.4> to <40.9>." + } + ] + }, + { + "video_id": "EED7D", + "question_id": "EED7D", + "video_file": "Charades/videos/EED7D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes off their wet shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <36.9> to <50.0>." + } + ] + }, + { + "video_id": "EED7D", + "question_id": "EED7D", + "video_file": "Charades/videos/EED7D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes a picture of themself with their phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.4> to <40.9>." + } + ] + }, + { + "video_id": "KQNOV", + "question_id": "KQNOV", + "video_file": "Charades/videos/KQNOV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is undressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.4> to <13.4>." + } + ] + }, + { + "video_id": "KQNOV", + "question_id": "KQNOV", + "video_file": "Charades/videos/KQNOV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is in the entryway of their home undressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <13.4>." + } + ] + }, + { + "video_id": "FQ282", + "question_id": "FQ282", + "video_file": "Charades/videos/FQ282.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person put clothing into a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "FQ282", + "question_id": "FQ282", + "video_file": "Charades/videos/FQ282.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.2> to <15.2>." + } + ] + }, + { + "video_id": "YNRI0", + "question_id": "YNRI0", + "video_file": "Charades/videos/YNRI0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person close the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <23.0>." + } + ] + }, + { + "video_id": "TYZ05", + "question_id": "TYZ05", + "video_file": "Charades/videos/TYZ05.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.7>." + } + ] + }, + { + "video_id": "6VJI0", + "question_id": "6VJI0", + "video_file": "Charades/videos/6VJI0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is smiling at a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.5> to <32.0>." + } + ] + }, + { + "video_id": "S2XJW", + "question_id": "S2XJW", + "video_file": "Charades/videos/S2XJW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits on the sofa' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.3> to <13.2>." + } + ] + }, + { + "video_id": "S2XJW", + "question_id": "S2XJW", + "video_file": "Charades/videos/S2XJW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts eating some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.0> to <14.3>." + } + ] + }, + { + "video_id": "S2XJW", + "question_id": "S2XJW", + "video_file": "Charades/videos/S2XJW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes a drink from a cup'." + }, + { + "from": "gpt", + "value": "From <10.3> to <20.6>." + } + ] + }, + { + "video_id": "S2XJW", + "question_id": "S2XJW", + "video_file": "Charades/videos/S2XJW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is opening a bag of food'." + }, + { + "from": "gpt", + "value": "From <4.8> to <9.8>." + } + ] + }, + { + "video_id": "WXC7V", + "question_id": "WXC7V", + "video_file": "Charades/videos/WXC7V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take a bite of food from the dish' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <32.5> to <74.7>." + } + ] + }, + { + "video_id": "NMG2Z", + "question_id": "NMG2Z", + "video_file": "Charades/videos/NMG2Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is in a basement undressing from their clothes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.0>." + } + ] + }, + { + "video_id": "3U6OZ", + "question_id": "3U6OZ", + "video_file": "Charades/videos/3U6OZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person closes a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <10.6>." + } + ] + }, + { + "video_id": "3U6OZ", + "question_id": "3U6OZ", + "video_file": "Charades/videos/3U6OZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <10.6>." + } + ] + }, + { + "video_id": "3U6OZ", + "question_id": "3U6OZ", + "video_file": "Charades/videos/3U6OZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.6> to <24.9>." + } + ] + }, + { + "video_id": "TXQWJ", + "question_id": "TXQWJ", + "video_file": "Charades/videos/TXQWJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person drinks from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.5>." + } + ] + }, + { + "video_id": "TXQWJ", + "question_id": "TXQWJ", + "video_file": "Charades/videos/TXQWJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking something from a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.5>." + } + ] + }, + { + "video_id": "ZPUVC", + "question_id": "ZPUVC", + "video_file": "Charades/videos/ZPUVC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs back in' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.6>." + } + ] + }, + { + "video_id": "ZPUVC", + "question_id": "ZPUVC", + "video_file": "Charades/videos/ZPUVC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.0>." + } + ] + }, + { + "video_id": "ZPUVC", + "question_id": "ZPUVC", + "video_file": "Charades/videos/ZPUVC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts undressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <30.1>." + } + ] + }, + { + "video_id": "ZPUVC", + "question_id": "ZPUVC", + "video_file": "Charades/videos/ZPUVC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person runs in through the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.6>." + } + ] + }, + { + "video_id": "YK7FT", + "question_id": "YK7FT", + "video_file": "Charades/videos/YK7FT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person struggles to get a door open' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.9> to <8.9>." + } + ] + }, + { + "video_id": "YK7FT", + "question_id": "YK7FT", + "video_file": "Charades/videos/YK7FT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person struggling to open the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.9> to <8.9>." + } + ] + }, + { + "video_id": "YK7FT", + "question_id": "YK7FT", + "video_file": "Charades/videos/YK7FT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person puts running shoes into a garage cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.6> to <14.1>." + } + ] + }, + { + "video_id": "YK7FT", + "question_id": "YK7FT", + "video_file": "Charades/videos/YK7FT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts some shoes behind a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <14.1>." + } + ] + }, + { + "video_id": "OCYT9", + "question_id": "OCYT9", + "video_file": "Charades/videos/OCYT9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is holding an open laptop'." + }, + { + "from": "gpt", + "value": "From <0.1> to <15.2>." + } + ] + }, + { + "video_id": "OCYT9", + "question_id": "OCYT9", + "video_file": "Charades/videos/OCYT9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.9> to <15.8>." + } + ] + }, + { + "video_id": "RANB1", + "question_id": "RANB1", + "video_file": "Charades/videos/RANB1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they throw their clothes on the bed'." + }, + { + "from": "gpt", + "value": "From <19.4> to <25.9>." + } + ] + }, + { + "video_id": "AOQ7C", + "question_id": "AOQ7C", + "video_file": "Charades/videos/AOQ7C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person close the second door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "AOQ7C", + "question_id": "AOQ7C", + "video_file": "Charades/videos/AOQ7C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs down the hallway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <17.4>." + } + ] + }, + { + "video_id": "AOQ7C", + "question_id": "AOQ7C", + "video_file": "Charades/videos/AOQ7C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person run through to a second door'." + }, + { + "from": "gpt", + "value": "From <7.5> to <17.4>." + } + ] + }, + { + "video_id": "AOQ7C", + "question_id": "AOQ7C", + "video_file": "Charades/videos/AOQ7C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person closed the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "AOQ7C", + "question_id": "AOQ7C", + "video_file": "Charades/videos/AOQ7C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "AOQ7C", + "question_id": "AOQ7C", + "video_file": "Charades/videos/AOQ7C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "X7J7H", + "question_id": "X7J7H", + "video_file": "Charades/videos/X7J7H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they stand up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.6> to <15.2>." + } + ] + }, + { + "video_id": "2DTZ3", + "question_id": "2DTZ3", + "video_file": "Charades/videos/2DTZ3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is seen closing a door in the doorway'." + }, + { + "from": "gpt", + "value": "From <0.3> to <7.3>." + } + ] + }, + { + "video_id": "2DTZ3", + "question_id": "2DTZ3", + "video_file": "Charades/videos/2DTZ3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person closes a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.3> to <7.3>." + } + ] + }, + { + "video_id": "00MFE", + "question_id": "00MFE", + "video_file": "Charades/videos/00MFE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take a broom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.7> to <21.5>." + } + ] + }, + { + "video_id": "00MFE", + "question_id": "00MFE", + "video_file": "Charades/videos/00MFE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a broom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <21.5>." + } + ] + }, + { + "video_id": "XWKKZ", + "question_id": "XWKKZ", + "video_file": "Charades/videos/XWKKZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening door to a room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.2> to <6.3>." + } + ] + }, + { + "video_id": "WUL5P", + "question_id": "WUL5P", + "video_file": "Charades/videos/WUL5P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person after opening the door to a doorway'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "WUL5P", + "question_id": "WUL5P", + "video_file": "Charades/videos/WUL5P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person lays down on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.0> to <33.0>." + } + ] + }, + { + "video_id": "OWZCZ", + "question_id": "OWZCZ", + "video_file": "Charades/videos/OWZCZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "OWZCZ", + "question_id": "OWZCZ", + "video_file": "Charades/videos/OWZCZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is opening a refrigerator door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "CFWP7", + "question_id": "CFWP7", + "video_file": "Charades/videos/CFWP7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws their shoes on the ground'." + }, + { + "from": "gpt", + "value": "From <31.7> to <37.0>." + } + ] + }, + { + "video_id": "C5BMA", + "question_id": "C5BMA", + "video_file": "Charades/videos/C5BMA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begin to eat it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.1> to <22.8>." + } + ] + }, + { + "video_id": "C5BMA", + "question_id": "C5BMA", + "video_file": "Charades/videos/C5BMA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person walks through the doorway into the kitchen'." + }, + { + "from": "gpt", + "value": "From <0.7> to <7.5>." + } + ] + }, + { + "video_id": "C5BMA", + "question_id": "C5BMA", + "video_file": "Charades/videos/C5BMA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person the sandwich was put into the sink'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.0> to <34.0>." + } + ] + }, + { + "video_id": "C5BMA", + "question_id": "C5BMA", + "video_file": "Charades/videos/C5BMA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <25.3>." + } + ] + }, + { + "video_id": "2GGH3", + "question_id": "2GGH3", + "video_file": "Charades/videos/2GGH3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person throws a box on the counter' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.2>." + } + ] + }, + { + "video_id": "99UA6", + "question_id": "99UA6", + "video_file": "Charades/videos/99UA6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closed the door of the washer' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <50.0> to <60.0>." + } + ] + }, + { + "video_id": "99UA6", + "question_id": "99UA6", + "video_file": "Charades/videos/99UA6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turned off the light'." + }, + { + "from": "gpt", + "value": "From <52.1> to <59.3>." + } + ] + }, + { + "video_id": "99UA6", + "question_id": "99UA6", + "video_file": "Charades/videos/99UA6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person turns off the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <52.1> to <59.3>." + } + ] + }, + { + "video_id": "99UA6", + "question_id": "99UA6", + "video_file": "Charades/videos/99UA6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes a door'." + }, + { + "from": "gpt", + "value": "From <50.0> to <60.0>." + } + ] + }, + { + "video_id": "H5ZG9", + "question_id": "H5ZG9", + "video_file": "Charades/videos/H5ZG9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person closes the refrigerator'." + }, + { + "from": "gpt", + "value": "From <9.1> to <14.0>." + } + ] + }, + { + "video_id": "H5ZG9", + "question_id": "H5ZG9", + "video_file": "Charades/videos/H5ZG9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "H5ZG9", + "question_id": "H5ZG9", + "video_file": "Charades/videos/H5ZG9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes some food out'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <12.1>." + } + ] + }, + { + "video_id": "H5ZG9", + "question_id": "H5ZG9", + "video_file": "Charades/videos/H5ZG9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes a sandwich out of the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.8> to <12.0>." + } + ] + }, + { + "video_id": "H5ZG9", + "question_id": "H5ZG9", + "video_file": "Charades/videos/H5ZG9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "450TW", + "question_id": "450TW", + "video_file": "Charades/videos/450TW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person sitting on the couch opens a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <8.7>." + } + ] + }, + { + "video_id": "450TW", + "question_id": "450TW", + "video_file": "Charades/videos/450TW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person reading the book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.3> to <21.9>." + } + ] + }, + { + "video_id": "450TW", + "question_id": "450TW", + "video_file": "Charades/videos/450TW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins eating food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.0> to <16.3>." + } + ] + }, + { + "video_id": "450TW", + "question_id": "450TW", + "video_file": "Charades/videos/450TW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they eat some food'." + }, + { + "from": "gpt", + "value": "From <10.0> to <16.3>." + } + ] + }, + { + "video_id": "5J7FQ", + "question_id": "5J7FQ", + "video_file": "Charades/videos/5J7FQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they watch their reflection in the mirror' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.8> to <17.6>." + } + ] + }, + { + "video_id": "0HD0T", + "question_id": "0HD0T", + "video_file": "Charades/videos/0HD0T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws it on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.1> to <28.2>." + } + ] + }, + { + "video_id": "0HD0T", + "question_id": "0HD0T", + "video_file": "Charades/videos/0HD0T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.1> to <23.3>." + } + ] + }, + { + "video_id": "0HD0T", + "question_id": "0HD0T", + "video_file": "Charades/videos/0HD0T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.1> to <23.3>." + } + ] + }, + { + "video_id": "0HD0T", + "question_id": "0HD0T", + "video_file": "Charades/videos/0HD0T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes out a pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <25.3>." + } + ] + }, + { + "video_id": "N0N4I", + "question_id": "N0N4I", + "video_file": "Charades/videos/N0N4I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they open a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.4> to <12.5>." + } + ] + }, + { + "video_id": "N0N4I", + "question_id": "N0N4I", + "video_file": "Charades/videos/N0N4I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is cooking on the stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "N0N4I", + "question_id": "N0N4I", + "video_file": "Charades/videos/N0N4I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take out some medicine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.6> to <19.4>." + } + ] + }, + { + "video_id": "55UAT", + "question_id": "55UAT", + "video_file": "Charades/videos/55UAT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting items on the shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.2>." + } + ] + }, + { + "video_id": "6E7Q4", + "question_id": "6E7Q4", + "video_file": "Charades/videos/6E7Q4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is putting a laptop on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.7> to <23.7>." + } + ] + }, + { + "video_id": "6E7Q4", + "question_id": "6E7Q4", + "video_file": "Charades/videos/6E7Q4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.9> to <32.0>." + } + ] + }, + { + "video_id": "QZ94G", + "question_id": "QZ94G", + "video_file": "Charades/videos/QZ94G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person picks up a camera to play with' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <28.0>." + } + ] + }, + { + "video_id": "QZ94G", + "question_id": "QZ94G", + "video_file": "Charades/videos/QZ94G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.4> to <14.8>." + } + ] + }, + { + "video_id": "QZ94G", + "question_id": "QZ94G", + "video_file": "Charades/videos/QZ94G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person is looking through a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.1>." + } + ] + }, + { + "video_id": "IU5TH", + "question_id": "IU5TH", + "video_file": "Charades/videos/IU5TH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they close the laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <14.2>." + } + ] + }, + { + "video_id": "IU5TH", + "question_id": "IU5TH", + "video_file": "Charades/videos/IU5TH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person closes their laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <14.2>." + } + ] + }, + { + "video_id": "A6JLX", + "question_id": "A6JLX", + "video_file": "Charades/videos/A6JLX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes out a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.6> to <12.7>." + } + ] + }, + { + "video_id": "A6JLX", + "question_id": "A6JLX", + "video_file": "Charades/videos/A6JLX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs into the room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "A6JLX", + "question_id": "A6JLX", + "video_file": "Charades/videos/A6JLX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is running in with dishes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "CGPBZ", + "question_id": "CGPBZ", + "video_file": "Charades/videos/CGPBZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sit down in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.6> to <36.8>." + } + ] + }, + { + "video_id": "CGPBZ", + "question_id": "CGPBZ", + "video_file": "Charades/videos/CGPBZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person start undressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.3> to <36.1>." + } + ] + }, + { + "video_id": "CGPBZ", + "question_id": "CGPBZ", + "video_file": "Charades/videos/CGPBZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off her shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.0> to <29.3>." + } + ] + }, + { + "video_id": "CSLEP", + "question_id": "CSLEP", + "video_file": "Charades/videos/CSLEP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person dress in a shirt' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <22.2>." + } + ] + }, + { + "video_id": "CSLEP", + "question_id": "CSLEP", + "video_file": "Charades/videos/CSLEP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is dressing under a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <22.2>." + } + ] + }, + { + "video_id": "CSLEP", + "question_id": "CSLEP", + "video_file": "Charades/videos/CSLEP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person dress themselves'." + }, + { + "from": "gpt", + "value": "From <3.9> to <22.2>." + } + ] + }, + { + "video_id": "3Z7KH", + "question_id": "3Z7KH", + "video_file": "Charades/videos/3Z7KH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting on a floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.5> to <32.0>." + } + ] + }, + { + "video_id": "3Z7KH", + "question_id": "3Z7KH", + "video_file": "Charades/videos/3Z7KH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sneezing on a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.5> to <20.4>." + } + ] + }, + { + "video_id": "XQDQ4", + "question_id": "XQDQ4", + "video_file": "Charades/videos/XQDQ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is looking out a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.2>." + } + ] + }, + { + "video_id": "XQDQ4", + "question_id": "XQDQ4", + "video_file": "Charades/videos/XQDQ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is seen watching out the window'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.2>." + } + ] + }, + { + "video_id": "DPKMU", + "question_id": "DPKMU", + "video_file": "Charades/videos/DPKMU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is holding a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "DPKMU", + "question_id": "DPKMU", + "video_file": "Charades/videos/DPKMU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person puts down a box by the door'." + }, + { + "from": "gpt", + "value": "From <0.2> to <6.3>." + } + ] + }, + { + "video_id": "9X6DL", + "question_id": "9X6DL", + "video_file": "Charades/videos/9X6DL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person reading a book'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "9X6DL", + "question_id": "9X6DL", + "video_file": "Charades/videos/9X6DL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes a picture of themselves with their phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <8.5>." + } + ] + }, + { + "video_id": "8QL4N", + "question_id": "8QL4N", + "video_file": "Charades/videos/8QL4N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <23.0>." + } + ] + }, + { + "video_id": "8QL4N", + "question_id": "8QL4N", + "video_file": "Charades/videos/8QL4N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they eat the sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.0> to <23.0>." + } + ] + }, + { + "video_id": "40NIM", + "question_id": "40NIM", + "video_file": "Charades/videos/40NIM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person as the are holding it the start laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.0> to <19.1>." + } + ] + }, + { + "video_id": "JELL2", + "question_id": "JELL2", + "video_file": "Charades/videos/JELL2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.2>." + } + ] + }, + { + "video_id": "JELL2", + "question_id": "JELL2", + "video_file": "Charades/videos/JELL2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person smiling in front of a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.2>." + } + ] + }, + { + "video_id": "G71VM", + "question_id": "G71VM", + "video_file": "Charades/videos/G71VM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they first open up cabinets' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.7> to <13.2>." + } + ] + }, + { + "video_id": "G71VM", + "question_id": "G71VM", + "video_file": "Charades/videos/G71VM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking from a glass of water'." + }, + { + "from": "gpt", + "value": "From <26.9> to <33.6>." + } + ] + }, + { + "video_id": "C83CD", + "question_id": "C83CD", + "video_file": "Charades/videos/C83CD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person is holding a grocery bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "51J43", + "question_id": "51J43", + "video_file": "Charades/videos/51J43.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person is running with a bag of food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.1> to <10.6>." + } + ] + }, + { + "video_id": "51J43", + "question_id": "51J43", + "video_file": "Charades/videos/51J43.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person holding a bag runs through the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.1> to <10.6>." + } + ] + }, + { + "video_id": "PQX96", + "question_id": "PQX96", + "video_file": "Charades/videos/PQX96.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person closes a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.2> to <14.8>." + } + ] + }, + { + "video_id": "PQX96", + "question_id": "PQX96", + "video_file": "Charades/videos/PQX96.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person plays with a flashlight' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.5> to <31.0>." + } + ] + }, + { + "video_id": "PQX96", + "question_id": "PQX96", + "video_file": "Charades/videos/PQX96.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits on the couch'." + }, + { + "from": "gpt", + "value": "From <13.2> to <19.2>." + } + ] + }, + { + "video_id": "PQX96", + "question_id": "PQX96", + "video_file": "Charades/videos/PQX96.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is looking out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.1>." + } + ] + }, + { + "video_id": "PQX96", + "question_id": "PQX96", + "video_file": "Charades/videos/PQX96.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sitting in a chair with a flashlight'." + }, + { + "from": "gpt", + "value": "From <20.5> to <31.0>." + } + ] + }, + { + "video_id": "PQX96", + "question_id": "PQX96", + "video_file": "Charades/videos/PQX96.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person lies on the sofa playing with a flashlight' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.5> to <31.0>." + } + ] + }, + { + "video_id": "PPXPA", + "question_id": "PPXPA", + "video_file": "Charades/videos/PPXPA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a paper towel'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <10.5>." + } + ] + }, + { + "video_id": "PPXPA", + "question_id": "PPXPA", + "video_file": "Charades/videos/PPXPA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person sneezing was opening a book by the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <9.2>." + } + ] + }, + { + "video_id": "0SE6A", + "question_id": "0SE6A", + "video_file": "Charades/videos/0SE6A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person upon awakening they look confused'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <16.9>." + } + ] + }, + { + "video_id": "2FW38", + "question_id": "2FW38", + "video_file": "Charades/videos/2FW38.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "2FW38", + "question_id": "2FW38", + "video_file": "Charades/videos/2FW38.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person in their home office is closing their door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "DP7KP", + "question_id": "DP7KP", + "video_file": "Charades/videos/DP7KP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws the book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.2> to <17.2>." + } + ] + }, + { + "video_id": "2O2LG", + "question_id": "2O2LG", + "video_file": "Charades/videos/2O2LG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking a cup of coffee' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "2O2LG", + "question_id": "2O2LG", + "video_file": "Charades/videos/2O2LG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinks coffee from a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "6TJVS", + "question_id": "6TJVS", + "video_file": "Charades/videos/6TJVS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting at a table eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.4> to <9.0>." + } + ] + }, + { + "video_id": "6TJVS", + "question_id": "6TJVS", + "video_file": "Charades/videos/6TJVS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.2> to <9.5>." + } + ] + }, + { + "video_id": "6TJVS", + "question_id": "6TJVS", + "video_file": "Charades/videos/6TJVS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <9.0>." + } + ] + }, + { + "video_id": "UETKQ", + "question_id": "UETKQ", + "video_file": "Charades/videos/UETKQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is putting a sandwhich in a bag'." + }, + { + "from": "gpt", + "value": "From <15.5> to <20.7>." + } + ] + }, + { + "video_id": "UETKQ", + "question_id": "UETKQ", + "video_file": "Charades/videos/UETKQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the bag in a closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.5> to <20.7>." + } + ] + }, + { + "video_id": "UETKQ", + "question_id": "UETKQ", + "video_file": "Charades/videos/UETKQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws the bag inside' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.9> to <19.3>." + } + ] + }, + { + "video_id": "UETKQ", + "question_id": "UETKQ", + "video_file": "Charades/videos/UETKQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'next the person opens the closet door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <23.1>." + } + ] + }, + { + "video_id": "UETKQ", + "question_id": "UETKQ", + "video_file": "Charades/videos/UETKQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.2> to <10.1>." + } + ] + }, + { + "video_id": "UETKQ", + "question_id": "UETKQ", + "video_file": "Charades/videos/UETKQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts two sandwiches inside'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <10.1>." + } + ] + }, + { + "video_id": "UETKQ", + "question_id": "UETKQ", + "video_file": "Charades/videos/UETKQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person closes the door throws their head back laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.4> to <28.2>." + } + ] + }, + { + "video_id": "UETKQ", + "question_id": "UETKQ", + "video_file": "Charades/videos/UETKQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a door'." + }, + { + "from": "gpt", + "value": "From <13.0> to <25.7>." + } + ] + }, + { + "video_id": "UETKQ", + "question_id": "UETKQ", + "video_file": "Charades/videos/UETKQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person laughs at the ceiling' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.4> to <28.2>." + } + ] + }, + { + "video_id": "UETKQ", + "question_id": "UETKQ", + "video_file": "Charades/videos/UETKQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "AGJH7", + "question_id": "AGJH7", + "video_file": "Charades/videos/AGJH7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes some clothes from a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.6> to <28.1>." + } + ] + }, + { + "video_id": "AGJH7", + "question_id": "AGJH7", + "video_file": "Charades/videos/AGJH7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <8.4>." + } + ] + }, + { + "video_id": "A2Z46", + "question_id": "A2Z46", + "video_file": "Charades/videos/A2Z46.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person close the book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.5> to <21.3>." + } + ] + }, + { + "video_id": "A2Z46", + "question_id": "A2Z46", + "video_file": "Charades/videos/A2Z46.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person awakens on the couch' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.5>." + } + ] + }, + { + "video_id": "YWO77", + "question_id": "YWO77", + "video_file": "Charades/videos/YWO77.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is walking through the doorway holding a camera' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <6.1>." + } + ] + }, + { + "video_id": "1MU5L", + "question_id": "1MU5L", + "video_file": "Charades/videos/1MU5L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.2> to <9.4>." + } + ] + }, + { + "video_id": "1MU5L", + "question_id": "1MU5L", + "video_file": "Charades/videos/1MU5L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is opening a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <9.4>." + } + ] + }, + { + "video_id": "IJN3M", + "question_id": "IJN3M", + "video_file": "Charades/videos/IJN3M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person start playing with the camera on their phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.0> to <42.0>." + } + ] + }, + { + "video_id": "UANDB", + "question_id": "UANDB", + "video_file": "Charades/videos/UANDB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts reading a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <27.0>." + } + ] + }, + { + "video_id": "UANDB", + "question_id": "UANDB", + "video_file": "Charades/videos/UANDB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person reading a book in a laundry room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.1> to <27.0>." + } + ] + }, + { + "video_id": "391VK", + "question_id": "391VK", + "video_file": "Charades/videos/391VK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting clothes in a wardrobe' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <20.8>." + } + ] + }, + { + "video_id": "Q1SV4", + "question_id": "Q1SV4", + "video_file": "Charades/videos/Q1SV4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throwing the box around'." + }, + { + "from": "gpt", + "value": "From <1.1> to <7.6>." + } + ] + }, + { + "video_id": "02XLP", + "question_id": "02XLP", + "video_file": "Charades/videos/02XLP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'eating a sandwich in a recreation room / man cave'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.4> to <32.0>." + } + ] + }, + { + "video_id": "02XLP", + "question_id": "02XLP", + "video_file": "Charades/videos/02XLP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they take a sandwich from a plate' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.6> to <20.1>." + } + ] + }, + { + "video_id": "RXIAX", + "question_id": "RXIAX", + "video_file": "Charades/videos/RXIAX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person is walking to box to put groceries in there'." + }, + { + "from": "gpt", + "value": "From <8.9> to <14.8>." + } + ] + }, + { + "video_id": "RXIAX", + "question_id": "RXIAX", + "video_file": "Charades/videos/RXIAX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <23.5>." + } + ] + }, + { + "video_id": "RXIAX", + "question_id": "RXIAX", + "video_file": "Charades/videos/RXIAX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person puts a bag down on the bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <11.4>." + } + ] + }, + { + "video_id": "FN457", + "question_id": "FN457", + "video_file": "Charades/videos/FN457.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looks at a picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.9> to <33.0>." + } + ] + }, + { + "video_id": "FN457", + "question_id": "FN457", + "video_file": "Charades/videos/FN457.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is smiling into a mirror'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "4EM25", + "question_id": "4EM25", + "video_file": "Charades/videos/4EM25.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throwing a pillow in a hallway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.4> to <11.1>." + } + ] + }, + { + "video_id": "4EM25", + "question_id": "4EM25", + "video_file": "Charades/videos/4EM25.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.2> to <12.7>." + } + ] + }, + { + "video_id": "4EM25", + "question_id": "4EM25", + "video_file": "Charades/videos/4EM25.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is smiling into a phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "PSSJ8", + "question_id": "PSSJ8", + "video_file": "Charades/videos/PSSJ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns off the lights' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.1> to <29.7>." + } + ] + }, + { + "video_id": "PSSJ8", + "question_id": "PSSJ8", + "video_file": "Charades/videos/PSSJ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turning off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.1> to <29.7>." + } + ] + }, + { + "video_id": "PSSJ8", + "question_id": "PSSJ8", + "video_file": "Charades/videos/PSSJ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin running towards the desk' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.1> to <10.8>." + } + ] + }, + { + "video_id": "PSSJ8", + "question_id": "PSSJ8", + "video_file": "Charades/videos/PSSJ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they close the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.3> to <14.7>." + } + ] + }, + { + "video_id": "PSSJ8", + "question_id": "PSSJ8", + "video_file": "Charades/videos/PSSJ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.3> to <14.7>." + } + ] + }, + { + "video_id": "0JXN3", + "question_id": "0JXN3", + "video_file": "Charades/videos/0JXN3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person throws clothes on a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.7> to <9.9>." + } + ] + }, + { + "video_id": "0JXN3", + "question_id": "0JXN3", + "video_file": "Charades/videos/0JXN3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sneezing a lot'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <19.2>." + } + ] + }, + { + "video_id": "45P7D", + "question_id": "45P7D", + "video_file": "Charades/videos/45P7D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they close the laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.1> to <18.9>." + } + ] + }, + { + "video_id": "45P7D", + "question_id": "45P7D", + "video_file": "Charades/videos/45P7D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person closes their laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.1> to <18.9>." + } + ] + }, + { + "video_id": "4DZ33", + "question_id": "4DZ33", + "video_file": "Charades/videos/4DZ33.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts a towel in the sink' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <17.4>." + } + ] + }, + { + "video_id": "0L0KS", + "question_id": "0L0KS", + "video_file": "Charades/videos/0L0KS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin talking on the phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.4> to <32.0>." + } + ] + }, + { + "video_id": "ES2QH", + "question_id": "ES2QH", + "video_file": "Charades/videos/ES2QH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.5> to <40.0>." + } + ] + }, + { + "video_id": "ES2QH", + "question_id": "ES2QH", + "video_file": "Charades/videos/ES2QH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.1> to <22.9>." + } + ] + }, + { + "video_id": "ES2QH", + "question_id": "ES2QH", + "video_file": "Charades/videos/ES2QH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sitting on a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <33.5> to <40.0>." + } + ] + }, + { + "video_id": "A3AV0", + "question_id": "A3AV0", + "video_file": "Charades/videos/A3AV0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person proceeds to open the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.5> to <29.0>." + } + ] + }, + { + "video_id": "A3AV0", + "question_id": "A3AV0", + "video_file": "Charades/videos/A3AV0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person reaches for the doorknob to open the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.5> to <29.0>." + } + ] + }, + { + "video_id": "HKY1M", + "question_id": "HKY1M", + "video_file": "Charades/videos/HKY1M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sneezing into a mirror in a hallway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.6> to <19.3>." + } + ] + }, + { + "video_id": "5XAMJ", + "question_id": "5XAMJ", + "video_file": "Charades/videos/5XAMJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person watches a video on their laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.0>." + } + ] + }, + { + "video_id": "5XAMJ", + "question_id": "5XAMJ", + "video_file": "Charades/videos/5XAMJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person begins to laugh'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.0> to <33.0>." + } + ] + }, + { + "video_id": "UO607", + "question_id": "UO607", + "video_file": "Charades/videos/UO607.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put the groceries away' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.4> to <17.8>." + } + ] + }, + { + "video_id": "KKGJ8", + "question_id": "KKGJ8", + "video_file": "Charades/videos/KKGJ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looks out a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.4> to <32.0>." + } + ] + }, + { + "video_id": "KKGJ8", + "question_id": "KKGJ8", + "video_file": "Charades/videos/KKGJ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person watches out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.4> to <32.0>." + } + ] + }, + { + "video_id": "KKGJ8", + "question_id": "KKGJ8", + "video_file": "Charades/videos/KKGJ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.7> to <10.6>." + } + ] + }, + { + "video_id": "KKGJ8", + "question_id": "KKGJ8", + "video_file": "Charades/videos/KKGJ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walks through a doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.8> to <10.8>." + } + ] + }, + { + "video_id": "KKGJ8", + "question_id": "KKGJ8", + "video_file": "Charades/videos/KKGJ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting on a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "7DZY6", + "question_id": "7DZY6", + "video_file": "Charades/videos/7DZY6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person is laying on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "7DZY6", + "question_id": "7DZY6", + "video_file": "Charades/videos/7DZY6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person is walking into the room holding a broom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.2> to <29.9>." + } + ] + }, + { + "video_id": "CF92N", + "question_id": "CF92N", + "video_file": "Charades/videos/CF92N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person wearing glasses walks through a doorway'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "OSIKP", + "question_id": "OSIKP", + "video_file": "Charades/videos/OSIKP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person start eating some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.2> to <27.6>." + } + ] + }, + { + "video_id": "ZFT87", + "question_id": "ZFT87", + "video_file": "Charades/videos/ZFT87.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person take out a picture' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <25.1>." + } + ] + }, + { + "video_id": "ZFT87", + "question_id": "ZFT87", + "video_file": "Charades/videos/ZFT87.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looks at a picture'." + }, + { + "from": "gpt", + "value": "From <21.7> to <30.0>." + } + ] + }, + { + "video_id": "ZFT87", + "question_id": "ZFT87", + "video_file": "Charades/videos/ZFT87.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person the open a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.8> to <18.7>." + } + ] + }, + { + "video_id": "MC45U", + "question_id": "MC45U", + "video_file": "Charades/videos/MC45U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.3> to <37.0>." + } + ] + }, + { + "video_id": "MC45U", + "question_id": "MC45U", + "video_file": "Charades/videos/MC45U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'next the person undresses by taking off a jacket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.4> to <30.5>." + } + ] + }, + { + "video_id": "MC45U", + "question_id": "MC45U", + "video_file": "Charades/videos/MC45U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person watches television' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.3> to <37.0>." + } + ] + }, + { + "video_id": "MC45U", + "question_id": "MC45U", + "video_file": "Charades/videos/MC45U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person watching television' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.3> to <37.0>." + } + ] + }, + { + "video_id": "MC45U", + "question_id": "MC45U", + "video_file": "Charades/videos/MC45U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes off their shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.6> to <8.6>." + } + ] + }, + { + "video_id": "MC45U", + "question_id": "MC45U", + "video_file": "Charades/videos/MC45U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes off shoes'." + }, + { + "from": "gpt", + "value": "From <2.6> to <8.6>." + } + ] + }, + { + "video_id": "BWG9B", + "question_id": "BWG9B", + "video_file": "Charades/videos/BWG9B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person drinks from a cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.2> to <5.3>." + } + ] + }, + { + "video_id": "BWG9B", + "question_id": "BWG9B", + "video_file": "Charades/videos/BWG9B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes a glass of water from a table'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.9>." + } + ] + }, + { + "video_id": "U5M7E", + "question_id": "U5M7E", + "video_file": "Charades/videos/U5M7E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person walking out of the room closing the door'." + }, + { + "from": "gpt", + "value": "From <16.3> to <24.3>." + } + ] + }, + { + "video_id": "SBHID", + "question_id": "SBHID", + "video_file": "Charades/videos/SBHID.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opened the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.7> to <10.7>." + } + ] + }, + { + "video_id": "SBHID", + "question_id": "SBHID", + "video_file": "Charades/videos/SBHID.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person holds a doorknob' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.8>." + } + ] + }, + { + "video_id": "SBHID", + "question_id": "SBHID", + "video_file": "Charades/videos/SBHID.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the door to enter the room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.7> to <10.7>." + } + ] + }, + { + "video_id": "SBHID", + "question_id": "SBHID", + "video_file": "Charades/videos/SBHID.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person one opens up door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.7> to <10.7>." + } + ] + }, + { + "video_id": "SBHID", + "question_id": "SBHID", + "video_file": "Charades/videos/SBHID.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens it takes a bottle out' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.7> to <35.6>." + } + ] + }, + { + "video_id": "97NNW", + "question_id": "97NNW", + "video_file": "Charades/videos/97NNW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is cooking on a stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.2>." + } + ] + }, + { + "video_id": "97NNW", + "question_id": "97NNW", + "video_file": "Charades/videos/97NNW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is cooking food on the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.2>." + } + ] + }, + { + "video_id": "97NNW", + "question_id": "97NNW", + "video_file": "Charades/videos/97NNW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is putting dishes in the sink' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.8> to <31.0>." + } + ] + }, + { + "video_id": "1KFLX", + "question_id": "1KFLX", + "video_file": "Charades/videos/1KFLX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.0> to <20.6>." + } + ] + }, + { + "video_id": "1KFLX", + "question_id": "1KFLX", + "video_file": "Charades/videos/1KFLX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closing a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.0> to <20.6>." + } + ] + }, + { + "video_id": "1KFLX", + "question_id": "1KFLX", + "video_file": "Charades/videos/1KFLX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they get dressed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.2> to <14.8>." + } + ] + }, + { + "video_id": "MI4SN", + "question_id": "MI4SN", + "video_file": "Charades/videos/MI4SN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person look out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.7> to <46.0>." + } + ] + }, + { + "video_id": "MI4SN", + "question_id": "MI4SN", + "video_file": "Charades/videos/MI4SN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is awakening on a sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.0> to <27.0>." + } + ] + }, + { + "video_id": "L09BS", + "question_id": "L09BS", + "video_file": "Charades/videos/L09BS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'opening a door in a recreation room / man cave' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.1> to <27.0>." + } + ] + }, + { + "video_id": "L09BS", + "question_id": "L09BS", + "video_file": "Charades/videos/L09BS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opens a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.1> to <27.0>." + } + ] + }, + { + "video_id": "6YH19", + "question_id": "6YH19", + "video_file": "Charades/videos/6YH19.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person drinks from the bottle of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.1> to <23.9>." + } + ] + }, + { + "video_id": "6YH19", + "question_id": "6YH19", + "video_file": "Charades/videos/6YH19.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sitting on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "FV9AL", + "question_id": "FV9AL", + "video_file": "Charades/videos/FV9AL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is opening a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "8WZRF", + "question_id": "8WZRF", + "video_file": "Charades/videos/8WZRF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person off a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.2> to <16.9>." + } + ] + }, + { + "video_id": "8WZRF", + "question_id": "8WZRF", + "video_file": "Charades/videos/8WZRF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns the light off'." + }, + { + "from": "gpt", + "value": "From <10.2> to <16.9>." + } + ] + }, + { + "video_id": "8WZRF", + "question_id": "8WZRF", + "video_file": "Charades/videos/8WZRF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns on the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.7> to <10.4>." + } + ] + }, + { + "video_id": "8WZRF", + "question_id": "8WZRF", + "video_file": "Charades/videos/8WZRF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take a drink of water from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.9> to <32.8>." + } + ] + }, + { + "video_id": "8WZRF", + "question_id": "8WZRF", + "video_file": "Charades/videos/8WZRF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns light back off' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.2> to <16.9>." + } + ] + }, + { + "video_id": "8WZRF", + "question_id": "8WZRF", + "video_file": "Charades/videos/8WZRF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turning the light back on' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <10.4>." + } + ] + }, + { + "video_id": "WEJRX", + "question_id": "WEJRX", + "video_file": "Charades/videos/WEJRX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person laughs as they walk down the hallway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <10.9>." + } + ] + }, + { + "video_id": "WEJRX", + "question_id": "WEJRX", + "video_file": "Charades/videos/WEJRX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person laughing together' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <10.9>." + } + ] + }, + { + "video_id": "WEJRX", + "question_id": "WEJRX", + "video_file": "Charades/videos/WEJRX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.4> to <16.4>." + } + ] + }, + { + "video_id": "7B1CR", + "question_id": "7B1CR", + "video_file": "Charades/videos/7B1CR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sneezing at their phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.3> to <17.0>." + } + ] + }, + { + "video_id": "7B1CR", + "question_id": "7B1CR", + "video_file": "Charades/videos/7B1CR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking out of a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <29.0>." + } + ] + }, + { + "video_id": "VX39P", + "question_id": "VX39P", + "video_file": "Charades/videos/VX39P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.3> to <34.1>." + } + ] + }, + { + "video_id": "VX39P", + "question_id": "VX39P", + "video_file": "Charades/videos/VX39P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they close the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.3> to <34.1>." + } + ] + }, + { + "video_id": "TSYR3", + "question_id": "TSYR3", + "video_file": "Charades/videos/TSYR3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.1>." + } + ] + }, + { + "video_id": "TSYR3", + "question_id": "TSYR3", + "video_file": "Charades/videos/TSYR3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they close the cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.4> to <20.5>." + } + ] + }, + { + "video_id": "ANXB8", + "question_id": "ANXB8", + "video_file": "Charades/videos/ANXB8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating some food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "ANXB8", + "question_id": "ANXB8", + "video_file": "Charades/videos/ANXB8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walk through the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.3> to <22.1>." + } + ] + }, + { + "video_id": "ANXB8", + "question_id": "ANXB8", + "video_file": "Charades/videos/ANXB8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is drinking a bottle of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <14.1>." + } + ] + }, + { + "video_id": "ANXB8", + "question_id": "ANXB8", + "video_file": "Charades/videos/ANXB8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person stands in a doorway eating' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "9OH7W", + "question_id": "9OH7W", + "video_file": "Charades/videos/9OH7W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats a snack'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.3> to <31.5>." + } + ] + }, + { + "video_id": "W1VO2", + "question_id": "W1VO2", + "video_file": "Charades/videos/W1VO2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks into the bathroom holding a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <8.2>." + } + ] + }, + { + "video_id": "W1VO2", + "question_id": "W1VO2", + "video_file": "Charades/videos/W1VO2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person reads a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.5> to <13.1>." + } + ] + }, + { + "video_id": "W1VO2", + "question_id": "W1VO2", + "video_file": "Charades/videos/W1VO2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <5.9>." + } + ] + }, + { + "video_id": "W1VO2", + "question_id": "W1VO2", + "video_file": "Charades/videos/W1VO2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person looks at a book'." + }, + { + "from": "gpt", + "value": "From <4.5> to <13.1>." + } + ] + }, + { + "video_id": "W1VO2", + "question_id": "W1VO2", + "video_file": "Charades/videos/W1VO2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens the door'." + }, + { + "from": "gpt", + "value": "From <0.1> to <7.4>." + } + ] + }, + { + "video_id": "DJVZQ", + "question_id": "DJVZQ", + "video_file": "Charades/videos/DJVZQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours milk into a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.1> to <34.1>." + } + ] + }, + { + "video_id": "DJVZQ", + "question_id": "DJVZQ", + "video_file": "Charades/videos/DJVZQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person pours a cup of coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.1> to <34.1>." + } + ] + }, + { + "video_id": "BE51K", + "question_id": "BE51K", + "video_file": "Charades/videos/BE51K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looking out the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.0>." + } + ] + }, + { + "video_id": "BE51K", + "question_id": "BE51K", + "video_file": "Charades/videos/BE51K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person holding a bag looks out a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.0>." + } + ] + }, + { + "video_id": "BE51K", + "question_id": "BE51K", + "video_file": "Charades/videos/BE51K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits down on their bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <19.8>." + } + ] + }, + { + "video_id": "BE51K", + "question_id": "BE51K", + "video_file": "Charades/videos/BE51K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts down their bag of groceries'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "8C8FI", + "question_id": "8C8FI", + "video_file": "Charades/videos/8C8FI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looks out a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.8> to <29.0>." + } + ] + }, + { + "video_id": "BBQIS", + "question_id": "BBQIS", + "video_file": "Charades/videos/BBQIS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes a pair of shoes from a box'." + }, + { + "from": "gpt", + "value": "From <11.8> to <23.7>." + } + ] + }, + { + "video_id": "KZW3K", + "question_id": "KZW3K", + "video_file": "Charades/videos/KZW3K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the picture down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <17.9>." + } + ] + }, + { + "video_id": "EGNZB", + "question_id": "EGNZB", + "video_file": "Charades/videos/EGNZB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person put the bag down on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <20.0>." + } + ] + }, + { + "video_id": "C4HHB", + "question_id": "C4HHB", + "video_file": "Charades/videos/C4HHB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <17.0>." + } + ] + }, + { + "video_id": "LQEB5", + "question_id": "LQEB5", + "video_file": "Charades/videos/LQEB5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they throw their clothes in a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.5> to <11.8>." + } + ] + }, + { + "video_id": "LQEB5", + "question_id": "LQEB5", + "video_file": "Charades/videos/LQEB5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is seen undressing in the laundry room'." + }, + { + "from": "gpt", + "value": "From <1.5> to <9.6>." + } + ] + }, + { + "video_id": "VUOQF", + "question_id": "VUOQF", + "video_file": "Charades/videos/VUOQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a cabinet'." + }, + { + "from": "gpt", + "value": "From <5.5> to <12.2>." + } + ] + }, + { + "video_id": "VUOQF", + "question_id": "VUOQF", + "video_file": "Charades/videos/VUOQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes some food out of the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <12.8>." + } + ] + }, + { + "video_id": "VUOQF", + "question_id": "VUOQF", + "video_file": "Charades/videos/VUOQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person walks through a doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.1> to <5.5>." + } + ] + }, + { + "video_id": "R4ENI", + "question_id": "R4ENI", + "video_file": "Charades/videos/R4ENI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eat a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <22.7>." + } + ] + }, + { + "video_id": "R4ENI", + "question_id": "R4ENI", + "video_file": "Charades/videos/R4ENI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <15.0> to <22.7>." + } + ] + }, + { + "video_id": "R4ENI", + "question_id": "R4ENI", + "video_file": "Charades/videos/R4ENI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person holding a laptop opens window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.8>." + } + ] + }, + { + "video_id": "R4ENI", + "question_id": "R4ENI", + "video_file": "Charades/videos/R4ENI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is holding a laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.8>." + } + ] + }, + { + "video_id": "R4ENI", + "question_id": "R4ENI", + "video_file": "Charades/videos/R4ENI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looks out the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <11.6>." + } + ] + }, + { + "video_id": "77KWE", + "question_id": "77KWE", + "video_file": "Charades/videos/77KWE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person throws a book onto the pantry shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.4> to <15.8>." + } + ] + }, + { + "video_id": "77KWE", + "question_id": "77KWE", + "video_file": "Charades/videos/77KWE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person throws the book on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <15.8>." + } + ] + }, + { + "video_id": "TON6G", + "question_id": "TON6G", + "video_file": "Charades/videos/TON6G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person runs up the stairs' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <25.6>." + } + ] + }, + { + "video_id": "TON6G", + "question_id": "TON6G", + "video_file": "Charades/videos/TON6G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs out of the garage' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <25.6>." + } + ] + }, + { + "video_id": "R1DCJ", + "question_id": "R1DCJ", + "video_file": "Charades/videos/R1DCJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws a phone onto the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <15.7>." + } + ] + }, + { + "video_id": "R1DCJ", + "question_id": "R1DCJ", + "video_file": "Charades/videos/R1DCJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throwing candy on floor'." + }, + { + "from": "gpt", + "value": "From <10.4> to <15.7>." + } + ] + }, + { + "video_id": "50N4E", + "question_id": "50N4E", + "video_file": "Charades/videos/50N4E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turning off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.4> to <26.8>." + } + ] + }, + { + "video_id": "50N4E", + "question_id": "50N4E", + "video_file": "Charades/videos/50N4E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns off the light'." + }, + { + "from": "gpt", + "value": "From <22.4> to <26.8>." + } + ] + }, + { + "video_id": "26YPV", + "question_id": "26YPV", + "video_file": "Charades/videos/26YPV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the cabinet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "VQN81", + "question_id": "VQN81", + "video_file": "Charades/videos/VQN81.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sneezing on their food'." + }, + { + "from": "gpt", + "value": "From <8.6> to <14.7>." + } + ] + }, + { + "video_id": "3YY88", + "question_id": "3YY88", + "video_file": "Charades/videos/3YY88.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person washes the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <22.7>." + } + ] + }, + { + "video_id": "3YY88", + "question_id": "3YY88", + "video_file": "Charades/videos/3YY88.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening cabinet to get a cloth' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.7>." + } + ] + }, + { + "video_id": "80YWL", + "question_id": "80YWL", + "video_file": "Charades/videos/80YWL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person puts a glass on a table'." + }, + { + "from": "gpt", + "value": "From <2.5> to <8.1>." + } + ] + }, + { + "video_id": "80YWL", + "question_id": "80YWL", + "video_file": "Charades/videos/80YWL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the closet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.7> to <14.0>." + } + ] + }, + { + "video_id": "80YWL", + "question_id": "80YWL", + "video_file": "Charades/videos/80YWL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the cup down on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.5> to <8.1>." + } + ] + }, + { + "video_id": "80YWL", + "question_id": "80YWL", + "video_file": "Charades/videos/80YWL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is drinking from a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "M9KMN", + "question_id": "M9KMN", + "video_file": "Charades/videos/M9KMN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person walks in holding another book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <32.0>." + } + ] + }, + { + "video_id": "M9KMN", + "question_id": "M9KMN", + "video_file": "Charades/videos/M9KMN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <32.0>." + } + ] + }, + { + "video_id": "VKXLL", + "question_id": "VKXLL", + "video_file": "Charades/videos/VKXLL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.1> to <28.6>." + } + ] + }, + { + "video_id": "GAM0E", + "question_id": "GAM0E", + "video_file": "Charades/videos/GAM0E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts dishes on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <8.7>." + } + ] + }, + { + "video_id": "GAM0E", + "question_id": "GAM0E", + "video_file": "Charades/videos/GAM0E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting dishes onto a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.1> to <8.7>." + } + ] + }, + { + "video_id": "68QO4", + "question_id": "68QO4", + "video_file": "Charades/videos/68QO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a sandwich from the cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.3> to <23.0>." + } + ] + }, + { + "video_id": "68QO4", + "question_id": "68QO4", + "video_file": "Charades/videos/68QO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.6> to <31.3>." + } + ] + }, + { + "video_id": "68QO4", + "question_id": "68QO4", + "video_file": "Charades/videos/68QO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the book on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <20.1>." + } + ] + }, + { + "video_id": "68QO4", + "question_id": "68QO4", + "video_file": "Charades/videos/68QO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person closes the book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.2> to <16.9>." + } + ] + }, + { + "video_id": "5T607", + "question_id": "5T607", + "video_file": "Charades/videos/5T607.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person is running to the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.1> to <34.8>." + } + ] + }, + { + "video_id": "5T607", + "question_id": "5T607", + "video_file": "Charades/videos/5T607.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <29.5> to <37.9>." + } + ] + }, + { + "video_id": "W4U3V", + "question_id": "W4U3V", + "video_file": "Charades/videos/W4U3V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eat a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.9> to <33.4>." + } + ] + }, + { + "video_id": "W4U3V", + "question_id": "W4U3V", + "video_file": "Charades/videos/W4U3V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts a glass down on a table'." + }, + { + "from": "gpt", + "value": "From <8.0> to <20.1>." + } + ] + }, + { + "video_id": "5D1FD", + "question_id": "5D1FD", + "video_file": "Charades/videos/5D1FD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is eating some food from a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.1> to <17.4>." + } + ] + }, + { + "video_id": "5D1FD", + "question_id": "5D1FD", + "video_file": "Charades/videos/5D1FD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turn on a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <29.9>." + } + ] + }, + { + "video_id": "5D1FD", + "question_id": "5D1FD", + "video_file": "Charades/videos/5D1FD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they turn on a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <29.9>." + } + ] + }, + { + "video_id": "W8TE5", + "question_id": "W8TE5", + "video_file": "Charades/videos/W8TE5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes a picture of themselves with the phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "FU5AP", + "question_id": "FU5AP", + "video_file": "Charades/videos/FU5AP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person he puts on a pair of shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.2> to <34.0>." + } + ] + }, + { + "video_id": "FU5AP", + "question_id": "FU5AP", + "video_file": "Charades/videos/FU5AP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person walking to go put on her shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.2> to <34.0>." + } + ] + }, + { + "video_id": "FU5AP", + "question_id": "FU5AP", + "video_file": "Charades/videos/FU5AP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts on shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.2> to <34.0>." + } + ] + }, + { + "video_id": "WNYV6", + "question_id": "WNYV6", + "video_file": "Charades/videos/WNYV6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they open a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <15.6>." + } + ] + }, + { + "video_id": "WNYV6", + "question_id": "WNYV6", + "video_file": "Charades/videos/WNYV6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person close the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <18.7>." + } + ] + }, + { + "video_id": "WNYV6", + "question_id": "WNYV6", + "video_file": "Charades/videos/WNYV6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put it on a shelf in a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.5> to <16.3>." + } + ] + }, + { + "video_id": "WNYV6", + "question_id": "WNYV6", + "video_file": "Charades/videos/WNYV6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person puts a laptop into the cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <18.4>." + } + ] + }, + { + "video_id": "PN4MI", + "question_id": "PN4MI", + "video_file": "Charades/videos/PN4MI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.9> to <13.1>." + } + ] + }, + { + "video_id": "PN4MI", + "question_id": "PN4MI", + "video_file": "Charades/videos/PN4MI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'two persons are eating'." + }, + { + "from": "gpt", + "value": "From <17.7> to <25.4>." + } + ] + }, + { + "video_id": "PN4MI", + "question_id": "PN4MI", + "video_file": "Charades/videos/PN4MI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'laughing in the doorway as another person awakens with homework'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.6> to <29.9>." + } + ] + }, + { + "video_id": "GPSIM", + "question_id": "GPSIM", + "video_file": "Charades/videos/GPSIM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts food on a plate'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <7.7>." + } + ] + }, + { + "video_id": "GPSIM", + "question_id": "GPSIM", + "video_file": "Charades/videos/GPSIM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'where person is fixing something to eat' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.4> to <29.8>." + } + ] + }, + { + "video_id": "GPSIM", + "question_id": "GPSIM", + "video_file": "Charades/videos/GPSIM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits on a sofa'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.1> to <31.8>." + } + ] + }, + { + "video_id": "GPSIM", + "question_id": "GPSIM", + "video_file": "Charades/videos/GPSIM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eat the food'." + }, + { + "from": "gpt", + "value": "From <18.4> to <29.8>." + } + ] + }, + { + "video_id": "GPSIM", + "question_id": "GPSIM", + "video_file": "Charades/videos/GPSIM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sat on the sofa to watch television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.3> to <29.5>." + } + ] + }, + { + "video_id": "GPSIM", + "question_id": "GPSIM", + "video_file": "Charades/videos/GPSIM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person decides on eating in there instead'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.4> to <29.8>." + } + ] + }, + { + "video_id": "GPSIM", + "question_id": "GPSIM", + "video_file": "Charades/videos/GPSIM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person watches television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.3> to <29.5>." + } + ] + }, + { + "video_id": "7GF4E", + "question_id": "7GF4E", + "video_file": "Charades/videos/7GF4E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts the bag down on the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.0> to <25.0>." + } + ] + }, + { + "video_id": "ETS92", + "question_id": "ETS92", + "video_file": "Charades/videos/ETS92.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they open another door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.7> to <19.8>." + } + ] + }, + { + "video_id": "ETS92", + "question_id": "ETS92", + "video_file": "Charades/videos/ETS92.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person closes a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.9> to <16.5>." + } + ] + }, + { + "video_id": "ETS92", + "question_id": "ETS92", + "video_file": "Charades/videos/ETS92.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person someone closes a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.9> to <16.5>." + } + ] + }, + { + "video_id": "ETS92", + "question_id": "ETS92", + "video_file": "Charades/videos/ETS92.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opening a different door'." + }, + { + "from": "gpt", + "value": "From <14.7> to <19.8>." + } + ] + }, + { + "video_id": "ETS92", + "question_id": "ETS92", + "video_file": "Charades/videos/ETS92.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits down on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.6> to <32.0>." + } + ] + }, + { + "video_id": "QIBAJ", + "question_id": "QIBAJ", + "video_file": "Charades/videos/QIBAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <27.8>." + } + ] + }, + { + "video_id": "QIBAJ", + "question_id": "QIBAJ", + "video_file": "Charades/videos/QIBAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the glass back on the floor,stands' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.5> to <27.8>." + } + ] + }, + { + "video_id": "QIBAJ", + "question_id": "QIBAJ", + "video_file": "Charades/videos/QIBAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take a sip of water from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.7> to <19.2>." + } + ] + }, + { + "video_id": "K8ZIK", + "question_id": "K8ZIK", + "video_file": "Charades/videos/K8ZIK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person cooking at the stove in a kitchen' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.1> to <30.0>." + } + ] + }, + { + "video_id": "K8ZIK", + "question_id": "K8ZIK", + "video_file": "Charades/videos/K8ZIK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they pour a glass of milk'." + }, + { + "from": "gpt", + "value": "From <7.3> to <15.2>." + } + ] + }, + { + "video_id": "7RI1Y", + "question_id": "7RI1Y", + "video_file": "Charades/videos/7RI1Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens cabinet doors'." + }, + { + "from": "gpt", + "value": "From <19.0> to <26.1>." + } + ] + }, + { + "video_id": "DB3Y6", + "question_id": "DB3Y6", + "video_file": "Charades/videos/DB3Y6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person puts some food on a plate'." + }, + { + "from": "gpt", + "value": "From <1.4> to <7.1>." + } + ] + }, + { + "video_id": "DB3Y6", + "question_id": "DB3Y6", + "video_file": "Charades/videos/DB3Y6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person is putting food on a plate' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <7.1>." + } + ] + }, + { + "video_id": "0DAO5", + "question_id": "0DAO5", + "video_file": "Charades/videos/0DAO5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts away the bag on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.9> to <14.7>." + } + ] + }, + { + "video_id": "0DAO5", + "question_id": "0DAO5", + "video_file": "Charades/videos/0DAO5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.3> to <28.2>." + } + ] + }, + { + "video_id": "G3TEA", + "question_id": "G3TEA", + "video_file": "Charades/videos/G3TEA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is tidying up some medicine on the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <17.2>." + } + ] + }, + { + "video_id": "G3TEA", + "question_id": "G3TEA", + "video_file": "Charades/videos/G3TEA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turning off the light'." + }, + { + "from": "gpt", + "value": "From <38.6> to <43.0>." + } + ] + }, + { + "video_id": "G3TEA", + "question_id": "G3TEA", + "video_file": "Charades/videos/G3TEA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person washes hands'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "G3TEA", + "question_id": "G3TEA", + "video_file": "Charades/videos/G3TEA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they wash their hands' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "OMX6W", + "question_id": "OMX6W", + "video_file": "Charades/videos/OMX6W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a closet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.3> to <7.2>." + } + ] + }, + { + "video_id": "OMX6W", + "question_id": "OMX6W", + "video_file": "Charades/videos/OMX6W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts opening the closet door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <7.2>." + } + ] + }, + { + "video_id": "P81OP", + "question_id": "P81OP", + "video_file": "Charades/videos/P81OP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person reading a book'." + }, + { + "from": "gpt", + "value": "From <9.4> to <22.9>." + } + ] + }, + { + "video_id": "P81OP", + "question_id": "P81OP", + "video_file": "Charades/videos/P81OP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'as another person is lying own reading a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.4> to <22.9>." + } + ] + }, + { + "video_id": "31TDO", + "question_id": "31TDO", + "video_file": "Charades/videos/31TDO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns on light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.6> to <29.1>." + } + ] + }, + { + "video_id": "31TDO", + "question_id": "31TDO", + "video_file": "Charades/videos/31TDO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits down at the table'." + }, + { + "from": "gpt", + "value": "From <37.4> to <49.7>." + } + ] + }, + { + "video_id": "31TDO", + "question_id": "31TDO", + "video_file": "Charades/videos/31TDO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns on the light'." + }, + { + "from": "gpt", + "value": "From <12.6> to <29.1>." + } + ] + }, + { + "video_id": "31TDO", + "question_id": "31TDO", + "video_file": "Charades/videos/31TDO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits down at a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <37.4> to <49.7>." + } + ] + }, + { + "video_id": "31TDO", + "question_id": "31TDO", + "video_file": "Charades/videos/31TDO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats some food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <32.7>." + } + ] + }, + { + "video_id": "31TDO", + "question_id": "31TDO", + "video_file": "Charades/videos/31TDO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person turns on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.6> to <29.1>." + } + ] + }, + { + "video_id": "4N5P9", + "question_id": "4N5P9", + "video_file": "Charades/videos/4N5P9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put their phone on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.9> to <24.4>." + } + ] + }, + { + "video_id": "4N5P9", + "question_id": "4N5P9", + "video_file": "Charades/videos/4N5P9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person undressing in their closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.3> to <27.1>." + } + ] + }, + { + "video_id": "WVOQP", + "question_id": "WVOQP", + "video_file": "Charades/videos/WVOQP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.2> to <25.7>." + } + ] + }, + { + "video_id": "WVOQP", + "question_id": "WVOQP", + "video_file": "Charades/videos/WVOQP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <10.7> to <17.8>." + } + ] + }, + { + "video_id": "646CU", + "question_id": "646CU", + "video_file": "Charades/videos/646CU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person is holding a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.2> to <12.8>." + } + ] + }, + { + "video_id": "646CU", + "question_id": "646CU", + "video_file": "Charades/videos/646CU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person starts playing with a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.2> to <12.8>." + } + ] + }, + { + "video_id": "52GJU", + "question_id": "52GJU", + "video_file": "Charades/videos/52GJU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walks to a window which they look out of' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.0> to <11.0>." + } + ] + }, + { + "video_id": "52GJU", + "question_id": "52GJU", + "video_file": "Charades/videos/52GJU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person looks out window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.0> to <11.0>." + } + ] + }, + { + "video_id": "EI17N", + "question_id": "EI17N", + "video_file": "Charades/videos/EI17N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person is smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <9.6>." + } + ] + }, + { + "video_id": "0UFGC", + "question_id": "0UFGC", + "video_file": "Charades/videos/0UFGC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <16.6>." + } + ] + }, + { + "video_id": "0UFGC", + "question_id": "0UFGC", + "video_file": "Charades/videos/0UFGC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "0UFGC", + "question_id": "0UFGC", + "video_file": "Charades/videos/0UFGC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <16.3>." + } + ] + }, + { + "video_id": "0UFGC", + "question_id": "0UFGC", + "video_file": "Charades/videos/0UFGC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person is opening the door knob in the entryway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "UKLQ1", + "question_id": "UKLQ1", + "video_file": "Charades/videos/UKLQ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walks to turn the light on' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.9> to <12.9>." + } + ] + }, + { + "video_id": "UKLQ1", + "question_id": "UKLQ1", + "video_file": "Charades/videos/UKLQ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turning the light on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.9> to <12.9>." + } + ] + }, + { + "video_id": "UKLQ1", + "question_id": "UKLQ1", + "video_file": "Charades/videos/UKLQ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.9> to <12.9>." + } + ] + }, + { + "video_id": "XZ0KV", + "question_id": "XZ0KV", + "video_file": "Charades/videos/XZ0KV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person is opening the laptop on the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.3> to <14.9>." + } + ] + }, + { + "video_id": "XZ0KV", + "question_id": "XZ0KV", + "video_file": "Charades/videos/XZ0KV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens up his laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <14.9>." + } + ] + }, + { + "video_id": "K87AC", + "question_id": "K87AC", + "video_file": "Charades/videos/K87AC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.7> to <10.2>." + } + ] + }, + { + "video_id": "K87AC", + "question_id": "K87AC", + "video_file": "Charades/videos/K87AC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person leaves the room closing the door behind them' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.7> to <10.2>." + } + ] + }, + { + "video_id": "K87AC", + "question_id": "K87AC", + "video_file": "Charades/videos/K87AC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person dress into something else' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.6> to <12.7>." + } + ] + }, + { + "video_id": "K87AC", + "question_id": "K87AC", + "video_file": "Charades/videos/K87AC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "K87AC", + "question_id": "K87AC", + "video_file": "Charades/videos/K87AC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person dressing with items from a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.6> to <12.7>." + } + ] + }, + { + "video_id": "K87AC", + "question_id": "K87AC", + "video_file": "Charades/videos/K87AC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.7> to <10.2>." + } + ] + }, + { + "video_id": "K87AC", + "question_id": "K87AC", + "video_file": "Charades/videos/K87AC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.6> to <12.7>." + } + ] + }, + { + "video_id": "K87AC", + "question_id": "K87AC", + "video_file": "Charades/videos/K87AC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the door behind them' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.7> to <10.2>." + } + ] + }, + { + "video_id": "RSTVT", + "question_id": "RSTVT", + "video_file": "Charades/videos/RSTVT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs down the stairs eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <7.4>." + } + ] + }, + { + "video_id": "RSTVT", + "question_id": "RSTVT", + "video_file": "Charades/videos/RSTVT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is running down the stairs' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.1> to <7.4>." + } + ] + }, + { + "video_id": "VJ2QS", + "question_id": "VJ2QS", + "video_file": "Charades/videos/VJ2QS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sit down at a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <21.4>." + } + ] + }, + { + "video_id": "VJ2QS", + "question_id": "VJ2QS", + "video_file": "Charades/videos/VJ2QS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits at the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <21.4>." + } + ] + }, + { + "video_id": "VJ2QS", + "question_id": "VJ2QS", + "video_file": "Charades/videos/VJ2QS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person sits in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <21.3>." + } + ] + }, + { + "video_id": "RW587", + "question_id": "RW587", + "video_file": "Charades/videos/RW587.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'that person started laughing non-stop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.1> to <26.3>." + } + ] + }, + { + "video_id": "EU10X", + "question_id": "EU10X", + "video_file": "Charades/videos/EU10X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating a snack' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.2>." + } + ] + }, + { + "video_id": "EU10X", + "question_id": "EU10X", + "video_file": "Charades/videos/EU10X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person is watching television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <13.7>." + } + ] + }, + { + "video_id": "EU10X", + "question_id": "EU10X", + "video_file": "Charades/videos/EU10X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person eats some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.2>." + } + ] + }, + { + "video_id": "EU10X", + "question_id": "EU10X", + "video_file": "Charades/videos/EU10X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person watches television' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <13.7>." + } + ] + }, + { + "video_id": "HS521", + "question_id": "HS521", + "video_file": "Charades/videos/HS521.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turned off the lights' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.1> to <29.7>." + } + ] + }, + { + "video_id": "HS521", + "question_id": "HS521", + "video_file": "Charades/videos/HS521.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open the door turn off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.1> to <29.7>." + } + ] + }, + { + "video_id": "PN7N9", + "question_id": "PN7N9", + "video_file": "Charades/videos/PN7N9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person close the box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.6> to <19.9>." + } + ] + }, + { + "video_id": "PN7N9", + "question_id": "PN7N9", + "video_file": "Charades/videos/PN7N9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts the bag into a larger box'." + }, + { + "from": "gpt", + "value": "From <4.0> to <13.7>." + } + ] + }, + { + "video_id": "PN7N9", + "question_id": "PN7N9", + "video_file": "Charades/videos/PN7N9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "VVMYX", + "question_id": "VVMYX", + "video_file": "Charades/videos/VVMYX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pours something into a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "TJKGJ", + "question_id": "TJKGJ", + "video_file": "Charades/videos/TJKGJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sit down in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <33.0>." + } + ] + }, + { + "video_id": "TJKGJ", + "question_id": "TJKGJ", + "video_file": "Charades/videos/TJKGJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person closes the cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <18.4>." + } + ] + }, + { + "video_id": "TJKGJ", + "question_id": "TJKGJ", + "video_file": "Charades/videos/TJKGJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person look out the window'." + }, + { + "from": "gpt", + "value": "From <23.5> to <33.0>." + } + ] + }, + { + "video_id": "TJKGJ", + "question_id": "TJKGJ", + "video_file": "Charades/videos/TJKGJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sit in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <33.0>." + } + ] + }, + { + "video_id": "TJKGJ", + "question_id": "TJKGJ", + "video_file": "Charades/videos/TJKGJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.4> to <17.7>." + } + ] + }, + { + "video_id": "TJKGJ", + "question_id": "TJKGJ", + "video_file": "Charades/videos/TJKGJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person looking out a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.5> to <33.0>." + } + ] + }, + { + "video_id": "SGAGB", + "question_id": "SGAGB", + "video_file": "Charades/videos/SGAGB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person snuggles a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.1>." + } + ] + }, + { + "video_id": "EP19T", + "question_id": "EP19T", + "video_file": "Charades/videos/EP19T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs into his room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <10.9>." + } + ] + }, + { + "video_id": "EP19T", + "question_id": "EP19T", + "video_file": "Charades/videos/EP19T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a picture with his camera' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.2> to <26.3>." + } + ] + }, + { + "video_id": "EP19T", + "question_id": "EP19T", + "video_file": "Charades/videos/EP19T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is running from the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.5> to <10.9>." + } + ] + }, + { + "video_id": "G2UYF", + "question_id": "G2UYF", + "video_file": "Charades/videos/G2UYF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "CJP4R", + "question_id": "CJP4R", + "video_file": "Charades/videos/CJP4R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes out a camera' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.0> to <11.2>." + } + ] + }, + { + "video_id": "CJP4R", + "question_id": "CJP4R", + "video_file": "Charades/videos/CJP4R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person throws some food into a pan'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "B69CF", + "question_id": "B69CF", + "video_file": "Charades/videos/B69CF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.3> to <11.9>." + } + ] + }, + { + "video_id": "B69CF", + "question_id": "B69CF", + "video_file": "Charades/videos/B69CF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins awakening' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.6> to <9.6>." + } + ] + }, + { + "video_id": "B69CF", + "question_id": "B69CF", + "video_file": "Charades/videos/B69CF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person gets up to close a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <20.6>." + } + ] + }, + { + "video_id": "B69CF", + "question_id": "B69CF", + "video_file": "Charades/videos/B69CF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person go close an open window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <20.6>." + } + ] + }, + { + "video_id": "UIHDF", + "question_id": "UIHDF", + "video_file": "Charades/videos/UIHDF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <13.8>." + } + ] + }, + { + "video_id": "UIHDF", + "question_id": "UIHDF", + "video_file": "Charades/videos/UIHDF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.5> to <13.8>." + } + ] + }, + { + "video_id": "UIHDF", + "question_id": "UIHDF", + "video_file": "Charades/videos/UIHDF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is standing in the doorway smiling' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "UIHDF", + "question_id": "UIHDF", + "video_file": "Charades/videos/UIHDF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person smiling into the camera'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "F6Z7U", + "question_id": "F6Z7U", + "video_file": "Charades/videos/F6Z7U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is closing a refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <12.5>." + } + ] + }, + { + "video_id": "DDV1G", + "question_id": "DDV1G", + "video_file": "Charades/videos/DDV1G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is awakening' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "S0LA8", + "question_id": "S0LA8", + "video_file": "Charades/videos/S0LA8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turning the light off' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.1> to <26.4>." + } + ] + }, + { + "video_id": "S0LA8", + "question_id": "S0LA8", + "video_file": "Charades/videos/S0LA8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.1> to <26.4>." + } + ] + }, + { + "video_id": "S0LA8", + "question_id": "S0LA8", + "video_file": "Charades/videos/S0LA8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turning the doorknob to open the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <21.6>." + } + ] + }, + { + "video_id": "S0LA8", + "question_id": "S0LA8", + "video_file": "Charades/videos/S0LA8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws the book onto the desk' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.4> to <12.2>." + } + ] + }, + { + "video_id": "S0LA8", + "question_id": "S0LA8", + "video_file": "Charades/videos/S0LA8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.2> to <21.6>." + } + ] + }, + { + "video_id": "S0LA8", + "question_id": "S0LA8", + "video_file": "Charades/videos/S0LA8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they turn off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.1> to <26.4>." + } + ] + }, + { + "video_id": "WIQ2O", + "question_id": "WIQ2O", + "video_file": "Charades/videos/WIQ2O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is seen drinking from a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.0> to <10.2>." + } + ] + }, + { + "video_id": "WIQ2O", + "question_id": "WIQ2O", + "video_file": "Charades/videos/WIQ2O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sitting on floor smiling holding a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "WIQ2O", + "question_id": "WIQ2O", + "video_file": "Charades/videos/WIQ2O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinks from glass'." + }, + { + "from": "gpt", + "value": "From <4.0> to <10.2>." + } + ] + }, + { + "video_id": "6HHCU", + "question_id": "6HHCU", + "video_file": "Charades/videos/6HHCU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person is smiling on the phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.9> to <17.4>." + } + ] + }, + { + "video_id": "6HHCU", + "question_id": "6HHCU", + "video_file": "Charades/videos/6HHCU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'takes a picture of other person holding paper' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.8> to <17.3>." + } + ] + }, + { + "video_id": "6HHCU", + "question_id": "6HHCU", + "video_file": "Charades/videos/6HHCU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a picture from a small table'." + }, + { + "from": "gpt", + "value": "From <10.8> to <17.3>." + } + ] + }, + { + "video_id": "DB07P", + "question_id": "DB07P", + "video_file": "Charades/videos/DB07P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is holding a box'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "DB07P", + "question_id": "DB07P", + "video_file": "Charades/videos/DB07P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person is laughing at a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "Y3YE1", + "question_id": "Y3YE1", + "video_file": "Charades/videos/Y3YE1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person is smiling at a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <8.7>." + } + ] + }, + { + "video_id": "40DSU", + "question_id": "40DSU", + "video_file": "Charades/videos/40DSU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is holding a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.8>." + } + ] + }, + { + "video_id": "40DSU", + "question_id": "40DSU", + "video_file": "Charades/videos/40DSU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opened a door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "MMRRV", + "question_id": "MMRRV", + "video_file": "Charades/videos/MMRRV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turn off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.18>." + } + ] + }, + { + "video_id": "MMRRV", + "question_id": "MMRRV", + "video_file": "Charades/videos/MMRRV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turning off the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.18>." + } + ] + }, + { + "video_id": "Q9DW7", + "question_id": "Q9DW7", + "video_file": "Charades/videos/Q9DW7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person lied down on the sofa to watch tv'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <25.5>." + } + ] + }, + { + "video_id": "OKL5S", + "question_id": "OKL5S", + "video_file": "Charades/videos/OKL5S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is pouring a glass of milk' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <12.1>." + } + ] + }, + { + "video_id": "OKL5S", + "question_id": "OKL5S", + "video_file": "Charades/videos/OKL5S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.4> to <22.3>." + } + ] + }, + { + "video_id": "OKL5S", + "question_id": "OKL5S", + "video_file": "Charades/videos/OKL5S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours its contents into a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <12.1>." + } + ] + }, + { + "video_id": "RTIP8", + "question_id": "RTIP8", + "video_file": "Charades/videos/RTIP8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a smiling person is sitting nest to a mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.0> to <27.6>." + } + ] + }, + { + "video_id": "BGQIF", + "question_id": "BGQIF", + "video_file": "Charades/videos/BGQIF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating some food from a few various dishes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.9> to <19.5>." + } + ] + }, + { + "video_id": "BGQIF", + "question_id": "BGQIF", + "video_file": "Charades/videos/BGQIF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.9> to <19.5>." + } + ] + }, + { + "video_id": "BGQIF", + "question_id": "BGQIF", + "video_file": "Charades/videos/BGQIF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting on the couch'." + }, + { + "from": "gpt", + "value": "From <1.6> to <6.7>." + } + ] + }, + { + "video_id": "LZ07A", + "question_id": "LZ07A", + "video_file": "Charades/videos/LZ07A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts some clothes on' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.7> to <16.0>." + } + ] + }, + { + "video_id": "VB1HT", + "question_id": "VB1HT", + "video_file": "Charades/videos/VB1HT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person close the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.2> to <35.2>." + } + ] + }, + { + "video_id": "VB1HT", + "question_id": "VB1HT", + "video_file": "Charades/videos/VB1HT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.2> to <35.2>." + } + ] + }, + { + "video_id": "C7V0X", + "question_id": "C7V0X", + "video_file": "Charades/videos/C7V0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is throwing a pillow on the sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "C7V0X", + "question_id": "C7V0X", + "video_file": "Charades/videos/C7V0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drink from a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.6> to <19.1>." + } + ] + }, + { + "video_id": "C7V0X", + "question_id": "C7V0X", + "video_file": "Charades/videos/C7V0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person in a dining room throws a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "ILKXV", + "question_id": "ILKXV", + "video_file": "Charades/videos/ILKXV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "ILKXV", + "question_id": "ILKXV", + "video_file": "Charades/videos/ILKXV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walks through the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <16.8>." + } + ] + }, + { + "video_id": "ILKXV", + "question_id": "ILKXV", + "video_file": "Charades/videos/ILKXV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens the door to the bathroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "ILKXV", + "question_id": "ILKXV", + "video_file": "Charades/videos/ILKXV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person appear to laugh a little'." + }, + { + "from": "gpt", + "value": "From <26.6> to <32.0>." + } + ] + }, + { + "video_id": "SBG5F", + "question_id": "SBG5F", + "video_file": "Charades/videos/SBG5F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is pouring water into a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <16.4>." + } + ] + }, + { + "video_id": "SBG5F", + "question_id": "SBG5F", + "video_file": "Charades/videos/SBG5F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pours it into a glass on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.0> to <16.4>." + } + ] + }, + { + "video_id": "KKONU", + "question_id": "KKONU", + "video_file": "Charades/videos/KKONU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats the snack' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.5> to <18.6>." + } + ] + }, + { + "video_id": "KKONU", + "question_id": "KKONU", + "video_file": "Charades/videos/KKONU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person under a blanket awakens'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "KKONU", + "question_id": "KKONU", + "video_file": "Charades/videos/KKONU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws off the blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <8.8>." + } + ] + }, + { + "video_id": "KKONU", + "question_id": "KKONU", + "video_file": "Charades/videos/KKONU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats something' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.5> to <18.6>." + } + ] + }, + { + "video_id": "N8N6U", + "question_id": "N8N6U", + "video_file": "Charades/videos/N8N6U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes some medicine with a drink of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.6> to <26.1>." + } + ] + }, + { + "video_id": "EFR6I", + "question_id": "EFR6I", + "video_file": "Charades/videos/EFR6I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person holds some papers' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <34.8>." + } + ] + }, + { + "video_id": "PKNT2", + "question_id": "PKNT2", + "video_file": "Charades/videos/PKNT2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put the book down'." + }, + { + "from": "gpt", + "value": "From <16.3> to <22.4>." + } + ] + }, + { + "video_id": "BOHLW", + "question_id": "BOHLW", + "video_file": "Charades/videos/BOHLW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person shines a light to see where to vacuum' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.9> to <29.8>." + } + ] + }, + { + "video_id": "BOHLW", + "question_id": "BOHLW", + "video_file": "Charades/videos/BOHLW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a flash light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.9> to <29.8>." + } + ] + }, + { + "video_id": "YUCU2", + "question_id": "YUCU2", + "video_file": "Charades/videos/YUCU2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person cooking on the stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.0>." + } + ] + }, + { + "video_id": "YUCU2", + "question_id": "YUCU2", + "video_file": "Charades/videos/YUCU2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person cooking at a stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.0>." + } + ] + }, + { + "video_id": "EYV38", + "question_id": "EYV38", + "video_file": "Charades/videos/EYV38.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is lying under a blanket just awakening' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.4> to <23.6>." + } + ] + }, + { + "video_id": "8ZHLV", + "question_id": "8ZHLV", + "video_file": "Charades/videos/8ZHLV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a second person takes a drink from the glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.2> to <29.2>." + } + ] + }, + { + "video_id": "8ZHLV", + "question_id": "8ZHLV", + "video_file": "Charades/videos/8ZHLV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is pouring juice into a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.7> to <16.4>." + } + ] + }, + { + "video_id": "8ZHLV", + "question_id": "8ZHLV", + "video_file": "Charades/videos/8ZHLV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pours a cup'." + }, + { + "from": "gpt", + "value": "From <7.7> to <16.4>." + } + ] + }, + { + "video_id": "8ZHLV", + "question_id": "8ZHLV", + "video_file": "Charades/videos/8ZHLV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting on a chair to drink it'." + }, + { + "from": "gpt", + "value": "From <19.7> to <32.0>." + } + ] + }, + { + "video_id": "8ZHLV", + "question_id": "8ZHLV", + "video_file": "Charades/videos/8ZHLV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pours a drink into a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.2> to <29.2>." + } + ] + }, + { + "video_id": "ACJBD", + "question_id": "ACJBD", + "video_file": "Charades/videos/ACJBD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they start dressing by putting a jacket on' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.2> to <23.7>." + } + ] + }, + { + "video_id": "QQRP2", + "question_id": "QQRP2", + "video_file": "Charades/videos/QQRP2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is eating food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.2>." + } + ] + }, + { + "video_id": "BMF6X", + "question_id": "BMF6X", + "video_file": "Charades/videos/BMF6X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws ishoes on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.8> to <24.3>." + } + ] + }, + { + "video_id": "QNPDC", + "question_id": "QNPDC", + "video_file": "Charades/videos/QNPDC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pouring coffee into a cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.1> to <24.6>." + } + ] + }, + { + "video_id": "QNPDC", + "question_id": "QNPDC", + "video_file": "Charades/videos/QNPDC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts phone on table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.8> to <13.0>." + } + ] + }, + { + "video_id": "QNPDC", + "question_id": "QNPDC", + "video_file": "Charades/videos/QNPDC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person puts a camera on the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.8> to <13.0>." + } + ] + }, + { + "video_id": "QNPDC", + "question_id": "QNPDC", + "video_file": "Charades/videos/QNPDC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks into a room holding a phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <11.5>." + } + ] + }, + { + "video_id": "NHLQ1", + "question_id": "NHLQ1", + "video_file": "Charades/videos/NHLQ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is seated' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.3> to <23.0>." + } + ] + }, + { + "video_id": "NHLQ1", + "question_id": "NHLQ1", + "video_file": "Charades/videos/NHLQ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <19.9>." + } + ] + }, + { + "video_id": "NHLQ1", + "question_id": "NHLQ1", + "video_file": "Charades/videos/NHLQ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a picture of themselves'." + }, + { + "from": "gpt", + "value": "From <24.1> to <32.0>." + } + ] + }, + { + "video_id": "NHLQ1", + "question_id": "NHLQ1", + "video_file": "Charades/videos/NHLQ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person start eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.3> to <32.0>." + } + ] + }, + { + "video_id": "59X8N", + "question_id": "59X8N", + "video_file": "Charades/videos/59X8N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating breakfast' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <29.3> to <38.0>." + } + ] + }, + { + "video_id": "59X8N", + "question_id": "59X8N", + "video_file": "Charades/videos/59X8N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.3> to <38.0>." + } + ] + }, + { + "video_id": "59X8N", + "question_id": "59X8N", + "video_file": "Charades/videos/59X8N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eat the cereal'." + }, + { + "from": "gpt", + "value": "From <29.3> to <38.0>." + } + ] + }, + { + "video_id": "HJ2JO", + "question_id": "HJ2JO", + "video_file": "Charades/videos/HJ2JO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sneezing into a towel' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.3> to <29.3>." + } + ] + }, + { + "video_id": "HJ2JO", + "question_id": "HJ2JO", + "video_file": "Charades/videos/HJ2JO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person by the refrigerator was eating'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.4>." + } + ] + }, + { + "video_id": "NHTSB", + "question_id": "NHTSB", + "video_file": "Charades/videos/NHTSB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turned off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.6> to <33.0>." + } + ] + }, + { + "video_id": "C8WLX", + "question_id": "C8WLX", + "video_file": "Charades/videos/C8WLX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins putting on a pair of shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.2> to <35.0>." + } + ] + }, + { + "video_id": "C8WLX", + "question_id": "C8WLX", + "video_file": "Charades/videos/C8WLX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person fixes his hair in the mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.1> to <23.9>." + } + ] + }, + { + "video_id": "C8WLX", + "question_id": "C8WLX", + "video_file": "Charades/videos/C8WLX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks through a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.7> to <16.8>." + } + ] + }, + { + "video_id": "2ZG3U", + "question_id": "2ZG3U", + "video_file": "Charades/videos/2ZG3U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is watching television' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "2ZG3U", + "question_id": "2ZG3U", + "video_file": "Charades/videos/2ZG3U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours a glass of juice'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.3> to <32.0>." + } + ] + }, + { + "video_id": "2ZG3U", + "question_id": "2ZG3U", + "video_file": "Charades/videos/2ZG3U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins pouring soda into a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.3> to <32.0>." + } + ] + }, + { + "video_id": "2ZG3U", + "question_id": "2ZG3U", + "video_file": "Charades/videos/2ZG3U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pour juice in a cup that on the counter' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.3> to <32.0>." + } + ] + }, + { + "video_id": "P3JMV", + "question_id": "P3JMV", + "video_file": "Charades/videos/P3JMV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting a glass on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <26.4>." + } + ] + }, + { + "video_id": "RF4OI", + "question_id": "RF4OI", + "video_file": "Charades/videos/RF4OI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person washing their hands' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.7> to <17.6>." + } + ] + }, + { + "video_id": "113YU", + "question_id": "113YU", + "video_file": "Charades/videos/113YU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a cabinet in their entryway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <6.9>." + } + ] + }, + { + "video_id": "113YU", + "question_id": "113YU", + "video_file": "Charades/videos/113YU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.7> to <26.9>." + } + ] + }, + { + "video_id": "113YU", + "question_id": "113YU", + "video_file": "Charades/videos/113YU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the door on the cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.7> to <26.9>." + } + ] + }, + { + "video_id": "V8JOH", + "question_id": "V8JOH", + "video_file": "Charades/videos/V8JOH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <22.0>." + } + ] + }, + { + "video_id": "V8JOH", + "question_id": "V8JOH", + "video_file": "Charades/videos/V8JOH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes some medicine' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "V8JOH", + "question_id": "V8JOH", + "video_file": "Charades/videos/V8JOH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is standing in a doorway holding some medicine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.8>." + } + ] + }, + { + "video_id": "DNZD8", + "question_id": "DNZD8", + "video_file": "Charades/videos/DNZD8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person the hallway drinking from a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "OZPA9", + "question_id": "OZPA9", + "video_file": "Charades/videos/OZPA9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns the light on' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <8.5>." + } + ] + }, + { + "video_id": "Z3AGB", + "question_id": "Z3AGB", + "video_file": "Charades/videos/Z3AGB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens his closet door to put some cloths away' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.7> to <8.9>." + } + ] + }, + { + "video_id": "Z3AGB", + "question_id": "Z3AGB", + "video_file": "Charades/videos/Z3AGB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turned on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.3> to <10.5>." + } + ] + }, + { + "video_id": "Z3AGB", + "question_id": "Z3AGB", + "video_file": "Charades/videos/Z3AGB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opened the door to the closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.7> to <8.9>." + } + ] + }, + { + "video_id": "Z3AGB", + "question_id": "Z3AGB", + "video_file": "Charades/videos/Z3AGB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person put folded clothes on the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.2> to <24.7>." + } + ] + }, + { + "video_id": "R0OI6", + "question_id": "R0OI6", + "video_file": "Charades/videos/R0OI6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person is sitting on a chair laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <24.2>." + } + ] + }, + { + "video_id": "R0OI6", + "question_id": "R0OI6", + "video_file": "Charades/videos/R0OI6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person playing on a phone'." + }, + { + "from": "gpt", + "value": "From <14.0> to <25.4>." + } + ] + }, + { + "video_id": "LEOL6", + "question_id": "LEOL6", + "video_file": "Charades/videos/LEOL6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person is holding medicine' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.0> to <7.0>." + } + ] + }, + { + "video_id": "TTRAJ", + "question_id": "TTRAJ", + "video_file": "Charades/videos/TTRAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sneezing on the food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.9> to <17.2>." + } + ] + }, + { + "video_id": "TTRAJ", + "question_id": "TTRAJ", + "video_file": "Charades/videos/TTRAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person begins to start sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.9> to <17.2>." + } + ] + }, + { + "video_id": "R0OGE", + "question_id": "R0OGE", + "video_file": "Charades/videos/R0OGE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person the drinking from a glass in the kitchen' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.3> to <44.0>." + } + ] + }, + { + "video_id": "R0OGE", + "question_id": "R0OGE", + "video_file": "Charades/videos/R0OGE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinks from a glass'." + }, + { + "from": "gpt", + "value": "From <32.3> to <44.0>." + } + ] + }, + { + "video_id": "6912B", + "question_id": "6912B", + "video_file": "Charades/videos/6912B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they are holding a laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.0> to <18.7>." + } + ] + }, + { + "video_id": "UAT1F", + "question_id": "UAT1F", + "video_file": "Charades/videos/UAT1F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open up a laptop'." + }, + { + "from": "gpt", + "value": "From <17.7> to <26.2>." + } + ] + }, + { + "video_id": "UAT1F", + "question_id": "UAT1F", + "video_file": "Charades/videos/UAT1F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <26.2>." + } + ] + }, + { + "video_id": "UAT1F", + "question_id": "UAT1F", + "video_file": "Charades/videos/UAT1F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put the book down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <20.4>." + } + ] + }, + { + "video_id": "UAT1F", + "question_id": "UAT1F", + "video_file": "Charades/videos/UAT1F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the book down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <20.4>." + } + ] + }, + { + "video_id": "O5JIS", + "question_id": "O5JIS", + "video_file": "Charades/videos/O5JIS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person at the sink takes some medicine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.4> to <29.8>." + } + ] + }, + { + "video_id": "E6Q95", + "question_id": "E6Q95", + "video_file": "Charades/videos/E6Q95.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eat a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <26.0>." + } + ] + }, + { + "video_id": "E6Q95", + "question_id": "E6Q95", + "video_file": "Charades/videos/E6Q95.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is undressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <11.8>." + } + ] + }, + { + "video_id": "SVQL5", + "question_id": "SVQL5", + "video_file": "Charades/videos/SVQL5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws it on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.3> to <25.0>." + } + ] + }, + { + "video_id": "AMUDT", + "question_id": "AMUDT", + "video_file": "Charades/videos/AMUDT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they start continuously sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <16.6>." + } + ] + }, + { + "video_id": "AMUDT", + "question_id": "AMUDT", + "video_file": "Charades/videos/AMUDT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person is constantly sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <16.6>." + } + ] + }, + { + "video_id": "MX9XB", + "question_id": "MX9XB", + "video_file": "Charades/videos/MX9XB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person is sitting on chair opening bag of candy' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "MX9XB", + "question_id": "MX9XB", + "video_file": "Charades/videos/MX9XB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throwing food out of a bag at a vacuum' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.8> to <34.6>." + } + ] + }, + { + "video_id": "MX9XB", + "question_id": "MX9XB", + "video_file": "Charades/videos/MX9XB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throwing the bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.8> to <34.6>." + } + ] + }, + { + "video_id": "58YA5", + "question_id": "58YA5", + "video_file": "Charades/videos/58YA5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a phone from the table'." + }, + { + "from": "gpt", + "value": "From <12.5> to <17.7>." + } + ] + }, + { + "video_id": "58YA5", + "question_id": "58YA5", + "video_file": "Charades/videos/58YA5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting on the couch' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.8>." + } + ] + }, + { + "video_id": "H3CDC", + "question_id": "H3CDC", + "video_file": "Charades/videos/H3CDC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes out a towel' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <7.8>." + } + ] + }, + { + "video_id": "H3CDC", + "question_id": "H3CDC", + "video_file": "Charades/videos/H3CDC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is opens the cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "H3CDC", + "question_id": "H3CDC", + "video_file": "Charades/videos/H3CDC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'there is a person that opens a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "3BVG1", + "question_id": "3BVG1", + "video_file": "Charades/videos/3BVG1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts a broom inside a hallway closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.1> to <21.7>." + } + ] + }, + { + "video_id": "BYSQ7", + "question_id": "BYSQ7", + "video_file": "Charades/videos/BYSQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sits in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "BYSQ7", + "question_id": "BYSQ7", + "video_file": "Charades/videos/BYSQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins to drink from a cup'." + }, + { + "from": "gpt", + "value": "From <0.6> to <14.9>." + } + ] + }, + { + "video_id": "BYSQ7", + "question_id": "BYSQ7", + "video_file": "Charades/videos/BYSQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes off their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.3> to <34.3>." + } + ] + }, + { + "video_id": "YXBXH", + "question_id": "YXBXH", + "video_file": "Charades/videos/YXBXH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person playing on a phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.5> to <14.3>." + } + ] + }, + { + "video_id": "YXBXH", + "question_id": "YXBXH", + "video_file": "Charades/videos/YXBXH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking from a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.8>." + } + ] + }, + { + "video_id": "KCLMY", + "question_id": "KCLMY", + "video_file": "Charades/videos/KCLMY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is looking in a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <16.9>." + } + ] + }, + { + "video_id": "DBN7R", + "question_id": "DBN7R", + "video_file": "Charades/videos/DBN7R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person smiling because the conversation they are having is humorous' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "3RSJV", + "question_id": "3RSJV", + "video_file": "Charades/videos/3RSJV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drink a glass of soda' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.1> to <22.5>." + } + ] + }, + { + "video_id": "3RSJV", + "question_id": "3RSJV", + "video_file": "Charades/videos/3RSJV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinks from a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.1> to <22.5>." + } + ] + }, + { + "video_id": "3RSJV", + "question_id": "3RSJV", + "video_file": "Charades/videos/3RSJV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drink a glass a water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.1> to <22.5>." + } + ] + }, + { + "video_id": "1BUFQ", + "question_id": "1BUFQ", + "video_file": "Charades/videos/1BUFQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sitting on the floor to work on homework' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.8> to <38.5>." + } + ] + }, + { + "video_id": "1BUFQ", + "question_id": "1BUFQ", + "video_file": "Charades/videos/1BUFQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they start sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.1> to <17.3>." + } + ] + }, + { + "video_id": "NNG2V", + "question_id": "NNG2V", + "video_file": "Charades/videos/NNG2V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person awakens as another'." + }, + { + "from": "gpt", + "value": "From <10.7> to <18.2>." + } + ] + }, + { + "video_id": "NNG2V", + "question_id": "NNG2V", + "video_file": "Charades/videos/NNG2V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'being woken up by another person to watch television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.8>." + } + ] + }, + { + "video_id": "NNG2V", + "question_id": "NNG2V", + "video_file": "Charades/videos/NNG2V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person watching television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.8>." + } + ] + }, + { + "video_id": "VFDQU", + "question_id": "VFDQU", + "video_file": "Charades/videos/VFDQU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "VFDQU", + "question_id": "VFDQU", + "video_file": "Charades/videos/VFDQU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opens their refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "VFDQU", + "question_id": "VFDQU", + "video_file": "Charades/videos/VFDQU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person stand watching the television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.8>." + } + ] + }, + { + "video_id": "138LG", + "question_id": "138LG", + "video_file": "Charades/videos/138LG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turning off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <14.7>." + } + ] + }, + { + "video_id": "138LG", + "question_id": "138LG", + "video_file": "Charades/videos/138LG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person shuts off the light as they leave' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <14.7>." + } + ] + }, + { + "video_id": "138LG", + "question_id": "138LG", + "video_file": "Charades/videos/138LG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is walking out of a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <12.8>." + } + ] + }, + { + "video_id": "WBT2F", + "question_id": "WBT2F", + "video_file": "Charades/videos/WBT2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.6> to <8.5>." + } + ] + }, + { + "video_id": "WBT2F", + "question_id": "WBT2F", + "video_file": "Charades/videos/WBT2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person puts a towel in a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.3>." + } + ] + }, + { + "video_id": "WBT2F", + "question_id": "WBT2F", + "video_file": "Charades/videos/WBT2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the towel in' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.3>." + } + ] + }, + { + "video_id": "WBT2F", + "question_id": "WBT2F", + "video_file": "Charades/videos/WBT2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <7.4> to <16.2>." + } + ] + }, + { + "video_id": "1UNXY", + "question_id": "1UNXY", + "video_file": "Charades/videos/1UNXY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.0> to <6.5>." + } + ] + }, + { + "video_id": "1UNXY", + "question_id": "1UNXY", + "video_file": "Charades/videos/1UNXY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts tidying up the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.7> to <24.4>." + } + ] + }, + { + "video_id": "KT3MJ", + "question_id": "KT3MJ", + "video_file": "Charades/videos/KT3MJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a refrigerator door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "KT3MJ", + "question_id": "KT3MJ", + "video_file": "Charades/videos/KT3MJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts cooking'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.8> to <34.0>." + } + ] + }, + { + "video_id": "TNEEH", + "question_id": "TNEEH", + "video_file": "Charades/videos/TNEEH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a drink from a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.8> to <25.5>." + } + ] + }, + { + "video_id": "7L3DI", + "question_id": "7L3DI", + "video_file": "Charades/videos/7L3DI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.4> to <28.0>." + } + ] + }, + { + "video_id": "7L3DI", + "question_id": "7L3DI", + "video_file": "Charades/videos/7L3DI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sits on the couch smiling watching something' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "7L3DI", + "question_id": "7L3DI", + "video_file": "Charades/videos/7L3DI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <9.9>." + } + ] + }, + { + "video_id": "7L3DI", + "question_id": "7L3DI", + "video_file": "Charades/videos/7L3DI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sitting on a sofa than gets up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "7L3DI", + "question_id": "7L3DI", + "video_file": "Charades/videos/7L3DI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting on the sofa' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "7L3DI", + "question_id": "7L3DI", + "video_file": "Charades/videos/7L3DI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'fixing their hair as the person does so'." + }, + { + "from": "gpt", + "value": "From <8.7> to <18.7>." + } + ] + }, + { + "video_id": "883AD", + "question_id": "883AD", + "video_file": "Charades/videos/883AD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person snuggling with a blanket on the stairs' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.2> to <34.0>." + } + ] + }, + { + "video_id": "BSFVT", + "question_id": "BSFVT", + "video_file": "Charades/videos/BSFVT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person decides to close a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "BSFVT", + "question_id": "BSFVT", + "video_file": "Charades/videos/BSFVT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person at a desk closes a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "THS0O", + "question_id": "THS0O", + "video_file": "Charades/videos/THS0O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the phone on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.1> to <24.5>." + } + ] + }, + { + "video_id": "THS0O", + "question_id": "THS0O", + "video_file": "Charades/videos/THS0O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts their phone onto the counter' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.1> to <24.5>." + } + ] + }, + { + "video_id": "11TTU", + "question_id": "11TTU", + "video_file": "Charades/videos/11TTU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person holding a phone enters the room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <19.9>." + } + ] + }, + { + "video_id": "11TTU", + "question_id": "11TTU", + "video_file": "Charades/videos/11TTU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is pouring a drink into a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <16.7>." + } + ] + }, + { + "video_id": "11TTU", + "question_id": "11TTU", + "video_file": "Charades/videos/11TTU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts their phone down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <9.7>." + } + ] + }, + { + "video_id": "11TTU", + "question_id": "11TTU", + "video_file": "Charades/videos/11TTU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pour water into a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.4> to <16.7>." + } + ] + }, + { + "video_id": "11TTU", + "question_id": "11TTU", + "video_file": "Charades/videos/11TTU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pouring themselves a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.4> to <16.7>." + } + ] + }, + { + "video_id": "11TTU", + "question_id": "11TTU", + "video_file": "Charades/videos/11TTU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the phone on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <9.7>." + } + ] + }, + { + "video_id": "FGEJI", + "question_id": "FGEJI", + "video_file": "Charades/videos/FGEJI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'watching another person running with shoes on' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.8> to <16.7>." + } + ] + }, + { + "video_id": "FGEJI", + "question_id": "FGEJI", + "video_file": "Charades/videos/FGEJI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is walking around a car holding their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.6> to <22.1>." + } + ] + }, + { + "video_id": "FGEJI", + "question_id": "FGEJI", + "video_file": "Charades/videos/FGEJI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person the other is holding shoes'." + }, + { + "from": "gpt", + "value": "From <11.6> to <22.1>." + } + ] + }, + { + "video_id": "FGEJI", + "question_id": "FGEJI", + "video_file": "Charades/videos/FGEJI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person running in circles'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.8> to <16.7>." + } + ] + }, + { + "video_id": "G4O9P", + "question_id": "G4O9P", + "video_file": "Charades/videos/G4O9P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person runs in circles'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "G4O9P", + "question_id": "G4O9P", + "video_file": "Charades/videos/G4O9P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person reads a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.9> to <23.4>." + } + ] + }, + { + "video_id": "G4O9P", + "question_id": "G4O9P", + "video_file": "Charades/videos/G4O9P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs into the garage'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "G4O9P", + "question_id": "G4O9P", + "video_file": "Charades/videos/G4O9P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes off their shoes'." + }, + { + "from": "gpt", + "value": "From <6.2> to <15.8>." + } + ] + }, + { + "video_id": "472B0", + "question_id": "472B0", + "video_file": "Charades/videos/472B0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person reading a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.7> to <33.0>." + } + ] + }, + { + "video_id": "472B0", + "question_id": "472B0", + "video_file": "Charades/videos/472B0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is awakening from the sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.7> to <12.6>." + } + ] + }, + { + "video_id": "472B0", + "question_id": "472B0", + "video_file": "Charades/videos/472B0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person lies in bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "472B0", + "question_id": "472B0", + "video_file": "Charades/videos/472B0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begin reading a book'." + }, + { + "from": "gpt", + "value": "From <21.7> to <33.0>." + } + ] + }, + { + "video_id": "BMPEV", + "question_id": "BMPEV", + "video_file": "Charades/videos/BMPEV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turning the light off' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.6> to <25.2>." + } + ] + }, + { + "video_id": "BMPEV", + "question_id": "BMPEV", + "video_file": "Charades/videos/BMPEV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sneezing into a bag of groceries' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.5> to <16.8>." + } + ] + }, + { + "video_id": "BMPEV", + "question_id": "BMPEV", + "video_file": "Charades/videos/BMPEV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turning off a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.6> to <25.2>." + } + ] + }, + { + "video_id": "YC3OE", + "question_id": "YC3OE", + "video_file": "Charades/videos/YC3OE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person undressing in a living room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "YC3OE", + "question_id": "YC3OE", + "video_file": "Charades/videos/YC3OE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits down on a couch' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.4> to <38.6>." + } + ] + }, + { + "video_id": "YC3OE", + "question_id": "YC3OE", + "video_file": "Charades/videos/YC3OE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats some food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.1> to <34.1>." + } + ] + }, + { + "video_id": "YC3OE", + "question_id": "YC3OE", + "video_file": "Charades/videos/YC3OE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they take their medicine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <25.1>." + } + ] + }, + { + "video_id": "CALKE", + "question_id": "CALKE", + "video_file": "Charades/videos/CALKE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put them on the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.5> to <44.6>." + } + ] + }, + { + "video_id": "CALKE", + "question_id": "CALKE", + "video_file": "Charades/videos/CALKE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is in a dining room dressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.0> to <27.6>." + } + ] + }, + { + "video_id": "CALKE", + "question_id": "CALKE", + "video_file": "Charades/videos/CALKE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting on shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <26.9>." + } + ] + }, + { + "video_id": "CALKE", + "question_id": "CALKE", + "video_file": "Charades/videos/CALKE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person smile as they tidy up the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.8> to <47.5>." + } + ] + }, + { + "video_id": "O8FBU", + "question_id": "O8FBU", + "video_file": "Charades/videos/O8FBU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person snuggles with a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.3> to <34.0>." + } + ] + }, + { + "video_id": "ZIU9C", + "question_id": "ZIU9C", + "video_file": "Charades/videos/ZIU9C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes a book from a bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "ZIU9C", + "question_id": "ZIU9C", + "video_file": "Charades/videos/ZIU9C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes a book off of the bed'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "ZIU9C", + "question_id": "ZIU9C", + "video_file": "Charades/videos/ZIU9C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts it on the table'." + }, + { + "from": "gpt", + "value": "From <9.8> to <15.7>." + } + ] + }, + { + "video_id": "ZIU9C", + "question_id": "ZIU9C", + "video_file": "Charades/videos/ZIU9C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <23.0>." + } + ] + }, + { + "video_id": "ZIU9C", + "question_id": "ZIU9C", + "video_file": "Charades/videos/ZIU9C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person opens the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <23.0>." + } + ] + }, + { + "video_id": "EAKL5", + "question_id": "EAKL5", + "video_file": "Charades/videos/EAKL5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'putting a cell phone in the person's pocket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.3> to <28.5>." + } + ] + }, + { + "video_id": "9PSZA", + "question_id": "9PSZA", + "video_file": "Charades/videos/9PSZA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is throwing shoes'." + }, + { + "from": "gpt", + "value": "From <3.3> to <13.4>." + } + ] + }, + { + "video_id": "9PSZA", + "question_id": "9PSZA", + "video_file": "Charades/videos/9PSZA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throw on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.0> to <13.9>." + } + ] + }, + { + "video_id": "9PSZA", + "question_id": "9PSZA", + "video_file": "Charades/videos/9PSZA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person throws their shoes down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.3> to <13.4>." + } + ] + }, + { + "video_id": "NEFHO", + "question_id": "NEFHO", + "video_file": "Charades/videos/NEFHO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking from a glass of water'." + }, + { + "from": "gpt", + "value": "From <8.6> to <15.0>." + } + ] + }, + { + "video_id": "9N44Y", + "question_id": "9N44Y", + "video_file": "Charades/videos/9N44Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person in the video opened the cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "9N44Y", + "question_id": "9N44Y", + "video_file": "Charades/videos/9N44Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closed the cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.0> to <17.5>." + } + ] + }, + { + "video_id": "AWGMW", + "question_id": "AWGMW", + "video_file": "Charades/videos/AWGMW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drink from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.3> to <16.9>." + } + ] + }, + { + "video_id": "AWGMW", + "question_id": "AWGMW", + "video_file": "Charades/videos/AWGMW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put on their shoes'." + }, + { + "from": "gpt", + "value": "From <6.2> to <12.6>." + } + ] + }, + { + "video_id": "AWGMW", + "question_id": "AWGMW", + "video_file": "Charades/videos/AWGMW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person running out of the room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.6> to <23.7>." + } + ] + }, + { + "video_id": "AWGMW", + "question_id": "AWGMW", + "video_file": "Charades/videos/AWGMW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person run out the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.6> to <23.7>." + } + ] + }, + { + "video_id": "AWGMW", + "question_id": "AWGMW", + "video_file": "Charades/videos/AWGMW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put on some shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.2> to <12.6>." + } + ] + }, + { + "video_id": "AWGMW", + "question_id": "AWGMW", + "video_file": "Charades/videos/AWGMW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put their shoes on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.2> to <12.6>." + } + ] + }, + { + "video_id": "AWGMW", + "question_id": "AWGMW", + "video_file": "Charades/videos/AWGMW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person runs out of the room'." + }, + { + "from": "gpt", + "value": "From <15.6> to <23.7>." + } + ] + }, + { + "video_id": "76BFU", + "question_id": "76BFU", + "video_file": "Charades/videos/76BFU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begin to talk on the phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <32.0>." + } + ] + }, + { + "video_id": "K56PA", + "question_id": "K56PA", + "video_file": "Charades/videos/K56PA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they sit down on the couch' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.9> to <21.7>." + } + ] + }, + { + "video_id": "K56PA", + "question_id": "K56PA", + "video_file": "Charades/videos/K56PA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is holding a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.2> to <11.5>." + } + ] + }, + { + "video_id": "K56PA", + "question_id": "K56PA", + "video_file": "Charades/videos/K56PA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put the box on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.2> to <10.7>." + } + ] + }, + { + "video_id": "K56PA", + "question_id": "K56PA", + "video_file": "Charades/videos/K56PA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person dressing in their living room'." + }, + { + "from": "gpt", + "value": "From <10.1> to <19.6>." + } + ] + }, + { + "video_id": "K56PA", + "question_id": "K56PA", + "video_file": "Charades/videos/K56PA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person walks into a room holding a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.2> to <11.5>." + } + ] + }, + { + "video_id": "Z1VB4", + "question_id": "Z1VB4", + "video_file": "Charades/videos/Z1VB4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts to cook' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.5> to <44.4>." + } + ] + }, + { + "video_id": "Z1VB4", + "question_id": "Z1VB4", + "video_file": "Charades/videos/Z1VB4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person start pouring water into a pot to begin cooking' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.5> to <44.4>." + } + ] + }, + { + "video_id": "CA5OI", + "question_id": "CA5OI", + "video_file": "Charades/videos/CA5OI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person is opening a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <7.3>." + } + ] + }, + { + "video_id": "CA5OI", + "question_id": "CA5OI", + "video_file": "Charades/videos/CA5OI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is opening a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <7.3>." + } + ] + }, + { + "video_id": "6VUKN", + "question_id": "6VUKN", + "video_file": "Charades/videos/6VUKN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins walking to the doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.2> to <31.1>." + } + ] + }, + { + "video_id": "5Q6N6", + "question_id": "5Q6N6", + "video_file": "Charades/videos/5Q6N6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person smiling as they take pictures of themselves' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.5> to <14.9>." + } + ] + }, + { + "video_id": "FT3D7", + "question_id": "FT3D7", + "video_file": "Charades/videos/FT3D7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting at a table eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.1>." + } + ] + }, + { + "video_id": "FT3D7", + "question_id": "FT3D7", + "video_file": "Charades/videos/FT3D7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person stand up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.9> to <10.8>." + } + ] + }, + { + "video_id": "FT3D7", + "question_id": "FT3D7", + "video_file": "Charades/videos/FT3D7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is sitting at a table eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.1>." + } + ] + }, + { + "video_id": "42SEP", + "question_id": "42SEP", + "video_file": "Charades/videos/42SEP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits down on chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <26.0>." + } + ] + }, + { + "video_id": "42SEP", + "question_id": "42SEP", + "video_file": "Charades/videos/42SEP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sit down on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <26.0>." + } + ] + }, + { + "video_id": "MAPVV", + "question_id": "MAPVV", + "video_file": "Charades/videos/MAPVV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person lies down on a bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.7> to <33.0>." + } + ] + }, + { + "video_id": "MAPVV", + "question_id": "MAPVV", + "video_file": "Charades/videos/MAPVV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts a book down on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.2> to <30.5>." + } + ] + }, + { + "video_id": "DSWWV", + "question_id": "DSWWV", + "video_file": "Charades/videos/DSWWV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <4.9>." + } + ] + }, + { + "video_id": "DSWWV", + "question_id": "DSWWV", + "video_file": "Charades/videos/DSWWV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.0> to <21.4>." + } + ] + }, + { + "video_id": "DSWWV", + "question_id": "DSWWV", + "video_file": "Charades/videos/DSWWV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put the camera down'." + }, + { + "from": "gpt", + "value": "From <13.6> to <19.1>." + } + ] + }, + { + "video_id": "DSWWV", + "question_id": "DSWWV", + "video_file": "Charades/videos/DSWWV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take a picture of theirselves with the phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "DSWWV", + "question_id": "DSWWV", + "video_file": "Charades/videos/DSWWV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take a picture' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "DSWWV", + "question_id": "DSWWV", + "video_file": "Charades/videos/DSWWV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <21.4>." + } + ] + }, + { + "video_id": "D3PPI", + "question_id": "D3PPI", + "video_file": "Charades/videos/D3PPI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.3> to <34.0>." + } + ] + }, + { + "video_id": "29T54", + "question_id": "29T54", + "video_file": "Charades/videos/29T54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they stand up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <7.9>." + } + ] + }, + { + "video_id": "29T54", + "question_id": "29T54", + "video_file": "Charades/videos/29T54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "29T54", + "question_id": "29T54", + "video_file": "Charades/videos/29T54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person sits in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "29T54", + "question_id": "29T54", + "video_file": "Charades/videos/29T54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open the closet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <19.3>." + } + ] + }, + { + "video_id": "29T54", + "question_id": "29T54", + "video_file": "Charades/videos/29T54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts in a closet where he removes other clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.7> to <29.0>." + } + ] + }, + { + "video_id": "29T54", + "question_id": "29T54", + "video_file": "Charades/videos/29T54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes off some clothes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <28.2>." + } + ] + }, + { + "video_id": "29T54", + "question_id": "29T54", + "video_file": "Charades/videos/29T54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes some clothes from the closet'." + }, + { + "from": "gpt", + "value": "From <19.5> to <28.2>." + } + ] + }, + { + "video_id": "29T54", + "question_id": "29T54", + "video_file": "Charades/videos/29T54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening a wardrobe cabinet to get clothes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <19.3>." + } + ] + }, + { + "video_id": "3C36H", + "question_id": "3C36H", + "video_file": "Charades/videos/3C36H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person opening bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <10.0>." + } + ] + }, + { + "video_id": "XZPAS", + "question_id": "XZPAS", + "video_file": "Charades/videos/XZPAS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a different person is looking at a picture' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.9> to <8.3>." + } + ] + }, + { + "video_id": "WULQ2", + "question_id": "WULQ2", + "video_file": "Charades/videos/WULQ2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.5> to <29.0>." + } + ] + }, + { + "video_id": "WZO6V", + "question_id": "WZO6V", + "video_file": "Charades/videos/WZO6V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person open door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "WZO6V", + "question_id": "WZO6V", + "video_file": "Charades/videos/WZO6V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a blanket off a sink' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <8.6>." + } + ] + }, + { + "video_id": "WZO6V", + "question_id": "WZO6V", + "video_file": "Charades/videos/WZO6V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the door behind them' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <17.1>." + } + ] + }, + { + "video_id": "3731F", + "question_id": "3731F", + "video_file": "Charades/videos/3731F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person stand up'." + }, + { + "from": "gpt", + "value": "From <12.1> to <17.7>." + } + ] + }, + { + "video_id": "3731F", + "question_id": "3731F", + "video_file": "Charades/videos/3731F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.4> to <23.6>." + } + ] + }, + { + "video_id": "KUO3N", + "question_id": "KUO3N", + "video_file": "Charades/videos/KUO3N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person closes a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <22.3>." + } + ] + }, + { + "video_id": "KUO3N", + "question_id": "KUO3N", + "video_file": "Charades/videos/KUO3N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <22.3>." + } + ] + }, + { + "video_id": "FW1L0", + "question_id": "FW1L0", + "video_file": "Charades/videos/FW1L0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating by the stove'." + }, + { + "from": "gpt", + "value": "From <2.0> to <8.5>." + } + ] + }, + { + "video_id": "FZ6AD", + "question_id": "FZ6AD", + "video_file": "Charades/videos/FZ6AD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts throwing clothes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.3> to <30.0>." + } + ] + }, + { + "video_id": "IPNEG", + "question_id": "IPNEG", + "video_file": "Charades/videos/IPNEG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person watching the laptop'." + }, + { + "from": "gpt", + "value": "From <30.3> to <36.0>." + } + ] + }, + { + "video_id": "T1OQ2", + "question_id": "T1OQ2", + "video_file": "Charades/videos/T1OQ2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting the glass down'." + }, + { + "from": "gpt", + "value": "From <17.7> to <22.1>." + } + ] + }, + { + "video_id": "XY2M3", + "question_id": "XY2M3", + "video_file": "Charades/videos/XY2M3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they grab a book they've been putting off reading' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.3> to <33.0>." + } + ] + }, + { + "video_id": "XY2M3", + "question_id": "XY2M3", + "video_file": "Charades/videos/XY2M3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person run towards the bedroom's doorway'." + }, + { + "from": "gpt", + "value": "From <25.9> to <33.0>." + } + ] + }, + { + "video_id": "XY2M3", + "question_id": "XY2M3", + "video_file": "Charades/videos/XY2M3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sitting on the bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.4> to <30.4>." + } + ] + }, + { + "video_id": "XY2M3", + "question_id": "XY2M3", + "video_file": "Charades/videos/XY2M3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person run out of the room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.9> to <33.0>." + } + ] + }, + { + "video_id": "SF6NU", + "question_id": "SF6NU", + "video_file": "Charades/videos/SF6NU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person runs in'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "SF6NU", + "question_id": "SF6NU", + "video_file": "Charades/videos/SF6NU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person running'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "L80CV", + "question_id": "L80CV", + "video_file": "Charades/videos/L80CV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin opening bags of groceries to put food away' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.9> to <31.2>." + } + ] + }, + { + "video_id": "L80CV", + "question_id": "L80CV", + "video_file": "Charades/videos/L80CV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person cooking food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.1> to <36.0>." + } + ] + }, + { + "video_id": "L80CV", + "question_id": "L80CV", + "video_file": "Charades/videos/L80CV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts cooking a meal' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.1> to <36.0>." + } + ] + }, + { + "video_id": "XRR22", + "question_id": "XRR22", + "video_file": "Charades/videos/XRR22.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the old light on the shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <35.4> to <46.8>." + } + ] + }, + { + "video_id": "XRR22", + "question_id": "XRR22", + "video_file": "Charades/videos/XRR22.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a pillow'." + }, + { + "from": "gpt", + "value": "From <36.9> to <46.8>." + } + ] + }, + { + "video_id": "OGK5Y", + "question_id": "OGK5Y", + "video_file": "Charades/videos/OGK5Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they stand up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <33.1> to <39.3>." + } + ] + }, + { + "video_id": "U8V9N", + "question_id": "U8V9N", + "video_file": "Charades/videos/U8V9N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they open the bathroom window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.9> to <31.3>." + } + ] + }, + { + "video_id": "U8V9N", + "question_id": "U8V9N", + "video_file": "Charades/videos/U8V9N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person look out the window'." + }, + { + "from": "gpt", + "value": "From <18.9> to <31.1>." + } + ] + }, + { + "video_id": "CRW0G", + "question_id": "CRW0G", + "video_file": "Charades/videos/CRW0G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sneezing into a mirror' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "CRW0G", + "question_id": "CRW0G", + "video_file": "Charades/videos/CRW0G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinks from the glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <28.0>." + } + ] + }, + { + "video_id": "1D31Z", + "question_id": "1D31Z", + "video_file": "Charades/videos/1D31Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking from a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <28.6> to <35.3>." + } + ] + }, + { + "video_id": "EUOT1", + "question_id": "EUOT1", + "video_file": "Charades/videos/EUOT1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person flips on a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <4.9>." + } + ] + }, + { + "video_id": "EUOT1", + "question_id": "EUOT1", + "video_file": "Charades/videos/EUOT1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns on a light'." + }, + { + "from": "gpt", + "value": "From <0.5> to <4.9>." + } + ] + }, + { + "video_id": "Z0IE9", + "question_id": "Z0IE9", + "video_file": "Charades/videos/Z0IE9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person fixing a light'." + }, + { + "from": "gpt", + "value": "From <12.2> to <18.5>." + } + ] + }, + { + "video_id": "Z0IE9", + "question_id": "Z0IE9", + "video_file": "Charades/videos/Z0IE9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is fixing a light in a room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.2> to <18.5>." + } + ] + }, + { + "video_id": "Z0IE9", + "question_id": "Z0IE9", + "video_file": "Charades/videos/Z0IE9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person when he turns it the light turns on brightly'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <18.5>." + } + ] + }, + { + "video_id": "U72GG", + "question_id": "U72GG", + "video_file": "Charades/videos/U72GG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put detergent on the shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.7> to <20.4>." + } + ] + }, + { + "video_id": "U72GG", + "question_id": "U72GG", + "video_file": "Charades/videos/U72GG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person put some shoes on the dryer'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.7> to <20.4>." + } + ] + }, + { + "video_id": "D444N", + "question_id": "D444N", + "video_file": "Charades/videos/D444N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts laughing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <19.1>." + } + ] + }, + { + "video_id": "IOL8Q", + "question_id": "IOL8Q", + "video_file": "Charades/videos/IOL8Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is throwing a bag out of the room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.9> to <12.0>." + } + ] + }, + { + "video_id": "IOL8Q", + "question_id": "IOL8Q", + "video_file": "Charades/videos/IOL8Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is tidying inside a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "3YNPY", + "question_id": "3YNPY", + "video_file": "Charades/videos/3YNPY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts a towel on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.6> to <7.1>." + } + ] + }, + { + "video_id": "3YNPY", + "question_id": "3YNPY", + "video_file": "Charades/videos/3YNPY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits down in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.4> to <15.0>." + } + ] + }, + { + "video_id": "3YNPY", + "question_id": "3YNPY", + "video_file": "Charades/videos/3YNPY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person puts a towel onto the shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.6> to <7.1>." + } + ] + }, + { + "video_id": "8TIC7", + "question_id": "8TIC7", + "video_file": "Charades/videos/8TIC7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs out' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <12.2>." + } + ] + }, + { + "video_id": "8TIC7", + "question_id": "8TIC7", + "video_file": "Charades/videos/8TIC7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.1> to <15.3>." + } + ] + }, + { + "video_id": "8TIC7", + "question_id": "8TIC7", + "video_file": "Charades/videos/8TIC7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs into a bathroom'." + }, + { + "from": "gpt", + "value": "From <5.0> to <12.2>." + } + ] + }, + { + "video_id": "8TIC7", + "question_id": "8TIC7", + "video_file": "Charades/videos/8TIC7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes a camera from the bathroom counter'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <14.1>." + } + ] + }, + { + "video_id": "8TIC7", + "question_id": "8TIC7", + "video_file": "Charades/videos/8TIC7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs through the bathroom doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <12.2>." + } + ] + }, + { + "video_id": "7XKUI", + "question_id": "7XKUI", + "video_file": "Charades/videos/7XKUI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person closes the laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.6> to <10.5>." + } + ] + }, + { + "video_id": "7XKUI", + "question_id": "7XKUI", + "video_file": "Charades/videos/7XKUI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close the laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.6> to <10.5>." + } + ] + }, + { + "video_id": "7XKUI", + "question_id": "7XKUI", + "video_file": "Charades/videos/7XKUI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person closes their laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.6> to <10.5>." + } + ] + }, + { + "video_id": "3V2HJ", + "question_id": "3V2HJ", + "video_file": "Charades/videos/3V2HJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts their bag on the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <15.0>." + } + ] + }, + { + "video_id": "3V2HJ", + "question_id": "3V2HJ", + "video_file": "Charades/videos/3V2HJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person she puts something on the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.4> to <15.0>." + } + ] + }, + { + "video_id": "PN7CX", + "question_id": "PN7CX", + "video_file": "Charades/videos/PN7CX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts shoes on'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <20.5>." + } + ] + }, + { + "video_id": "PN7CX", + "question_id": "PN7CX", + "video_file": "Charades/videos/PN7CX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts on their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <20.5>." + } + ] + }, + { + "video_id": "0J5T4", + "question_id": "0J5T4", + "video_file": "Charades/videos/0J5T4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes some food from the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.0> to <10.0>." + } + ] + }, + { + "video_id": "0J5T4", + "question_id": "0J5T4", + "video_file": "Charades/videos/0J5T4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes out some food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <10.0>." + } + ] + }, + { + "video_id": "1RNK6", + "question_id": "1RNK6", + "video_file": "Charades/videos/1RNK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is watching something out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "1RNK6", + "question_id": "1RNK6", + "video_file": "Charades/videos/1RNK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person running with a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.2> to <14.2>." + } + ] + }, + { + "video_id": "1RNK6", + "question_id": "1RNK6", + "video_file": "Charades/videos/1RNK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is looking out a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "1RNK6", + "question_id": "1RNK6", + "video_file": "Charades/videos/1RNK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take out a box'." + }, + { + "from": "gpt", + "value": "From <7.2> to <12.5>." + } + ] + }, + { + "video_id": "1RNK6", + "question_id": "1RNK6", + "video_file": "Charades/videos/1RNK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person run out of the room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.2> to <14.2>." + } + ] + }, + { + "video_id": "V9JJL", + "question_id": "V9JJL", + "video_file": "Charades/videos/V9JJL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <17.0>." + } + ] + }, + { + "video_id": "4PCSC", + "question_id": "4PCSC", + "video_file": "Charades/videos/4PCSC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'throws clothes to another person awakening from under a blanket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.0> to <23.0>." + } + ] + }, + { + "video_id": "4PCSC", + "question_id": "4PCSC", + "video_file": "Charades/videos/4PCSC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person is pouring water into a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "4PCSC", + "question_id": "4PCSC", + "video_file": "Charades/videos/4PCSC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pouring a glass of water from a jar' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "4PCSC", + "question_id": "4PCSC", + "video_file": "Charades/videos/4PCSC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'wakes up as person 1 throws clothes on them' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.0> to <21.8>." + } + ] + }, + { + "video_id": "SFH7G", + "question_id": "SFH7G", + "video_file": "Charades/videos/SFH7G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is putting food on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.6> to <8.8>." + } + ] + }, + { + "video_id": "XC9EY", + "question_id": "XC9EY", + "video_file": "Charades/videos/XC9EY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts a glass down on a table'." + }, + { + "from": "gpt", + "value": "From <15.9> to <21.5>." + } + ] + }, + { + "video_id": "XC9EY", + "question_id": "XC9EY", + "video_file": "Charades/videos/XC9EY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting a glass on a table'." + }, + { + "from": "gpt", + "value": "From <15.9> to <21.5>." + } + ] + }, + { + "video_id": "OY50Q", + "question_id": "OY50Q", + "video_file": "Charades/videos/OY50Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person walks through the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <18.2>." + } + ] + }, + { + "video_id": "OY50Q", + "question_id": "OY50Q", + "video_file": "Charades/videos/OY50Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the door behind them'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.7> to <18.8>." + } + ] + }, + { + "video_id": "K8MMX", + "question_id": "K8MMX", + "video_file": "Charades/videos/K8MMX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person enters the room holding some papers' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.2> to <10.1>." + } + ] + }, + { + "video_id": "K8MMX", + "question_id": "K8MMX", + "video_file": "Charades/videos/K8MMX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts a piece of paper on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.1> to <9.8>." + } + ] + }, + { + "video_id": "K8MMX", + "question_id": "K8MMX", + "video_file": "Charades/videos/K8MMX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person opens the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "KNIZK", + "question_id": "KNIZK", + "video_file": "Charades/videos/KNIZK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person awakens in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.6> to <31.0>." + } + ] + }, + { + "video_id": "ZGHLY", + "question_id": "ZGHLY", + "video_file": "Charades/videos/ZGHLY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they open the laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "P4WRI", + "question_id": "P4WRI", + "video_file": "Charades/videos/P4WRI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person in a bathroom watches themselves in a mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.8> to <18.8>." + } + ] + }, + { + "video_id": "LGS4C", + "question_id": "LGS4C", + "video_file": "Charades/videos/LGS4C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs in through their doorway'." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.3>." + } + ] + }, + { + "video_id": "LGS4C", + "question_id": "LGS4C", + "video_file": "Charades/videos/LGS4C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person slightly closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.8> to <24.5>." + } + ] + }, + { + "video_id": "LGS4C", + "question_id": "LGS4C", + "video_file": "Charades/videos/LGS4C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is running towards a room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.3>." + } + ] + }, + { + "video_id": "LGS4C", + "question_id": "LGS4C", + "video_file": "Charades/videos/LGS4C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is running into the room holding their phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.3>." + } + ] + }, + { + "video_id": "TF95Y", + "question_id": "TF95Y", + "video_file": "Charades/videos/TF95Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <15.6>." + } + ] + }, + { + "video_id": "TF95Y", + "question_id": "TF95Y", + "video_file": "Charades/videos/TF95Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <15.6>." + } + ] + }, + { + "video_id": "39SXL", + "question_id": "39SXL", + "video_file": "Charades/videos/39SXL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person watches some television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <31.7> to <42.0>." + } + ] + }, + { + "video_id": "VGAMG", + "question_id": "VGAMG", + "video_file": "Charades/videos/VGAMG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is in a office drinking from a glass'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.8>." + } + ] + }, + { + "video_id": "VGAMG", + "question_id": "VGAMG", + "video_file": "Charades/videos/VGAMG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sitting in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.5> to <29.3>." + } + ] + }, + { + "video_id": "VGAMG", + "question_id": "VGAMG", + "video_file": "Charades/videos/VGAMG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking from a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.8>." + } + ] + }, + { + "video_id": "VGAMG", + "question_id": "VGAMG", + "video_file": "Charades/videos/VGAMG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting in a chair at a desk' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.5> to <29.3>." + } + ] + }, + { + "video_id": "FYJHS", + "question_id": "FYJHS", + "video_file": "Charades/videos/FYJHS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person to start fixing the doorknob' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.8> to <28.5>." + } + ] + }, + { + "video_id": "FYJHS", + "question_id": "FYJHS", + "video_file": "Charades/videos/FYJHS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person walks holding a bag of groceries' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.5> to <14.3>." + } + ] + }, + { + "video_id": "1IIS5", + "question_id": "1IIS5", + "video_file": "Charades/videos/1IIS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.0> to <9.8>." + } + ] + }, + { + "video_id": "BEUCN", + "question_id": "BEUCN", + "video_file": "Charades/videos/BEUCN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.2> to <17.7>." + } + ] + }, + { + "video_id": "8W31Y", + "question_id": "8W31Y", + "video_file": "Charades/videos/8W31Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating their favorite kind of candy' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.3> to <7.5>." + } + ] + }, + { + "video_id": "9FHVW", + "question_id": "9FHVW", + "video_file": "Charades/videos/9FHVW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins to drink from the cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.6> to <30.8>." + } + ] + }, + { + "video_id": "9FHVW", + "question_id": "9FHVW", + "video_file": "Charades/videos/9FHVW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the other person opens the cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.0> to <21.5>." + } + ] + }, + { + "video_id": "E7QDZ", + "question_id": "E7QDZ", + "video_file": "Charades/videos/E7QDZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person working on putting the light back on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <22.6>." + } + ] + }, + { + "video_id": "E7QDZ", + "question_id": "E7QDZ", + "video_file": "Charades/videos/E7QDZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turned on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <22.6>." + } + ] + }, + { + "video_id": "E7QDZ", + "question_id": "E7QDZ", + "video_file": "Charades/videos/E7QDZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is laughing at themselves in the mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <24.8>." + } + ] + }, + { + "video_id": "2DPEC", + "question_id": "2DPEC", + "video_file": "Charades/videos/2DPEC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person poured some water in the glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.1> to <24.1>." + } + ] + }, + { + "video_id": "2DPEC", + "question_id": "2DPEC", + "video_file": "Charades/videos/2DPEC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person watching a movie on their laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.1> to <32.0>." + } + ] + }, + { + "video_id": "2DPEC", + "question_id": "2DPEC", + "video_file": "Charades/videos/2DPEC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person watching a movie in the laptop'." + }, + { + "from": "gpt", + "value": "From <22.1> to <32.0>." + } + ] + }, + { + "video_id": "EXPOA", + "question_id": "EXPOA", + "video_file": "Charades/videos/EXPOA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person undresses from work clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.4> to <27.7>." + } + ] + }, + { + "video_id": "EXPOA", + "question_id": "EXPOA", + "video_file": "Charades/videos/EXPOA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person dresses with comfortable clothes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.3> to <36.0>." + } + ] + }, + { + "video_id": "FDNWD", + "question_id": "FDNWD", + "video_file": "Charades/videos/FDNWD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person standing in the doorway sneezing several times in elbow'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "FDNWD", + "question_id": "FDNWD", + "video_file": "Charades/videos/FDNWD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "52MV9", + "question_id": "52MV9", + "video_file": "Charades/videos/52MV9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walks through doorway into dining room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <11.5>." + } + ] + }, + { + "video_id": "52MV9", + "question_id": "52MV9", + "video_file": "Charades/videos/52MV9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a bite of the sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.8> to <24.6>." + } + ] + }, + { + "video_id": "52MV9", + "question_id": "52MV9", + "video_file": "Charades/videos/52MV9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person holding a plate open a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.1> to <11.5>." + } + ] + }, + { + "video_id": "52MV9", + "question_id": "52MV9", + "video_file": "Charades/videos/52MV9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a sandwich'." + }, + { + "from": "gpt", + "value": "From <7.9> to <14.0>." + } + ] + }, + { + "video_id": "52MV9", + "question_id": "52MV9", + "video_file": "Charades/videos/52MV9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.2> to <8.4>." + } + ] + }, + { + "video_id": "52MV9", + "question_id": "52MV9", + "video_file": "Charades/videos/52MV9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walks into the kitchen eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.4> to <32.7>." + } + ] + }, + { + "video_id": "RSFII", + "question_id": "RSFII", + "video_file": "Charades/videos/RSFII.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <18.9>." + } + ] + }, + { + "video_id": "4WYQ5", + "question_id": "4WYQ5", + "video_file": "Charades/videos/4WYQ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they throw the bag into a garbage box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.7> to <36.6>." + } + ] + }, + { + "video_id": "6LBDM", + "question_id": "6LBDM", + "video_file": "Charades/videos/6LBDM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looks out a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <34.4>." + } + ] + }, + { + "video_id": "6LBDM", + "question_id": "6LBDM", + "video_file": "Charades/videos/6LBDM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person began washing the window furiously'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.0> to <48.4>." + } + ] + }, + { + "video_id": "TMTAC", + "question_id": "TMTAC", + "video_file": "Charades/videos/TMTAC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts a bag on the counter'." + }, + { + "from": "gpt", + "value": "From <4.5> to <9.7>." + } + ] + }, + { + "video_id": "TMTAC", + "question_id": "TMTAC", + "video_file": "Charades/videos/TMTAC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they begin opening a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.8> to <11.4>." + } + ] + }, + { + "video_id": "JY0I5", + "question_id": "JY0I5", + "video_file": "Charades/videos/JY0I5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes a towel out of a bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.8> to <9.0>." + } + ] + }, + { + "video_id": "V1SQH", + "question_id": "V1SQH", + "video_file": "Charades/videos/V1SQH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person a girl drinks from a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <20.8>." + } + ] + }, + { + "video_id": "V1SQH", + "question_id": "V1SQH", + "video_file": "Charades/videos/V1SQH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is drinking a cup of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <20.8>." + } + ] + }, + { + "video_id": "V1SQH", + "question_id": "V1SQH", + "video_file": "Charades/videos/V1SQH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person drinks from the cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <20.8>." + } + ] + }, + { + "video_id": "Y4P3C", + "question_id": "Y4P3C", + "video_file": "Charades/videos/Y4P3C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <24.0>." + } + ] + }, + { + "video_id": "Y4P3C", + "question_id": "Y4P3C", + "video_file": "Charades/videos/Y4P3C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person was drinking a glass of water'." + }, + { + "from": "gpt", + "value": "From <0.4> to <4.6>." + } + ] + }, + { + "video_id": "Y4P3C", + "question_id": "Y4P3C", + "video_file": "Charades/videos/Y4P3C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a drink out of a glass'." + }, + { + "from": "gpt", + "value": "From <0.4> to <4.6>." + } + ] + }, + { + "video_id": "6J00B", + "question_id": "6J00B", + "video_file": "Charades/videos/6J00B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pour a glass of water'." + }, + { + "from": "gpt", + "value": "From <17.8> to <23.8>." + } + ] + }, + { + "video_id": "6J00B", + "question_id": "6J00B", + "video_file": "Charades/videos/6J00B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats something' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.8> to <33.0>." + } + ] + }, + { + "video_id": "6J00B", + "question_id": "6J00B", + "video_file": "Charades/videos/6J00B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take a bite of their food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.7> to <28.3>." + } + ] + }, + { + "video_id": "XP6CJ", + "question_id": "XP6CJ", + "video_file": "Charades/videos/XP6CJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they stand up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.2> to <14.4>." + } + ] + }, + { + "video_id": "XP6CJ", + "question_id": "XP6CJ", + "video_file": "Charades/videos/XP6CJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sits on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.2> to <13.4>." + } + ] + }, + { + "video_id": "SFDTP", + "question_id": "SFDTP", + "video_file": "Charades/videos/SFDTP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes clothes out of the box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.8> to <22.3>." + } + ] + }, + { + "video_id": "VMOBC", + "question_id": "VMOBC", + "video_file": "Charades/videos/VMOBC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person forth eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.5> to <20.5>." + } + ] + }, + { + "video_id": "VMOBC", + "question_id": "VMOBC", + "video_file": "Charades/videos/VMOBC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.9> to <26.0>." + } + ] + }, + { + "video_id": "VMOBC", + "question_id": "VMOBC", + "video_file": "Charades/videos/VMOBC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <20.5>." + } + ] + }, + { + "video_id": "SB7CW", + "question_id": "SB7CW", + "video_file": "Charades/videos/SB7CW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the cabinet door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.7> to <7.5>." + } + ] + }, + { + "video_id": "SB7CW", + "question_id": "SB7CW", + "video_file": "Charades/videos/SB7CW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens the cabinet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.7> to <7.5>." + } + ] + }, + { + "video_id": "SB7CW", + "question_id": "SB7CW", + "video_file": "Charades/videos/SB7CW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person wrapped in a blanket runs in sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.7> to <10.0>." + } + ] + }, + { + "video_id": "SB7CW", + "question_id": "SB7CW", + "video_file": "Charades/videos/SB7CW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a door to get some medicine' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.7> to <7.5>." + } + ] + }, + { + "video_id": "KNN3G", + "question_id": "KNN3G", + "video_file": "Charades/videos/KNN3G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pouring a cup of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.7> to <22.0>." + } + ] + }, + { + "video_id": "KNN3G", + "question_id": "KNN3G", + "video_file": "Charades/videos/KNN3G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating food in a dining room'." + }, + { + "from": "gpt", + "value": "From <22.7> to <31.0>." + } + ] + }, + { + "video_id": "KNN3G", + "question_id": "KNN3G", + "video_file": "Charades/videos/KNN3G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take some food from a plate'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.4> to <31.0>." + } + ] + }, + { + "video_id": "KNN3G", + "question_id": "KNN3G", + "video_file": "Charades/videos/KNN3G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pours some coffee into a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.7> to <22.0>." + } + ] + }, + { + "video_id": "KNN3G", + "question_id": "KNN3G", + "video_file": "Charades/videos/KNN3G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin eating' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.7> to <31.0>." + } + ] + }, + { + "video_id": "MMTED", + "question_id": "MMTED", + "video_file": "Charades/videos/MMTED.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person lifts up their head to look out a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.7> to <23.0>." + } + ] + }, + { + "video_id": "MMTED", + "question_id": "MMTED", + "video_file": "Charades/videos/MMTED.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person look out the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.7> to <23.0>." + } + ] + }, + { + "video_id": "0GR11", + "question_id": "0GR11", + "video_file": "Charades/videos/0GR11.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they close the laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.8> to <24.7>." + } + ] + }, + { + "video_id": "0GR11", + "question_id": "0GR11", + "video_file": "Charades/videos/0GR11.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person is working on a laptop computer' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.3> to <25.3>." + } + ] + }, + { + "video_id": "0GR11", + "question_id": "0GR11", + "video_file": "Charades/videos/0GR11.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eat a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.4> to <32.0>." + } + ] + }, + { + "video_id": "0GR11", + "question_id": "0GR11", + "video_file": "Charades/videos/0GR11.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take a break to bite their sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.2> to <26.4>." + } + ] + }, + { + "video_id": "E4FN6", + "question_id": "E4FN6", + "video_file": "Charades/videos/E4FN6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person is standing in the bathroom washing their hands'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <24.0>." + } + ] + }, + { + "video_id": "75KQ6", + "question_id": "75KQ6", + "video_file": "Charades/videos/75KQ6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person make a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.5> to <30.0>." + } + ] + }, + { + "video_id": "75KQ6", + "question_id": "75KQ6", + "video_file": "Charades/videos/75KQ6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is in the kitchen washing dishes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.5>." + } + ] + }, + { + "video_id": "75KQ6", + "question_id": "75KQ6", + "video_file": "Charades/videos/75KQ6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is washing dishes in a sink' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.5>." + } + ] + }, + { + "video_id": "4ETXZ", + "question_id": "4ETXZ", + "video_file": "Charades/videos/4ETXZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns the light on'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.2> to <15.4>." + } + ] + }, + { + "video_id": "4ETXZ", + "question_id": "4ETXZ", + "video_file": "Charades/videos/4ETXZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person the light is off' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.2> to <15.4>." + } + ] + }, + { + "video_id": "4ETXZ", + "question_id": "4ETXZ", + "video_file": "Charades/videos/4ETXZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person walks over to the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.2> to <15.4>." + } + ] + }, + { + "video_id": "1TGKL", + "question_id": "1TGKL", + "video_file": "Charades/videos/1TGKL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drink something out of the glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <26.5>." + } + ] + }, + { + "video_id": "1TGKL", + "question_id": "1TGKL", + "video_file": "Charades/videos/1TGKL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinks something from a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <26.5>." + } + ] + }, + { + "video_id": "J88MJ", + "question_id": "J88MJ", + "video_file": "Charades/videos/J88MJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns off light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <28.4>." + } + ] + }, + { + "video_id": "J88MJ", + "question_id": "J88MJ", + "video_file": "Charades/videos/J88MJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they turn off a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <28.4>." + } + ] + }, + { + "video_id": "J88MJ", + "question_id": "J88MJ", + "video_file": "Charades/videos/J88MJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put towel up to face' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.9> to <22.6>." + } + ] + }, + { + "video_id": "J88MJ", + "question_id": "J88MJ", + "video_file": "Charades/videos/J88MJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they laugh' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <21.5>." + } + ] + }, + { + "video_id": "J88MJ", + "question_id": "J88MJ", + "video_file": "Charades/videos/J88MJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <21.5>." + } + ] + }, + { + "video_id": "X62WQ", + "question_id": "X62WQ", + "video_file": "Charades/videos/X62WQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they stand up'." + }, + { + "from": "gpt", + "value": "From <4.4> to <11.9>." + } + ] + }, + { + "video_id": "X62WQ", + "question_id": "X62WQ", + "video_file": "Charades/videos/X62WQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person still holding the phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <19.2>." + } + ] + }, + { + "video_id": "DR4R7", + "question_id": "DR4R7", + "video_file": "Charades/videos/DR4R7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put their laptop on the bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <9.0>." + } + ] + }, + { + "video_id": "LY10X", + "question_id": "LY10X", + "video_file": "Charades/videos/LY10X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take some medicine'." + }, + { + "from": "gpt", + "value": "From <24.1> to <31.6>." + } + ] + }, + { + "video_id": "LY10X", + "question_id": "LY10X", + "video_file": "Charades/videos/LY10X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes out medicine'." + }, + { + "from": "gpt", + "value": "From <24.1> to <31.6>." + } + ] + }, + { + "video_id": "LY10X", + "question_id": "LY10X", + "video_file": "Charades/videos/LY10X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.5> to <20.7>." + } + ] + }, + { + "video_id": "LY10X", + "question_id": "LY10X", + "video_file": "Charades/videos/LY10X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person washing dishes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <16.4>." + } + ] + }, + { + "video_id": "COT5L", + "question_id": "COT5L", + "video_file": "Charades/videos/COT5L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is holding a cup of coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.4> to <41.4>." + } + ] + }, + { + "video_id": "WESPT", + "question_id": "WESPT", + "video_file": "Charades/videos/WESPT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looks out the door window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <11.4>." + } + ] + }, + { + "video_id": "WESPT", + "question_id": "WESPT", + "video_file": "Charades/videos/WESPT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looks out a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <11.4>." + } + ] + }, + { + "video_id": "WESPT", + "question_id": "WESPT", + "video_file": "Charades/videos/WESPT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <14.1>." + } + ] + }, + { + "video_id": "WESPT", + "question_id": "WESPT", + "video_file": "Charades/videos/WESPT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person close the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.3> to <24.0>." + } + ] + }, + { + "video_id": "XLQ6X", + "question_id": "XLQ6X", + "video_file": "Charades/videos/XLQ6X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walks back into the room holding some blankets' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.2> to <13.1>." + } + ] + }, + { + "video_id": "XLQ6X", + "question_id": "XLQ6X", + "video_file": "Charades/videos/XLQ6X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws blankets onto sofa' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.3> to <15.7>." + } + ] + }, + { + "video_id": "XLQ6X", + "question_id": "XLQ6X", + "video_file": "Charades/videos/XLQ6X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person throws a pillow'." + }, + { + "from": "gpt", + "value": "From <13.0> to <18.0>." + } + ] + }, + { + "video_id": "XLQ6X", + "question_id": "XLQ6X", + "video_file": "Charades/videos/XLQ6X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.3> to <23.2>." + } + ] + }, + { + "video_id": "XLQ6X", + "question_id": "XLQ6X", + "video_file": "Charades/videos/XLQ6X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks through doorway out of room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.7> to <21.1>." + } + ] + }, + { + "video_id": "BQXVG", + "question_id": "BQXVG", + "video_file": "Charades/videos/BQXVG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs out the bathroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.2> to <18.1>." + } + ] + }, + { + "video_id": "BQXVG", + "question_id": "BQXVG", + "video_file": "Charades/videos/BQXVG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person quickly runs from the room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.2> to <18.1>." + } + ] + }, + { + "video_id": "ETQYM", + "question_id": "ETQYM", + "video_file": "Charades/videos/ETQYM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sneezing in a mirror'." + }, + { + "from": "gpt", + "value": "From <15.2> to <24.4>." + } + ] + }, + { + "video_id": "ETQYM", + "question_id": "ETQYM", + "video_file": "Charades/videos/ETQYM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <24.4>." + } + ] + }, + { + "video_id": "9GLQZ", + "question_id": "9GLQZ", + "video_file": "Charades/videos/9GLQZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person holding a box runs up stairs'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "C3CGB", + "question_id": "C3CGB", + "video_file": "Charades/videos/C3CGB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.7> to <10.7>." + } + ] + }, + { + "video_id": "O6FSN", + "question_id": "O6FSN", + "video_file": "Charades/videos/O6FSN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is opening the refrigerator door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "O6FSN", + "question_id": "O6FSN", + "video_file": "Charades/videos/O6FSN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person in a kitchen opens a refrigerator door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "O6FSN", + "question_id": "O6FSN", + "video_file": "Charades/videos/O6FSN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a plate of food out' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.3> to <24.5>." + } + ] + }, + { + "video_id": "0DDYA", + "question_id": "0DDYA", + "video_file": "Charades/videos/0DDYA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person turns on a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.5> to <9.9>." + } + ] + }, + { + "video_id": "0DDYA", + "question_id": "0DDYA", + "video_file": "Charades/videos/0DDYA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts on a pair of shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.1> to <37.1>." + } + ] + }, + { + "video_id": "0DDYA", + "question_id": "0DDYA", + "video_file": "Charades/videos/0DDYA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts their shoes on' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.1> to <37.1>." + } + ] + }, + { + "video_id": "0DDYA", + "question_id": "0DDYA", + "video_file": "Charades/videos/0DDYA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person turns off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <34.8> to <39.9>." + } + ] + }, + { + "video_id": "0DDYA", + "question_id": "0DDYA", + "video_file": "Charades/videos/0DDYA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is watching themselves in a mirror'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.6> to <27.3>." + } + ] + }, + { + "video_id": "ZK05Q", + "question_id": "ZK05Q", + "video_file": "Charades/videos/ZK05Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person takes a book from a bathroom shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.2> to <15.0>." + } + ] + }, + { + "video_id": "2URXS", + "question_id": "2URXS", + "video_file": "Charades/videos/2URXS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sneezing into their sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.7> to <34.0>." + } + ] + }, + { + "video_id": "GK7S7", + "question_id": "GK7S7", + "video_file": "Charades/videos/GK7S7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "GK7S7", + "question_id": "GK7S7", + "video_file": "Charades/videos/GK7S7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is holding a camera'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "MEZVN", + "question_id": "MEZVN", + "video_file": "Charades/videos/MEZVN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a sneezing person opens a door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "L0CK7", + "question_id": "L0CK7", + "video_file": "Charades/videos/L0CK7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits on the bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.8> to <32.0>." + } + ] + }, + { + "video_id": "L0CK7", + "question_id": "L0CK7", + "video_file": "Charades/videos/L0CK7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person sits down on a bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.8> to <32.0>." + } + ] + }, + { + "video_id": "5N7EW", + "question_id": "5N7EW", + "video_file": "Charades/videos/5N7EW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opened a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.5> to <8.9>." + } + ] + }, + { + "video_id": "41EQS", + "question_id": "41EQS", + "video_file": "Charades/videos/41EQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person throws a towel in a bag'." + }, + { + "from": "gpt", + "value": "From <6.8> to <13.8>." + } + ] + }, + { + "video_id": "41EQS", + "question_id": "41EQS", + "video_file": "Charades/videos/41EQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws the towel into the bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.8> to <13.8>." + } + ] + }, + { + "video_id": "41EQS", + "question_id": "41EQS", + "video_file": "Charades/videos/41EQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person is holding a towel'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "TU2GE", + "question_id": "TU2GE", + "video_file": "Charades/videos/TU2GE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person talking on the phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "TU2GE", + "question_id": "TU2GE", + "video_file": "Charades/videos/TU2GE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is talking on a phone'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "FCBEQ", + "question_id": "FCBEQ", + "video_file": "Charades/videos/FCBEQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person pours some food into a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.1> to <13.0>." + } + ] + }, + { + "video_id": "FCBEQ", + "question_id": "FCBEQ", + "video_file": "Charades/videos/FCBEQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is seen pouring food in a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <13.0>." + } + ] + }, + { + "video_id": "HHCFO", + "question_id": "HHCFO", + "video_file": "Charades/videos/HHCFO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turn off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.9> to <32.6>." + } + ] + }, + { + "video_id": "HHCFO", + "question_id": "HHCFO", + "video_file": "Charades/videos/HHCFO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person flipping off the hall light as they pass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.9> to <32.6>." + } + ] + }, + { + "video_id": "HHCFO", + "question_id": "HHCFO", + "video_file": "Charades/videos/HHCFO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turn off the lights' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.9> to <32.6>." + } + ] + }, + { + "video_id": "HHCFO", + "question_id": "HHCFO", + "video_file": "Charades/videos/HHCFO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they open the door'." + }, + { + "from": "gpt", + "value": "From <28.3> to <33.5>." + } + ] + }, + { + "video_id": "XIUJE", + "question_id": "XIUJE", + "video_file": "Charades/videos/XIUJE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throwing dirty clothes behind the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.0> to <14.1>." + } + ] + }, + { + "video_id": "XIUJE", + "question_id": "XIUJE", + "video_file": "Charades/videos/XIUJE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person starts undressing in front of a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <14.3>." + } + ] + }, + { + "video_id": "SEDKM", + "question_id": "SEDKM", + "video_file": "Charades/videos/SEDKM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person laughs some more' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.5> to <6.5>." + } + ] + }, + { + "video_id": "SEDKM", + "question_id": "SEDKM", + "video_file": "Charades/videos/SEDKM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person laughing at a picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.5> to <6.5>." + } + ] + }, + { + "video_id": "SEDKM", + "question_id": "SEDKM", + "video_file": "Charades/videos/SEDKM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person with hair in a towel laughs'." + }, + { + "from": "gpt", + "value": "From <0.5> to <6.5>." + } + ] + }, + { + "video_id": "NEE6E", + "question_id": "NEE6E", + "video_file": "Charades/videos/NEE6E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is smiling at a picture on a desk' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "NEE6E", + "question_id": "NEE6E", + "video_file": "Charades/videos/NEE6E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person is awakening on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.9> to <29.0>." + } + ] + }, + { + "video_id": "DIDRB", + "question_id": "DIDRB", + "video_file": "Charades/videos/DIDRB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is pouring a drink in a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.0>." + } + ] + }, + { + "video_id": "342XO", + "question_id": "342XO", + "video_file": "Charades/videos/342XO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.9> to <34.0>." + } + ] + }, + { + "video_id": "342XO", + "question_id": "342XO", + "video_file": "Charades/videos/342XO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the coffee cup down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.2> to <29.4>." + } + ] + }, + { + "video_id": "ZJPK8", + "question_id": "ZJPK8", + "video_file": "Charades/videos/ZJPK8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person carrying some clothes opens a door'." + }, + { + "from": "gpt", + "value": "From <5.5> to <15.0>." + } + ] + }, + { + "video_id": "ZJPK8", + "question_id": "ZJPK8", + "video_file": "Charades/videos/ZJPK8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is opening the door to the bedroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.5> to <15.0>." + } + ] + }, + { + "video_id": "ZJPK8", + "question_id": "ZJPK8", + "video_file": "Charades/videos/ZJPK8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts a blanket on top of the pile'." + }, + { + "from": "gpt", + "value": "From <24.2> to <30.8>." + } + ] + }, + { + "video_id": "0ACZ8", + "question_id": "0ACZ8", + "video_file": "Charades/videos/0ACZ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is smiling in the mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "J06RS", + "question_id": "J06RS", + "video_file": "Charades/videos/J06RS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person in a hall smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.0>." + } + ] + }, + { + "video_id": "J06RS", + "question_id": "J06RS", + "video_file": "Charades/videos/J06RS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is smiling' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.0>." + } + ] + }, + { + "video_id": "J06RS", + "question_id": "J06RS", + "video_file": "Charades/videos/J06RS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person laughing bends over' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.7> to <9.0>." + } + ] + }, + { + "video_id": "J06RS", + "question_id": "J06RS", + "video_file": "Charades/videos/J06RS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is laughing in the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <9.0>." + } + ] + }, + { + "video_id": "YC9FL", + "question_id": "YC9FL", + "video_file": "Charades/videos/YC9FL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <9.8>." + } + ] + }, + { + "video_id": "YC9FL", + "question_id": "YC9FL", + "video_file": "Charades/videos/YC9FL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the door behind them' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.8> to <9.8>." + } + ] + }, + { + "video_id": "TBCJA", + "question_id": "TBCJA", + "video_file": "Charades/videos/TBCJA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they close the laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.9> to <23.2>." + } + ] + }, + { + "video_id": "TBCJA", + "question_id": "TBCJA", + "video_file": "Charades/videos/TBCJA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes up the pantry by grabbing the doorknob' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.3> to <29.6>." + } + ] + }, + { + "video_id": "TBCJA", + "question_id": "TBCJA", + "video_file": "Charades/videos/TBCJA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person shut the light off' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.0> to <25.3>." + } + ] + }, + { + "video_id": "HOI88", + "question_id": "HOI88", + "video_file": "Charades/videos/HOI88.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits on the sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.7> to <21.6>." + } + ] + }, + { + "video_id": "HOI88", + "question_id": "HOI88", + "video_file": "Charades/videos/HOI88.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.7> to <17.4>." + } + ] + }, + { + "video_id": "HOI88", + "question_id": "HOI88", + "video_file": "Charades/videos/HOI88.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person is sitting on the couch' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.7> to <21.6>." + } + ] + }, + { + "video_id": "2O6RB", + "question_id": "2O6RB", + "video_file": "Charades/videos/2O6RB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens the refrigerators' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.2> to <13.3>." + } + ] + }, + { + "video_id": "2O6RB", + "question_id": "2O6RB", + "video_file": "Charades/videos/2O6RB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a refrigerator'." + }, + { + "from": "gpt", + "value": "From <2.2> to <13.3>." + } + ] + }, + { + "video_id": "6QDZI", + "question_id": "6QDZI", + "video_file": "Charades/videos/6QDZI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person sits down on the bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.5> to <31.0>." + } + ] + }, + { + "video_id": "6QDZI", + "question_id": "6QDZI", + "video_file": "Charades/videos/6QDZI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opens a book'." + }, + { + "from": "gpt", + "value": "From <14.3> to <19.9>." + } + ] + }, + { + "video_id": "6QDZI", + "question_id": "6QDZI", + "video_file": "Charades/videos/6QDZI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts reading the book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.7> to <31.0>." + } + ] + }, + { + "video_id": "S1PWD", + "question_id": "S1PWD", + "video_file": "Charades/videos/S1PWD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the picture back on the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "S1PWD", + "question_id": "S1PWD", + "video_file": "Charades/videos/S1PWD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person takes a picture from a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "GQ4ML", + "question_id": "GQ4ML", + "video_file": "Charades/videos/GQ4ML.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts it on the shelf next to a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <19.1>." + } + ] + }, + { + "video_id": "EW6K9", + "question_id": "EW6K9", + "video_file": "Charades/videos/EW6K9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throwing clothes down stairs' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.6> to <19.4>." + } + ] + }, + { + "video_id": "WRW74", + "question_id": "WRW74", + "video_file": "Charades/videos/WRW74.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person turns on a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.7> to <13.1>." + } + ] + }, + { + "video_id": "WRW74", + "question_id": "WRW74", + "video_file": "Charades/videos/WRW74.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.7> to <13.1>." + } + ] + }, + { + "video_id": "WRW74", + "question_id": "WRW74", + "video_file": "Charades/videos/WRW74.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns the light off' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.7> to <13.1>." + } + ] + }, + { + "video_id": "3YVPG", + "question_id": "3YVPG", + "video_file": "Charades/videos/3YVPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting down the glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.4> to <25.5>." + } + ] + }, + { + "video_id": "3YVPG", + "question_id": "3YVPG", + "video_file": "Charades/videos/3YVPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person lays down on the floor with a blanket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <32.0> to <39.0>." + } + ] + }, + { + "video_id": "3YVPG", + "question_id": "3YVPG", + "video_file": "Charades/videos/3YVPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pours it in a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.5> to <18.3>." + } + ] + }, + { + "video_id": "3YVPG", + "question_id": "3YVPG", + "video_file": "Charades/videos/3YVPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pouring a glass of wine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.5> to <18.3>." + } + ] + }, + { + "video_id": "3YVPG", + "question_id": "3YVPG", + "video_file": "Charades/videos/3YVPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sneezing into a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.0> to <32.0>." + } + ] + }, + { + "video_id": "3YVPG", + "question_id": "3YVPG", + "video_file": "Charades/videos/3YVPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person pours a glass of wine' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.5> to <18.3>." + } + ] + }, + { + "video_id": "3YVPG", + "question_id": "3YVPG", + "video_file": "Charades/videos/3YVPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits on the floor with it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <30.5> to <35.7>." + } + ] + }, + { + "video_id": "5VUQE", + "question_id": "5VUQE", + "video_file": "Charades/videos/5VUQE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a napkin out of a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.0> to <19.0>." + } + ] + }, + { + "video_id": "1BEFM", + "question_id": "1BEFM", + "video_file": "Charades/videos/1BEFM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they eat a cookie' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <9.8>." + } + ] + }, + { + "video_id": "1BEFM", + "question_id": "1BEFM", + "video_file": "Charades/videos/1BEFM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they are eating something from a bowl' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.8> to <9.8>." + } + ] + }, + { + "video_id": "1BEFM", + "question_id": "1BEFM", + "video_file": "Charades/videos/1BEFM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating food'." + }, + { + "from": "gpt", + "value": "From <1.8> to <9.8>." + } + ] + }, + { + "video_id": "DVTAR", + "question_id": "DVTAR", + "video_file": "Charades/videos/DVTAR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person walks around eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.9> to <17.3>." + } + ] + }, + { + "video_id": "R0M34", + "question_id": "R0M34", + "video_file": "Charades/videos/R0M34.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person opens the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <19.5>." + } + ] + }, + { + "video_id": "R0M34", + "question_id": "R0M34", + "video_file": "Charades/videos/R0M34.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts cooking'." + }, + { + "from": "gpt", + "value": "From <0.1> to <15.3>." + } + ] + }, + { + "video_id": "R0M34", + "question_id": "R0M34", + "video_file": "Charades/videos/R0M34.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is standing by a stove cooking some food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.1> to <15.3>." + } + ] + }, + { + "video_id": "AAGZ2", + "question_id": "AAGZ2", + "video_file": "Charades/videos/AAGZ2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts cooking in his pot' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <33.8>." + } + ] + }, + { + "video_id": "AAGZ2", + "question_id": "AAGZ2", + "video_file": "Charades/videos/AAGZ2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts cooking on a stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.0> to <33.8>." + } + ] + }, + { + "video_id": "L1MR4", + "question_id": "L1MR4", + "video_file": "Charades/videos/L1MR4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person plays with their phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.5> to <28.0>." + } + ] + }, + { + "video_id": "2C253", + "question_id": "2C253", + "video_file": "Charades/videos/2C253.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they drink something from a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.6> to <23.6>." + } + ] + }, + { + "video_id": "2C253", + "question_id": "2C253", + "video_file": "Charades/videos/2C253.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person putting a blanket on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.2> to <27.0>." + } + ] + }, + { + "video_id": "2C253", + "question_id": "2C253", + "video_file": "Charades/videos/2C253.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they start drinking from a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.6> to <23.6>." + } + ] + }, + { + "video_id": "2C253", + "question_id": "2C253", + "video_file": "Charades/videos/2C253.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put a blanket down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.2> to <27.0>." + } + ] + }, + { + "video_id": "DSGVL", + "question_id": "DSGVL", + "video_file": "Charades/videos/DSGVL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits down on a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.1> to <21.0>." + } + ] + }, + { + "video_id": "DSGVL", + "question_id": "DSGVL", + "video_file": "Charades/videos/DSGVL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the laptop on the desk' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.1> to <16.3>." + } + ] + }, + { + "video_id": "GCWNT", + "question_id": "GCWNT", + "video_file": "Charades/videos/GCWNT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts eating'." + }, + { + "from": "gpt", + "value": "From <14.7> to <24.0>." + } + ] + }, + { + "video_id": "GCWNT", + "question_id": "GCWNT", + "video_file": "Charades/videos/GCWNT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats some food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <24.0>." + } + ] + }, + { + "video_id": "GCWNT", + "question_id": "GCWNT", + "video_file": "Charades/videos/GCWNT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closing the door'." + }, + { + "from": "gpt", + "value": "From <8.4> to <14.9>." + } + ] + }, + { + "video_id": "GCWNT", + "question_id": "GCWNT", + "video_file": "Charades/videos/GCWNT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person closes a closet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.4> to <14.9>." + } + ] + }, + { + "video_id": "0F0WE", + "question_id": "0F0WE", + "video_file": "Charades/videos/0F0WE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person runs back to the middle of the hall'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <10.4>." + } + ] + }, + { + "video_id": "0F0WE", + "question_id": "0F0WE", + "video_file": "Charades/videos/0F0WE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person runs down the hall checking doors' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <10.4>." + } + ] + }, + { + "video_id": "0F0WE", + "question_id": "0F0WE", + "video_file": "Charades/videos/0F0WE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts to read a notebook found on the floor'." + }, + { + "from": "gpt", + "value": "From <21.6> to <34.0>." + } + ] + }, + { + "video_id": "0F0WE", + "question_id": "0F0WE", + "video_file": "Charades/videos/0F0WE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is running down the hallway closing doors'." + }, + { + "from": "gpt", + "value": "From <6.5> to <13.9>." + } + ] + }, + { + "video_id": "T9ZNR", + "question_id": "T9ZNR", + "video_file": "Charades/videos/T9ZNR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes out their vacuum'." + }, + { + "from": "gpt", + "value": "From <2.6> to <11.3>." + } + ] + }, + { + "video_id": "F8HLI", + "question_id": "F8HLI", + "video_file": "Charades/videos/F8HLI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts running' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.8> to <21.9>." + } + ] + }, + { + "video_id": "F8HLI", + "question_id": "F8HLI", + "video_file": "Charades/videos/F8HLI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs out of the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.8> to <21.9>." + } + ] + }, + { + "video_id": "9RSGO", + "question_id": "9RSGO", + "video_file": "Charades/videos/9RSGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'other person looking in mirror fixing hair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.8> to <28.2>." + } + ] + }, + { + "video_id": "1HGEX", + "question_id": "1HGEX", + "video_file": "Charades/videos/1HGEX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts looking at the book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.0> to <31.0>." + } + ] + }, + { + "video_id": "1HGEX", + "question_id": "1HGEX", + "video_file": "Charades/videos/1HGEX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a book from the cabinet'." + }, + { + "from": "gpt", + "value": "From <16.8> to <24.4>." + } + ] + }, + { + "video_id": "EFD6M", + "question_id": "EFD6M", + "video_file": "Charades/videos/EFD6M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.8> to <19.1>." + } + ] + }, + { + "video_id": "5M17J", + "question_id": "5M17J", + "video_file": "Charades/videos/5M17J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting them onto a shelf in a nearby cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.1> to <32.4>." + } + ] + }, + { + "video_id": "5M17J", + "question_id": "5M17J", + "video_file": "Charades/videos/5M17J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening cabinet doors' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.2> to <26.6>." + } + ] + }, + { + "video_id": "5M17J", + "question_id": "5M17J", + "video_file": "Charades/videos/5M17J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person putting things on a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.1> to <32.4>." + } + ] + }, + { + "video_id": "5M17J", + "question_id": "5M17J", + "video_file": "Charades/videos/5M17J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person opens a box containing spices'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.7> to <18.6>." + } + ] + }, + { + "video_id": "O52BL", + "question_id": "O52BL", + "video_file": "Charades/videos/O52BL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts a laptop in a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.3> to <29.4>." + } + ] + }, + { + "video_id": "O52BL", + "question_id": "O52BL", + "video_file": "Charades/videos/O52BL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they take their laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <31.0>." + } + ] + }, + { + "video_id": "USNFB", + "question_id": "USNFB", + "video_file": "Charades/videos/USNFB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.3>." + } + ] + }, + { + "video_id": "USNFB", + "question_id": "USNFB", + "video_file": "Charades/videos/USNFB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.3>." + } + ] + }, + { + "video_id": "91YZH", + "question_id": "91YZH", + "video_file": "Charades/videos/91YZH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.6> to <11.1>." + } + ] + }, + { + "video_id": "R2YKH", + "question_id": "R2YKH", + "video_file": "Charades/videos/R2YKH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person plays with their phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.0> to <20.8>." + } + ] + }, + { + "video_id": "KU656", + "question_id": "KU656", + "video_file": "Charades/videos/KU656.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens in bed'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "4JXAK", + "question_id": "4JXAK", + "video_file": "Charades/videos/4JXAK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person starts washing a mirror' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <23.0>." + } + ] + }, + { + "video_id": "XPOKK", + "question_id": "XPOKK", + "video_file": "Charades/videos/XPOKK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sitting on couch looking a picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <10.8>." + } + ] + }, + { + "video_id": "XPOKK", + "question_id": "XPOKK", + "video_file": "Charades/videos/XPOKK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person looks at a picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <10.8>." + } + ] + }, + { + "video_id": "M5YLS", + "question_id": "M5YLS", + "video_file": "Charades/videos/M5YLS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <22.2>." + } + ] + }, + { + "video_id": "M5YLS", + "question_id": "M5YLS", + "video_file": "Charades/videos/M5YLS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they turn on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <9.9>." + } + ] + }, + { + "video_id": "M5YLS", + "question_id": "M5YLS", + "video_file": "Charades/videos/M5YLS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a smiling person is sitting in their bathroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.3> to <22.2>." + } + ] + }, + { + "video_id": "5XCUV", + "question_id": "5XCUV", + "video_file": "Charades/videos/5XCUV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they are drinking a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "5XCUV", + "question_id": "5XCUV", + "video_file": "Charades/videos/5XCUV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person take off their shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.9> to <29.0>." + } + ] + }, + { + "video_id": "MCQO5", + "question_id": "MCQO5", + "video_file": "Charades/videos/MCQO5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is undressing in the closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.2> to <13.4>." + } + ] + }, + { + "video_id": "MCQO5", + "question_id": "MCQO5", + "video_file": "Charades/videos/MCQO5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drink a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.6> to <33.0>." + } + ] + }, + { + "video_id": "MCQO5", + "question_id": "MCQO5", + "video_file": "Charades/videos/MCQO5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drink from a bottle of water in a bedroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <29.6> to <33.0>." + } + ] + }, + { + "video_id": "RR8M2", + "question_id": "RR8M2", + "video_file": "Charades/videos/RR8M2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a cabinet door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.2> to <26.7>." + } + ] + }, + { + "video_id": "RR8M2", + "question_id": "RR8M2", + "video_file": "Charades/videos/RR8M2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the shoes on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.0> to <29.0>." + } + ] + }, + { + "video_id": "GZQ0V", + "question_id": "GZQ0V", + "video_file": "Charades/videos/GZQ0V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walk towards the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.4> to <24.3>." + } + ] + }, + { + "video_id": "IKZJE", + "question_id": "IKZJE", + "video_file": "Charades/videos/IKZJE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a photo of the picture' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.0> to <33.8>." + } + ] + }, + { + "video_id": "R6CIX", + "question_id": "R6CIX", + "video_file": "Charades/videos/R6CIX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is fixing a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "R6CIX", + "question_id": "R6CIX", + "video_file": "Charades/videos/R6CIX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits down on a couch' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.7> to <20.1>." + } + ] + }, + { + "video_id": "R6CIX", + "question_id": "R6CIX", + "video_file": "Charades/videos/R6CIX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits on a sofa' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.7> to <20.1>." + } + ] + }, + { + "video_id": "X8AP2", + "question_id": "X8AP2", + "video_file": "Charades/videos/X8AP2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is standing in a kitchen eating something' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <12.9>." + } + ] + }, + { + "video_id": "X8AP2", + "question_id": "X8AP2", + "video_file": "Charades/videos/X8AP2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is eating'." + }, + { + "from": "gpt", + "value": "From <1.5> to <12.9>." + } + ] + }, + { + "video_id": "X8AP2", + "question_id": "X8AP2", + "video_file": "Charades/videos/X8AP2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes opens a medicine bottle'." + }, + { + "from": "gpt", + "value": "From <22.3> to <31.0>." + } + ] + }, + { + "video_id": "KPXQ0", + "question_id": "KPXQ0", + "video_file": "Charades/videos/KPXQ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.2> to <24.4>." + } + ] + }, + { + "video_id": "H76DJ", + "question_id": "H76DJ", + "video_file": "Charades/videos/H76DJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throwing a pillow on the floor'." + }, + { + "from": "gpt", + "value": "From <27.3> to <33.0>." + } + ] + }, + { + "video_id": "H76DJ", + "question_id": "H76DJ", + "video_file": "Charades/videos/H76DJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws a pillow on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <27.3> to <33.0>." + } + ] + }, + { + "video_id": "FCQS8", + "question_id": "FCQS8", + "video_file": "Charades/videos/FCQS8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks through the doorway of a bathroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "FCQS8", + "question_id": "FCQS8", + "video_file": "Charades/videos/FCQS8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open the cabinet door'." + }, + { + "from": "gpt", + "value": "From <16.1> to <21.3>." + } + ] + }, + { + "video_id": "FCQS8", + "question_id": "FCQS8", + "video_file": "Charades/videos/FCQS8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks through a doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "VT624", + "question_id": "VT624", + "video_file": "Charades/videos/VT624.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is watching television in their bedroom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <12.9>." + } + ] + }, + { + "video_id": "PNVRJ", + "question_id": "PNVRJ", + "video_file": "Charades/videos/PNVRJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person still sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.9>." + } + ] + }, + { + "video_id": "PWJ3J", + "question_id": "PWJ3J", + "video_file": "Charades/videos/PWJ3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begin sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <13.3>." + } + ] + }, + { + "video_id": "PWJ3J", + "question_id": "PWJ3J", + "video_file": "Charades/videos/PWJ3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person start putting away groceries' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.2> to <29.1>." + } + ] + }, + { + "video_id": "1WJ6O", + "question_id": "1WJ6O", + "video_file": "Charades/videos/1WJ6O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits down on a couch' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <27.5>." + } + ] + }, + { + "video_id": "1WJ6O", + "question_id": "1WJ6O", + "video_file": "Charades/videos/1WJ6O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.8> to <29.4>." + } + ] + }, + { + "video_id": "1WJ6O", + "question_id": "1WJ6O", + "video_file": "Charades/videos/1WJ6O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is opening a door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "1WJ6O", + "question_id": "1WJ6O", + "video_file": "Charades/videos/1WJ6O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person watching television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.6> to <17.8>." + } + ] + }, + { + "video_id": "1WJ6O", + "question_id": "1WJ6O", + "video_file": "Charades/videos/1WJ6O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens the front door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "1WJ6O", + "question_id": "1WJ6O", + "video_file": "Charades/videos/1WJ6O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walking through the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.8> to <31.9>." + } + ] + }, + { + "video_id": "YOVS5", + "question_id": "YOVS5", + "video_file": "Charades/videos/YOVS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person the stand up'." + }, + { + "from": "gpt", + "value": "From <5.6> to <13.7>." + } + ] + }, + { + "video_id": "YOVS5", + "question_id": "YOVS5", + "video_file": "Charades/videos/YOVS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person close the door'." + }, + { + "from": "gpt", + "value": "From <15.8> to <22.9>." + } + ] + }, + { + "video_id": "YOVS5", + "question_id": "YOVS5", + "video_file": "Charades/videos/YOVS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person leave through a doorway closing the door behind them' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.8> to <22.9>." + } + ] + }, + { + "video_id": "YOVS5", + "question_id": "YOVS5", + "video_file": "Charades/videos/YOVS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person awakens on a floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <8.5>." + } + ] + }, + { + "video_id": "63TU2", + "question_id": "63TU2", + "video_file": "Charades/videos/63TU2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walk through a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.0> to <36.0>." + } + ] + }, + { + "video_id": "FZBK6", + "question_id": "FZBK6", + "video_file": "Charades/videos/FZBK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person start watching television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.5> to <21.0>." + } + ] + }, + { + "video_id": "FZBK6", + "question_id": "FZBK6", + "video_file": "Charades/videos/FZBK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put on a blanket' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.1> to <21.0>." + } + ] + }, + { + "video_id": "RS4ZC", + "question_id": "RS4ZC", + "video_file": "Charades/videos/RS4ZC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a picture of it'." + }, + { + "from": "gpt", + "value": "From <10.4> to <20.1>." + } + ] + }, + { + "video_id": "RS4ZC", + "question_id": "RS4ZC", + "video_file": "Charades/videos/RS4ZC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a cellphone picture with it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.4> to <20.1>." + } + ] + }, + { + "video_id": "RS4ZC", + "question_id": "RS4ZC", + "video_file": "Charades/videos/RS4ZC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person runs down the hall' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.9> to <26.2>." + } + ] + }, + { + "video_id": "RS4ZC", + "question_id": "RS4ZC", + "video_file": "Charades/videos/RS4ZC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is eating food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "RS4ZC", + "question_id": "RS4ZC", + "video_file": "Charades/videos/RS4ZC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person eats a snack'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "ENC45", + "question_id": "ENC45", + "video_file": "Charades/videos/ENC45.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.3>." + } + ] + }, + { + "video_id": "61UIO", + "question_id": "61UIO", + "video_file": "Charades/videos/61UIO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.8> to <27.8>." + } + ] + }, + { + "video_id": "Y5826", + "question_id": "Y5826", + "video_file": "Charades/videos/Y5826.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is cooking food on a stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "Y5826", + "question_id": "Y5826", + "video_file": "Charades/videos/Y5826.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking at a stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "GRBUE", + "question_id": "GRBUE", + "video_file": "Charades/videos/GRBUE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <14.1>." + } + ] + }, + { + "video_id": "GRBUE", + "question_id": "GRBUE", + "video_file": "Charades/videos/GRBUE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the first person turns off the light'." + }, + { + "from": "gpt", + "value": "From <14.0> to <18.9>." + } + ] + }, + { + "video_id": "GRBUE", + "question_id": "GRBUE", + "video_file": "Charades/videos/GRBUE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person begins undressing from their thick jacket'." + }, + { + "from": "gpt", + "value": "From <4.1> to <9.3>." + } + ] + }, + { + "video_id": "GRBUE", + "question_id": "GRBUE", + "video_file": "Charades/videos/GRBUE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throwing it on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.5> to <10.3>." + } + ] + }, + { + "video_id": "GRBUE", + "question_id": "GRBUE", + "video_file": "Charades/videos/GRBUE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns off the light'." + }, + { + "from": "gpt", + "value": "From <14.0> to <18.9>." + } + ] + }, + { + "video_id": "QTOJX", + "question_id": "QTOJX", + "video_file": "Charades/videos/QTOJX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take out their phone to check a message' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.5> to <31.7>." + } + ] + }, + { + "video_id": "QTOJX", + "question_id": "QTOJX", + "video_file": "Charades/videos/QTOJX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person laughs'." + }, + { + "from": "gpt", + "value": "From <2.2> to <15.0>." + } + ] + }, + { + "video_id": "Y4PAD", + "question_id": "Y4PAD", + "video_file": "Charades/videos/Y4PAD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person walks behind them holding a bottle of medicine' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.6> to <22.9>." + } + ] + }, + { + "video_id": "GGAN0", + "question_id": "GGAN0", + "video_file": "Charades/videos/GGAN0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts it on the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.4> to <24.0>." + } + ] + }, + { + "video_id": "VPBIO", + "question_id": "VPBIO", + "video_file": "Charades/videos/VPBIO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they start undressing'." + }, + { + "from": "gpt", + "value": "From <16.1> to <23.9>." + } + ] + }, + { + "video_id": "VPBIO", + "question_id": "VPBIO", + "video_file": "Charades/videos/VPBIO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is putting groceries on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.7> to <9.6>." + } + ] + }, + { + "video_id": "VPBIO", + "question_id": "VPBIO", + "video_file": "Charades/videos/VPBIO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they begin to undress' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.1> to <23.9>." + } + ] + }, + { + "video_id": "VPBIO", + "question_id": "VPBIO", + "video_file": "Charades/videos/VPBIO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is holding a bag of groceries' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.5> to <5.5>." + } + ] + }, + { + "video_id": "61N26", + "question_id": "61N26", + "video_file": "Charades/videos/61N26.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they put a broom away' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.9> to <20.9>." + } + ] + }, + { + "video_id": "2MGC1", + "question_id": "2MGC1", + "video_file": "Charades/videos/2MGC1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pours something into a cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.2> to <36.0>." + } + ] + }, + { + "video_id": "2MGC1", + "question_id": "2MGC1", + "video_file": "Charades/videos/2MGC1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person holding a bag opens their door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <9.0>." + } + ] + }, + { + "video_id": "CVTII", + "question_id": "CVTII", + "video_file": "Charades/videos/CVTII.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take a medicine bottle from a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.2> to <33.0>." + } + ] + }, + { + "video_id": "CVTII", + "question_id": "CVTII", + "video_file": "Charades/videos/CVTII.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take medicine in a kitchen' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.2> to <33.0>." + } + ] + }, + { + "video_id": "NDD35", + "question_id": "NDD35", + "video_file": "Charades/videos/NDD35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs down a hallway holding a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.9>." + } + ] + }, + { + "video_id": "FBIVN", + "question_id": "FBIVN", + "video_file": "Charades/videos/FBIVN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.6> to <10.3>." + } + ] + }, + { + "video_id": "FBIVN", + "question_id": "FBIVN", + "video_file": "Charades/videos/FBIVN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes door'." + }, + { + "from": "gpt", + "value": "From <1.6> to <10.3>." + } + ] + }, + { + "video_id": "FBIVN", + "question_id": "FBIVN", + "video_file": "Charades/videos/FBIVN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is closing a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <10.3>." + } + ] + }, + { + "video_id": "4U52R", + "question_id": "4U52R", + "video_file": "Charades/videos/4U52R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sit on the bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <30.0>." + } + ] + }, + { + "video_id": "4U52R", + "question_id": "4U52R", + "video_file": "Charades/videos/4U52R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put a box on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.5> to <15.0>." + } + ] + }, + { + "video_id": "DDWK5", + "question_id": "DDWK5", + "video_file": "Charades/videos/DDWK5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opened the cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "DDWK5", + "question_id": "DDWK5", + "video_file": "Charades/videos/DDWK5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closed the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <9.1>." + } + ] + }, + { + "video_id": "DDWK5", + "question_id": "DDWK5", + "video_file": "Charades/videos/DDWK5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opend the cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "M23X2", + "question_id": "M23X2", + "video_file": "Charades/videos/M23X2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the bag into a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <16.4>." + } + ] + }, + { + "video_id": "M23X2", + "question_id": "M23X2", + "video_file": "Charades/videos/M23X2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs into the garage' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <7.4>." + } + ] + }, + { + "video_id": "M23X2", + "question_id": "M23X2", + "video_file": "Charades/videos/M23X2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person smiling at the camera' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <23.0>." + } + ] + }, + { + "video_id": "WF8WM", + "question_id": "WF8WM", + "video_file": "Charades/videos/WF8WM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting their shoes on'." + }, + { + "from": "gpt", + "value": "From <0.4> to <7.1>." + } + ] + }, + { + "video_id": "WF8WM", + "question_id": "WF8WM", + "video_file": "Charades/videos/WF8WM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person putting on their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <7.1>." + } + ] + }, + { + "video_id": "WF8WM", + "question_id": "WF8WM", + "video_file": "Charades/videos/WF8WM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is putting on shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <7.1>." + } + ] + }, + { + "video_id": "WF8WM", + "question_id": "WF8WM", + "video_file": "Charades/videos/WF8WM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins to laugh' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "WF8WM", + "question_id": "WF8WM", + "video_file": "Charades/videos/WF8WM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person is looking in the mirror laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "IO712", + "question_id": "IO712", + "video_file": "Charades/videos/IO712.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts to read the book'." + }, + { + "from": "gpt", + "value": "From <6.7> to <31.5>." + } + ] + }, + { + "video_id": "IO712", + "question_id": "IO712", + "video_file": "Charades/videos/IO712.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person closes a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "IO712", + "question_id": "IO712", + "video_file": "Charades/videos/IO712.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits down at a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <29.7>." + } + ] + }, + { + "video_id": "H51V5", + "question_id": "H51V5", + "video_file": "Charades/videos/H51V5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put the food on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.6> to <23.5>." + } + ] + }, + { + "video_id": "H51V5", + "question_id": "H51V5", + "video_file": "Charades/videos/H51V5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person laughing they put it on a plate'." + }, + { + "from": "gpt", + "value": "From <6.4> to <15.0>." + } + ] + }, + { + "video_id": "H51V5", + "question_id": "H51V5", + "video_file": "Charades/videos/H51V5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they laugh'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <15.0>." + } + ] + }, + { + "video_id": "ITJZB", + "question_id": "ITJZB", + "video_file": "Charades/videos/ITJZB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is in a recreational room undressing'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.8>." + } + ] + }, + { + "video_id": "ITJZB", + "question_id": "ITJZB", + "video_file": "Charades/videos/ITJZB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they sit down on a sofa'." + }, + { + "from": "gpt", + "value": "From <16.0> to <26.0>." + } + ] + }, + { + "video_id": "ITJZB", + "question_id": "ITJZB", + "video_file": "Charades/videos/ITJZB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting their clothes in a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <15.3>." + } + ] + }, + { + "video_id": "FJXBX", + "question_id": "FJXBX", + "video_file": "Charades/videos/FJXBX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person closes the book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <13.0>." + } + ] + }, + { + "video_id": "R1X8P", + "question_id": "R1X8P", + "video_file": "Charades/videos/R1X8P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person is opening cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "R1X8P", + "question_id": "R1X8P", + "video_file": "Charades/videos/R1X8P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts smiling' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.9> to <22.0>." + } + ] + }, + { + "video_id": "HOZ76", + "question_id": "HOZ76", + "video_file": "Charades/videos/HOZ76.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.0> to <25.0>." + } + ] + }, + { + "video_id": "HD38O", + "question_id": "HD38O", + "video_file": "Charades/videos/HD38O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.4> to <16.1>." + } + ] + }, + { + "video_id": "HD38O", + "question_id": "HD38O", + "video_file": "Charades/videos/HD38O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a bite of a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.4> to <14.2>." + } + ] + }, + { + "video_id": "HD38O", + "question_id": "HD38O", + "video_file": "Charades/videos/HD38O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a smiling person playing a game takes medicine' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.0> to <32.0>." + } + ] + }, + { + "video_id": "NF9ZR", + "question_id": "NF9ZR", + "video_file": "Charades/videos/NF9ZR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person finally opens the door'." + }, + { + "from": "gpt", + "value": "From <5.6> to <13.6>." + } + ] + }, + { + "video_id": "NF9ZR", + "question_id": "NF9ZR", + "video_file": "Charades/videos/NF9ZR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person is trying to open a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <13.6>." + } + ] + }, + { + "video_id": "NF9ZR", + "question_id": "NF9ZR", + "video_file": "Charades/videos/NF9ZR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person wrapped in a blanket starts dressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.7> to <32.0>." + } + ] + }, + { + "video_id": "NF9ZR", + "question_id": "NF9ZR", + "video_file": "Charades/videos/NF9ZR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they open the door'." + }, + { + "from": "gpt", + "value": "From <5.6> to <13.6>." + } + ] + }, + { + "video_id": "CO1S9", + "question_id": "CO1S9", + "video_file": "Charades/videos/CO1S9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs out' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.7> to <12.7>." + } + ] + }, + { + "video_id": "SPUPH", + "question_id": "SPUPH", + "video_file": "Charades/videos/SPUPH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begin sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <9.7>." + } + ] + }, + { + "video_id": "SPUPH", + "question_id": "SPUPH", + "video_file": "Charades/videos/SPUPH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eat a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.6> to <13.0>." + } + ] + }, + { + "video_id": "0M2DO", + "question_id": "0M2DO", + "video_file": "Charades/videos/0M2DO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a cabinet'." + }, + { + "from": "gpt", + "value": "From <4.9> to <11.9>." + } + ] + }, + { + "video_id": "0M2DO", + "question_id": "0M2DO", + "video_file": "Charades/videos/0M2DO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws the clothes on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.8> to <25.7>." + } + ] + }, + { + "video_id": "0M2DO", + "question_id": "0M2DO", + "video_file": "Charades/videos/0M2DO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person in a bathroom opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <11.9>." + } + ] + }, + { + "video_id": "0M2DO", + "question_id": "0M2DO", + "video_file": "Charades/videos/0M2DO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes out clothes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <14.8>." + } + ] + }, + { + "video_id": "0M2DO", + "question_id": "0M2DO", + "video_file": "Charades/videos/0M2DO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walks through a doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <27.1>." + } + ] + }, + { + "video_id": "NVGDG", + "question_id": "NVGDG", + "video_file": "Charades/videos/NVGDG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person covered by a blanket awakens'." + }, + { + "from": "gpt", + "value": "From <1.7> to <15.2>." + } + ] + }, + { + "video_id": "NVGDG", + "question_id": "NVGDG", + "video_file": "Charades/videos/NVGDG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person covered in a blanket awakens' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <15.2>." + } + ] + }, + { + "video_id": "7G6F0", + "question_id": "7G6F0", + "video_file": "Charades/videos/7G6F0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking something from a cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.4> to <27.2>." + } + ] + }, + { + "video_id": "73NG4", + "question_id": "73NG4", + "video_file": "Charades/videos/73NG4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours a glass of juice'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <13.1>." + } + ] + }, + { + "video_id": "73NG4", + "question_id": "73NG4", + "video_file": "Charades/videos/73NG4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits at the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.9> to <33.0>." + } + ] + }, + { + "video_id": "UL65N", + "question_id": "UL65N", + "video_file": "Charades/videos/UL65N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put it on the bedside table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.2> to <17.8>." + } + ] + }, + { + "video_id": "12VVC", + "question_id": "12VVC", + "video_file": "Charades/videos/12VVC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes a book off a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.6> to <14.9>." + } + ] + }, + { + "video_id": "FQ9QX", + "question_id": "FQ9QX", + "video_file": "Charades/videos/FQ9QX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts sneezing'." + }, + { + "from": "gpt", + "value": "From <20.0> to <26.6>." + } + ] + }, + { + "video_id": "SHMU7", + "question_id": "SHMU7", + "video_file": "Charades/videos/SHMU7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening a bag of groceries' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <17.6>." + } + ] + }, + { + "video_id": "SHMU7", + "question_id": "SHMU7", + "video_file": "Charades/videos/SHMU7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts some bags onto a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <7.0>." + } + ] + }, + { + "video_id": "SHMU7", + "question_id": "SHMU7", + "video_file": "Charades/videos/SHMU7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they take some food out of the bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.0> to <15.5>." + } + ] + }, + { + "video_id": "SHP4R", + "question_id": "SHP4R", + "video_file": "Charades/videos/SHP4R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens the refrigerator'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "SHP4R", + "question_id": "SHP4R", + "video_file": "Charades/videos/SHP4R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes out some food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.8> to <10.7>." + } + ] + }, + { + "video_id": "GJNI4", + "question_id": "GJNI4", + "video_file": "Charades/videos/GJNI4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinks from a bottle of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.1> to <28.2>." + } + ] + }, + { + "video_id": "GJNI4", + "question_id": "GJNI4", + "video_file": "Charades/videos/GJNI4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drink from a bottle of soda' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.1> to <28.2>." + } + ] + }, + { + "video_id": "LNR61", + "question_id": "LNR61", + "video_file": "Charades/videos/LNR61.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding a towel' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.7> to <19.0>." + } + ] + }, + { + "video_id": "FDPHZ", + "question_id": "FDPHZ", + "video_file": "Charades/videos/FDPHZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns off a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.9> to <21.0>." + } + ] + }, + { + "video_id": "FDPHZ", + "question_id": "FDPHZ", + "video_file": "Charades/videos/FDPHZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.9> to <21.0>." + } + ] + }, + { + "video_id": "W2DAK", + "question_id": "W2DAK", + "video_file": "Charades/videos/W2DAK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes out a dryer sheet from a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.9> to <19.6>." + } + ] + }, + { + "video_id": "W2DAK", + "question_id": "W2DAK", + "video_file": "Charades/videos/W2DAK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes a dryer sheet out of a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.9> to <19.6>." + } + ] + }, + { + "video_id": "LCZ8F", + "question_id": "LCZ8F", + "video_file": "Charades/videos/LCZ8F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person slid open a door of a closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "KR0SH", + "question_id": "KR0SH", + "video_file": "Charades/videos/KR0SH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a book on the desk'." + }, + { + "from": "gpt", + "value": "From <10.5> to <16.2>." + } + ] + }, + { + "video_id": "ZP8AL", + "question_id": "ZP8AL", + "video_file": "Charades/videos/ZP8AL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person throws the pillow back out through the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.7> to <16.4>." + } + ] + }, + { + "video_id": "ZP8AL", + "question_id": "ZP8AL", + "video_file": "Charades/videos/ZP8AL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs into the room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <19.0>." + } + ] + }, + { + "video_id": "ZP8AL", + "question_id": "ZP8AL", + "video_file": "Charades/videos/ZP8AL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throwing the pillow onto the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.7> to <16.4>." + } + ] + }, + { + "video_id": "ZP8AL", + "question_id": "ZP8AL", + "video_file": "Charades/videos/ZP8AL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs through a doorway holding a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.6> to <16.1>." + } + ] + }, + { + "video_id": "ZP8AL", + "question_id": "ZP8AL", + "video_file": "Charades/videos/ZP8AL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.6> to <16.1>." + } + ] + }, + { + "video_id": "OHLSE", + "question_id": "OHLSE", + "video_file": "Charades/videos/OHLSE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the box back' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <28.1> to <32.0>." + } + ] + }, + { + "video_id": "OHLSE", + "question_id": "OHLSE", + "video_file": "Charades/videos/OHLSE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a box off the shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.0> to <12.9>." + } + ] + }, + { + "video_id": "OHLSE", + "question_id": "OHLSE", + "video_file": "Charades/videos/OHLSE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person he eats a snack'." + }, + { + "from": "gpt", + "value": "From <17.0> to <22.2>." + } + ] + }, + { + "video_id": "OHLSE", + "question_id": "OHLSE", + "video_file": "Charades/videos/OHLSE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person he opens the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "OHLSE", + "question_id": "OHLSE", + "video_file": "Charades/videos/OHLSE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens the closet'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "OHLSE", + "question_id": "OHLSE", + "video_file": "Charades/videos/OHLSE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person begins eating the sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.6> to <22.0>." + } + ] + }, + { + "video_id": "OHLSE", + "question_id": "OHLSE", + "video_file": "Charades/videos/OHLSE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person he puts the box back on the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.1> to <31.7>." + } + ] + }, + { + "video_id": "7NW4V", + "question_id": "7NW4V", + "video_file": "Charades/videos/7NW4V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.8> to <26.9>." + } + ] + }, + { + "video_id": "7NW4V", + "question_id": "7NW4V", + "video_file": "Charades/videos/7NW4V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sit on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.8> to <26.9>." + } + ] + }, + { + "video_id": "8NJD2", + "question_id": "8NJD2", + "video_file": "Charades/videos/8NJD2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.3> to <26.4>." + } + ] + }, + { + "video_id": "8NJD2", + "question_id": "8NJD2", + "video_file": "Charades/videos/8NJD2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a large cabinet to get a broom'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <7.0>." + } + ] + }, + { + "video_id": "8NJD2", + "question_id": "8NJD2", + "video_file": "Charades/videos/8NJD2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <7.0>." + } + ] + }, + { + "video_id": "UNQD7", + "question_id": "UNQD7", + "video_file": "Charades/videos/UNQD7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person is holding a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.5> to <33.0>." + } + ] + }, + { + "video_id": "UNQD7", + "question_id": "UNQD7", + "video_file": "Charades/videos/UNQD7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person closes a wardrobe door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.3> to <7.6>." + } + ] + }, + { + "video_id": "HONL7", + "question_id": "HONL7", + "video_file": "Charades/videos/HONL7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person lying near the doorway throws down a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.1>." + } + ] + }, + { + "video_id": "HONL7", + "question_id": "HONL7", + "video_file": "Charades/videos/HONL7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throwing it on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.8>." + } + ] + }, + { + "video_id": "F1MWB", + "question_id": "F1MWB", + "video_file": "Charades/videos/F1MWB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person is walking to box to put groceries in there'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.7> to <16.1>." + } + ] + }, + { + "video_id": "F1MWB", + "question_id": "F1MWB", + "video_file": "Charades/videos/F1MWB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person go over to turn off light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <33.9> to <38.0>." + } + ] + }, + { + "video_id": "F1MWB", + "question_id": "F1MWB", + "video_file": "Charades/videos/F1MWB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <24.0>." + } + ] + }, + { + "video_id": "7NBPG", + "question_id": "7NBPG", + "video_file": "Charades/videos/7NBPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person looks at the picture'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.7> to <31.0>." + } + ] + }, + { + "video_id": "7NBPG", + "question_id": "7NBPG", + "video_file": "Charades/videos/7NBPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person throws the book onto the counter' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.2> to <25.1>." + } + ] + }, + { + "video_id": "7NBPG", + "question_id": "7NBPG", + "video_file": "Charades/videos/7NBPG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person picking up picture looking at it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.7> to <31.0>." + } + ] + }, + { + "video_id": "I6PBQ", + "question_id": "I6PBQ", + "video_file": "Charades/videos/I6PBQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sneezing into a sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "82R7V", + "question_id": "82R7V", + "video_file": "Charades/videos/82R7V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a cup of tea'." + }, + { + "from": "gpt", + "value": "From <2.8> to <8.2>." + } + ] + }, + { + "video_id": "82R7V", + "question_id": "82R7V", + "video_file": "Charades/videos/82R7V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a woman is smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "09F15", + "question_id": "09F15", + "video_file": "Charades/videos/09F15.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens the pantry door'." + }, + { + "from": "gpt", + "value": "From <0.4> to <6.5>." + } + ] + }, + { + "video_id": "09F15", + "question_id": "09F15", + "video_file": "Charades/videos/09F15.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person holding a book'." + }, + { + "from": "gpt", + "value": "From <0.4> to <9.2>." + } + ] + }, + { + "video_id": "EF2YJ", + "question_id": "EF2YJ", + "video_file": "Charades/videos/EF2YJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is standing at the stove cooking'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <25.6>." + } + ] + }, + { + "video_id": "OXC4X", + "question_id": "OXC4X", + "video_file": "Charades/videos/OXC4X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person gets dressed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.1> to <12.9>." + } + ] + }, + { + "video_id": "OXC4X", + "question_id": "OXC4X", + "video_file": "Charades/videos/OXC4X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they get dressed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.1> to <12.9>." + } + ] + }, + { + "video_id": "OXC4X", + "question_id": "OXC4X", + "video_file": "Charades/videos/OXC4X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes some clothes from a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "CU9RE", + "question_id": "CU9RE", + "video_file": "Charades/videos/CU9RE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pick up a glass drink from the glass'." + }, + { + "from": "gpt", + "value": "From <24.9> to <31.0>." + } + ] + }, + { + "video_id": "TTZVW", + "question_id": "TTZVW", + "video_file": "Charades/videos/TTZVW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they are holding a plate of food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.7> to <16.3>." + } + ] + }, + { + "video_id": "TTZVW", + "question_id": "TTZVW", + "video_file": "Charades/videos/TTZVW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person hold a cup of coffee' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <23.1>." + } + ] + }, + { + "video_id": "TTZVW", + "question_id": "TTZVW", + "video_file": "Charades/videos/TTZVW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is opening a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "TTZVW", + "question_id": "TTZVW", + "video_file": "Charades/videos/TTZVW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking a cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.0> to <27.2>." + } + ] + }, + { + "video_id": "TTZVW", + "question_id": "TTZVW", + "video_file": "Charades/videos/TTZVW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "537I0", + "question_id": "537I0", + "video_file": "Charades/videos/537I0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <17.8>." + } + ] + }, + { + "video_id": "537I0", + "question_id": "537I0", + "video_file": "Charades/videos/537I0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts a broom in the closet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <11.9>." + } + ] + }, + { + "video_id": "0CX32", + "question_id": "0CX32", + "video_file": "Charades/videos/0CX32.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs into the living room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "0CX32", + "question_id": "0CX32", + "video_file": "Charades/videos/0CX32.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <22.8>." + } + ] + }, + { + "video_id": "0CX32", + "question_id": "0CX32", + "video_file": "Charades/videos/0CX32.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs into the room carrying a broom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "0CX32", + "question_id": "0CX32", + "video_file": "Charades/videos/0CX32.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the towel'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.3> to <9.1>." + } + ] + }, + { + "video_id": "HUNW6", + "question_id": "HUNW6", + "video_file": "Charades/videos/HUNW6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <18.7>." + } + ] + }, + { + "video_id": "HUNW6", + "question_id": "HUNW6", + "video_file": "Charades/videos/HUNW6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating food they got off a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <18.7>." + } + ] + }, + { + "video_id": "DED50", + "question_id": "DED50", + "video_file": "Charades/videos/DED50.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person awakens on a sofa in their home office'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.6> to <9.0>." + } + ] + }, + { + "video_id": "DED50", + "question_id": "DED50", + "video_file": "Charades/videos/DED50.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person pours a cup of coffee' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.2> to <21.9>." + } + ] + }, + { + "video_id": "H9U38", + "question_id": "H9U38", + "video_file": "Charades/videos/H9U38.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person closes the laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.1> to <26.0>." + } + ] + }, + { + "video_id": "8AKRP", + "question_id": "8AKRP", + "video_file": "Charades/videos/8AKRP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the cup into the sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <18.8>." + } + ] + }, + { + "video_id": "8AKRP", + "question_id": "8AKRP", + "video_file": "Charades/videos/8AKRP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.7> to <15.9>." + } + ] + }, + { + "video_id": "8AKRP", + "question_id": "8AKRP", + "video_file": "Charades/videos/8AKRP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put the cup into the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.6> to <18.8>." + } + ] + }, + { + "video_id": "4MKH4", + "question_id": "4MKH4", + "video_file": "Charades/videos/4MKH4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holds a tray of dishes in the hallway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.9> to <14.5>." + } + ] + }, + { + "video_id": "LNEAG", + "question_id": "LNEAG", + "video_file": "Charades/videos/LNEAG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the picture on top of a pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <13.9>." + } + ] + }, + { + "video_id": "LNEAG", + "question_id": "LNEAG", + "video_file": "Charades/videos/LNEAG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts picture on top of pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <13.9>." + } + ] + }, + { + "video_id": "5JVMD", + "question_id": "5JVMD", + "video_file": "Charades/videos/5JVMD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pouring a drink into a cup the consuming it'." + }, + { + "from": "gpt", + "value": "From <18.0> to <24.0>." + } + ] + }, + { + "video_id": "5JVMD", + "question_id": "5JVMD", + "video_file": "Charades/videos/5JVMD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pours a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.2> to <22.6>." + } + ] + }, + { + "video_id": "5JVMD", + "question_id": "5JVMD", + "video_file": "Charades/videos/5JVMD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person puts clothes next to a vacuum' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.0>." + } + ] + }, + { + "video_id": "8FH96", + "question_id": "8FH96", + "video_file": "Charades/videos/8FH96.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open the refrigerator door'." + }, + { + "from": "gpt", + "value": "From <8.6> to <14.5>." + } + ] + }, + { + "video_id": "8FH96", + "question_id": "8FH96", + "video_file": "Charades/videos/8FH96.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person laugh at what's inside' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <18.1>." + } + ] + }, + { + "video_id": "8FH96", + "question_id": "8FH96", + "video_file": "Charades/videos/8FH96.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person she opens the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.6> to <14.5>." + } + ] + }, + { + "video_id": "SP8UG", + "question_id": "SP8UG", + "video_file": "Charades/videos/SP8UG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take their medicine'." + }, + { + "from": "gpt", + "value": "From <29.0> to <34.0>." + } + ] + }, + { + "video_id": "SP8UG", + "question_id": "SP8UG", + "video_file": "Charades/videos/SP8UG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes medicine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.0> to <34.0>." + } + ] + }, + { + "video_id": "DTJ6B", + "question_id": "DTJ6B", + "video_file": "Charades/videos/DTJ6B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks through the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "WDCGH", + "question_id": "WDCGH", + "video_file": "Charades/videos/WDCGH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person reading a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "B8IYJ", + "question_id": "B8IYJ", + "video_file": "Charades/videos/B8IYJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin eating' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.5> to <19.0>." + } + ] + }, + { + "video_id": "G4WCE", + "question_id": "G4WCE", + "video_file": "Charades/videos/G4WCE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put the bag on a desk' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.3> to <15.5>." + } + ] + }, + { + "video_id": "D0V0Y", + "question_id": "D0V0Y", + "video_file": "Charades/videos/D0V0Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person this they sit in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.5> to <30.0>." + } + ] + }, + { + "video_id": "D0V0Y", + "question_id": "D0V0Y", + "video_file": "Charades/videos/D0V0Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks through a doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.2> to <12.0>." + } + ] + }, + { + "video_id": "D0V0Y", + "question_id": "D0V0Y", + "video_file": "Charades/videos/D0V0Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <20.5>." + } + ] + }, + { + "video_id": "D0V0Y", + "question_id": "D0V0Y", + "video_file": "Charades/videos/D0V0Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turn on a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.5> to <20.5>." + } + ] + }, + { + "video_id": "D0V0Y", + "question_id": "D0V0Y", + "video_file": "Charades/videos/D0V0Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns on a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.5> to <20.5>." + } + ] + }, + { + "video_id": "D0V0Y", + "question_id": "D0V0Y", + "video_file": "Charades/videos/D0V0Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes off a shoe' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.3> to <35.0>." + } + ] + }, + { + "video_id": "D0V0Y", + "question_id": "D0V0Y", + "video_file": "Charades/videos/D0V0Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts another shoe on' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.0> to <35.0>." + } + ] + }, + { + "video_id": "D0V0Y", + "question_id": "D0V0Y", + "video_file": "Charades/videos/D0V0Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begin to put on their shoes'." + }, + { + "from": "gpt", + "value": "From <26.0> to <35.0>." + } + ] + }, + { + "video_id": "XXMN1", + "question_id": "XXMN1", + "video_file": "Charades/videos/XXMN1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts laughing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.2> to <32.0>." + } + ] + }, + { + "video_id": "NDBXV", + "question_id": "NDBXV", + "video_file": "Charades/videos/NDBXV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take a camera from a shelf'." + }, + { + "from": "gpt", + "value": "From <17.7> to <23.7>." + } + ] + }, + { + "video_id": "48CYX", + "question_id": "48CYX", + "video_file": "Charades/videos/48CYX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take a book from atop a shelf'." + }, + { + "from": "gpt", + "value": "From <17.5> to <22.1>." + } + ] + }, + { + "video_id": "48CYX", + "question_id": "48CYX", + "video_file": "Charades/videos/48CYX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person undresses in the bathroom'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "WHJ70", + "question_id": "WHJ70", + "video_file": "Charades/videos/WHJ70.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person a towel opens a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <11.6>." + } + ] + }, + { + "video_id": "WHJ70", + "question_id": "WHJ70", + "video_file": "Charades/videos/WHJ70.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is opening a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.0> to <11.6>." + } + ] + }, + { + "video_id": "MHQYD", + "question_id": "MHQYD", + "video_file": "Charades/videos/MHQYD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looking out a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.2> to <19.1>." + } + ] + }, + { + "video_id": "MHQYD", + "question_id": "MHQYD", + "video_file": "Charades/videos/MHQYD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is smiling at a window'." + }, + { + "from": "gpt", + "value": "From <8.3> to <21.0>." + } + ] + }, + { + "video_id": "MHQYD", + "question_id": "MHQYD", + "video_file": "Charades/videos/MHQYD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person one is laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <21.0>." + } + ] + }, + { + "video_id": "LB006", + "question_id": "LB006", + "video_file": "Charades/videos/LB006.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person takes an item out the box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.4> to <16.0>." + } + ] + }, + { + "video_id": "CZY4Q", + "question_id": "CZY4Q", + "video_file": "Charades/videos/CZY4Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts their phone onto a shelf'." + }, + { + "from": "gpt", + "value": "From <6.8> to <14.9>." + } + ] + }, + { + "video_id": "CZY4Q", + "question_id": "CZY4Q", + "video_file": "Charades/videos/CZY4Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'that person puts the phone on a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.8> to <14.9>." + } + ] + }, + { + "video_id": "IA5TC", + "question_id": "IA5TC", + "video_file": "Charades/videos/IA5TC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person enters their house holding boxes'." + }, + { + "from": "gpt", + "value": "From <2.3> to <12.3>." + } + ] + }, + { + "video_id": "IA5TC", + "question_id": "IA5TC", + "video_file": "Charades/videos/IA5TC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.4> to <32.0>." + } + ] + }, + { + "video_id": "IA5TC", + "question_id": "IA5TC", + "video_file": "Charades/videos/IA5TC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is putting a box on the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.6> to <20.1>." + } + ] + }, + { + "video_id": "UBUKL", + "question_id": "UBUKL", + "video_file": "Charades/videos/UBUKL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits on couch' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.6> to <12.0>." + } + ] + }, + { + "video_id": "UBUKL", + "question_id": "UBUKL", + "video_file": "Charades/videos/UBUKL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits on a couch'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.6> to <12.0>." + } + ] + }, + { + "video_id": "UBUKL", + "question_id": "UBUKL", + "video_file": "Charades/videos/UBUKL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they take a coffee cup off a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <13.6>." + } + ] + }, + { + "video_id": "UBUKL", + "question_id": "UBUKL", + "video_file": "Charades/videos/UBUKL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put it on another table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <15.4>." + } + ] + }, + { + "video_id": "1YSGI", + "question_id": "1YSGI", + "video_file": "Charades/videos/1YSGI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the box on a dryer'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <25.5>." + } + ] + }, + { + "video_id": "6LNYW", + "question_id": "6LNYW", + "video_file": "Charades/videos/6LNYW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person b throws a pillow on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.1> to <26.0>." + } + ] + }, + { + "video_id": "4SN6Z", + "question_id": "4SN6Z", + "video_file": "Charades/videos/4SN6Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "4SN6Z", + "question_id": "4SN6Z", + "video_file": "Charades/videos/4SN6Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person close the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.6> to <21.0>." + } + ] + }, + { + "video_id": "4SN6Z", + "question_id": "4SN6Z", + "video_file": "Charades/videos/4SN6Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put dishes on shelves in cabinets'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <12.2>." + } + ] + }, + { + "video_id": "RZIAJ", + "question_id": "RZIAJ", + "video_file": "Charades/videos/RZIAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sitting on a bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "RZIAJ", + "question_id": "RZIAJ", + "video_file": "Charades/videos/RZIAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person close the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.5> to <31.0>." + } + ] + }, + { + "video_id": "RZIAJ", + "question_id": "RZIAJ", + "video_file": "Charades/videos/RZIAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closing the door'." + }, + { + "from": "gpt", + "value": "From <22.5> to <31.0>." + } + ] + }, + { + "video_id": "RZIAJ", + "question_id": "RZIAJ", + "video_file": "Charades/videos/RZIAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is playing with the phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.8> to <12.6>." + } + ] + }, + { + "video_id": "95HQC", + "question_id": "95HQC", + "video_file": "Charades/videos/95HQC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is undressing'." + }, + { + "from": "gpt", + "value": "From <2.3> to <11.3>." + } + ] + }, + { + "video_id": "8L9AJ", + "question_id": "8L9AJ", + "video_file": "Charades/videos/8L9AJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes out a laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.7> to <20.3>." + } + ] + }, + { + "video_id": "8L9AJ", + "question_id": "8L9AJ", + "video_file": "Charades/videos/8L9AJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a laptop out of a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.7> to <20.3>." + } + ] + }, + { + "video_id": "8L9AJ", + "question_id": "8L9AJ", + "video_file": "Charades/videos/8L9AJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person closes the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.6> to <31.2>." + } + ] + }, + { + "video_id": "NPEA3", + "question_id": "NPEA3", + "video_file": "Charades/videos/NPEA3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closed the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.1> to <29.9>." + } + ] + }, + { + "video_id": "NPEA3", + "question_id": "NPEA3", + "video_file": "Charades/videos/NPEA3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.1> to <29.9>." + } + ] + }, + { + "video_id": "UIPXF", + "question_id": "UIPXF", + "video_file": "Charades/videos/UIPXF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person walking through the doorway to the kitchen' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.2> to <29.6>." + } + ] + }, + { + "video_id": "UIPXF", + "question_id": "UIPXF", + "video_file": "Charades/videos/UIPXF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.5> to <31.8>." + } + ] + }, + { + "video_id": "UIPXF", + "question_id": "UIPXF", + "video_file": "Charades/videos/UIPXF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.5> to <31.8>." + } + ] + }, + { + "video_id": "HYD3N", + "question_id": "HYD3N", + "video_file": "Charades/videos/HYD3N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts a sandwich down on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.4> to <27.0>." + } + ] + }, + { + "video_id": "HYD3N", + "question_id": "HYD3N", + "video_file": "Charades/videos/HYD3N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting a sandwich in a towel' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.4> to <27.0>." + } + ] + }, + { + "video_id": "D8BTU", + "question_id": "D8BTU", + "video_file": "Charades/videos/D8BTU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a broom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.2> to <10.3>." + } + ] + }, + { + "video_id": "D8BTU", + "question_id": "D8BTU", + "video_file": "Charades/videos/D8BTU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens the cabinet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.4> to <12.5>." + } + ] + }, + { + "video_id": "D8BTU", + "question_id": "D8BTU", + "video_file": "Charades/videos/D8BTU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the broom up'." + }, + { + "from": "gpt", + "value": "From <6.2> to <10.9>." + } + ] + }, + { + "video_id": "D8BTU", + "question_id": "D8BTU", + "video_file": "Charades/videos/D8BTU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put the broom in the corner'." + }, + { + "from": "gpt", + "value": "From <6.2> to <10.9>." + } + ] + }, + { + "video_id": "D8BTU", + "question_id": "D8BTU", + "video_file": "Charades/videos/D8BTU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the cabinet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.4> to <12.5>." + } + ] + }, + { + "video_id": "ZGUTC", + "question_id": "ZGUTC", + "video_file": "Charades/videos/ZGUTC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opens the medicine cabinet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.2> to <25.6>." + } + ] + }, + { + "video_id": "UD7R8", + "question_id": "UD7R8", + "video_file": "Charades/videos/UD7R8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts on their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <12.7>." + } + ] + }, + { + "video_id": "UD7R8", + "question_id": "UD7R8", + "video_file": "Charades/videos/UD7R8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting on their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.6> to <12.7>." + } + ] + }, + { + "video_id": "X2LBW", + "question_id": "X2LBW", + "video_file": "Charades/videos/X2LBW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person b are dressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.8>." + } + ] + }, + { + "video_id": "X2LBW", + "question_id": "X2LBW", + "video_file": "Charades/videos/X2LBW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put the pillow on a chair'." + }, + { + "from": "gpt", + "value": "From <40.6> to <47.6>." + } + ] + }, + { + "video_id": "X2LBW", + "question_id": "X2LBW", + "video_file": "Charades/videos/X2LBW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person grab pillows to sit on at a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <49.1> to <59.9>." + } + ] + }, + { + "video_id": "I6UIG", + "question_id": "I6UIG", + "video_file": "Charades/videos/I6UIG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws some food'." + }, + { + "from": "gpt", + "value": "From <20.2> to <31.5>." + } + ] + }, + { + "video_id": "I6UIG", + "question_id": "I6UIG", + "video_file": "Charades/videos/I6UIG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person begins throwing food towards the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.2> to <31.5>." + } + ] + }, + { + "video_id": "0GH5O", + "question_id": "0GH5O", + "video_file": "Charades/videos/0GH5O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walk through the doorway smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "0GH5O", + "question_id": "0GH5O", + "video_file": "Charades/videos/0GH5O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person opens the front door looks outside'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <13.7>." + } + ] + }, + { + "video_id": "0IAGO", + "question_id": "0IAGO", + "video_file": "Charades/videos/0IAGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they begin trying to fix a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <21.0>." + } + ] + }, + { + "video_id": "0IAGO", + "question_id": "0IAGO", + "video_file": "Charades/videos/0IAGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns the light on' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <21.0>." + } + ] + }, + { + "video_id": "0IAGO", + "question_id": "0IAGO", + "video_file": "Charades/videos/0IAGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.4> to <11.4>." + } + ] + }, + { + "video_id": "0IAGO", + "question_id": "0IAGO", + "video_file": "Charades/videos/0IAGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins to fix a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <21.0>." + } + ] + }, + { + "video_id": "0IAGO", + "question_id": "0IAGO", + "video_file": "Charades/videos/0IAGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person fixes a light'." + }, + { + "from": "gpt", + "value": "From <13.0> to <21.0>." + } + ] + }, + { + "video_id": "0IAGO", + "question_id": "0IAGO", + "video_file": "Charades/videos/0IAGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sits down in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.0> to <11.7>." + } + ] + }, + { + "video_id": "0IAGO", + "question_id": "0IAGO", + "video_file": "Charades/videos/0IAGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person stands on the chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <15.0>." + } + ] + }, + { + "video_id": "0IAGO", + "question_id": "0IAGO", + "video_file": "Charades/videos/0IAGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes off their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <11.4>." + } + ] + }, + { + "video_id": "0IAGO", + "question_id": "0IAGO", + "video_file": "Charades/videos/0IAGO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person stands up on a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.0> to <15.0>." + } + ] + }, + { + "video_id": "ZHF60", + "question_id": "ZHF60", + "video_file": "Charades/videos/ZHF60.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begin washing clothes in the sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <43.2> to <53.9>." + } + ] + }, + { + "video_id": "ZHF60", + "question_id": "ZHF60", + "video_file": "Charades/videos/ZHF60.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open the cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.7> to <19.1>." + } + ] + }, + { + "video_id": "ZHF60", + "question_id": "ZHF60", + "video_file": "Charades/videos/ZHF60.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person uses a towel to wash dry their hands' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <33.9> to <62.6>." + } + ] + }, + { + "video_id": "ZHF60", + "question_id": "ZHF60", + "video_file": "Charades/videos/ZHF60.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is putting clothes in the sink'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.7> to <16.7>." + } + ] + }, + { + "video_id": "FWQYI", + "question_id": "FWQYI", + "video_file": "Charades/videos/FWQYI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person tidying up a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.0> to <40.0>." + } + ] + }, + { + "video_id": "Q9BRQ", + "question_id": "Q9BRQ", + "video_file": "Charades/videos/Q9BRQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <17.0>." + } + ] + }, + { + "video_id": "Q9BRQ", + "question_id": "Q9BRQ", + "video_file": "Charades/videos/Q9BRQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is laying on the floor holding a bag'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.4>." + } + ] + }, + { + "video_id": "J99Z8", + "question_id": "J99Z8", + "video_file": "Charades/videos/J99Z8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they begin smiling as they finish their food'." + }, + { + "from": "gpt", + "value": "From <0.4> to <9.2>." + } + ] + }, + { + "video_id": "VSYFA", + "question_id": "VSYFA", + "video_file": "Charades/videos/VSYFA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person eats the food out of the bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "VSYFA", + "question_id": "VSYFA", + "video_file": "Charades/videos/VSYFA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts eating some food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "VSYFA", + "question_id": "VSYFA", + "video_file": "Charades/videos/VSYFA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits down on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.8>." + } + ] + }, + { + "video_id": "VSYFA", + "question_id": "VSYFA", + "video_file": "Charades/videos/VSYFA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sitting down with something to eat' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "VSYFA", + "question_id": "VSYFA", + "video_file": "Charades/videos/VSYFA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walks out a doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <32.0>." + } + ] + }, + { + "video_id": "VSYFA", + "question_id": "VSYFA", + "video_file": "Charades/videos/VSYFA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person undresses their shirt' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.9>." + } + ] + }, + { + "video_id": "UJXLL", + "question_id": "UJXLL", + "video_file": "Charades/videos/UJXLL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes some clothes off a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.0> to <16.1>." + } + ] + }, + { + "video_id": "XXCS6", + "question_id": "XXCS6", + "video_file": "Charades/videos/XXCS6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a picture out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <32.0>." + } + ] + }, + { + "video_id": "XXCS6", + "question_id": "XXCS6", + "video_file": "Charades/videos/XXCS6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'takes a picture of a second person with a phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <32.0>." + } + ] + }, + { + "video_id": "XXCS6", + "question_id": "XXCS6", + "video_file": "Charades/videos/XXCS6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'takes the picture of another person posing outside' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <32.0>." + } + ] + }, + { + "video_id": "JAAVR", + "question_id": "JAAVR", + "video_file": "Charades/videos/JAAVR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take medicine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <26.6>." + } + ] + }, + { + "video_id": "JAAVR", + "question_id": "JAAVR", + "video_file": "Charades/videos/JAAVR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes the medicine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <26.6>." + } + ] + }, + { + "video_id": "E7Q3R", + "question_id": "E7Q3R", + "video_file": "Charades/videos/E7Q3R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts their coffee on the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <12.4>." + } + ] + }, + { + "video_id": "E7Q3R", + "question_id": "E7Q3R", + "video_file": "Charades/videos/E7Q3R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <18.7>." + } + ] + }, + { + "video_id": "E7Q3R", + "question_id": "E7Q3R", + "video_file": "Charades/videos/E7Q3R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs into their dining room'." + }, + { + "from": "gpt", + "value": "From <19.2> to <23.9>." + } + ] + }, + { + "video_id": "XJH0X", + "question_id": "XJH0X", + "video_file": "Charades/videos/XJH0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take a pillow off of the ground' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.7> to <26.1>." + } + ] + }, + { + "video_id": "XJH0X", + "question_id": "XJH0X", + "video_file": "Charades/videos/XJH0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person proceeds to wash dishes in the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.7> to <22.1>." + } + ] + }, + { + "video_id": "ZU8Q1", + "question_id": "ZU8Q1", + "video_file": "Charades/videos/ZU8Q1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens their laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <20.9>." + } + ] + }, + { + "video_id": "HL9YR", + "question_id": "HL9YR", + "video_file": "Charades/videos/HL9YR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person starts laughing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <38.9>." + } + ] + }, + { + "video_id": "HL9YR", + "question_id": "HL9YR", + "video_file": "Charades/videos/HL9YR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.2> to <42.0>." + } + ] + }, + { + "video_id": "HL9YR", + "question_id": "HL9YR", + "video_file": "Charades/videos/HL9YR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they close their laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.2> to <42.0>." + } + ] + }, + { + "video_id": "HL9YR", + "question_id": "HL9YR", + "video_file": "Charades/videos/HL9YR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <9.5>." + } + ] + }, + { + "video_id": "HL9YR", + "question_id": "HL9YR", + "video_file": "Charades/videos/HL9YR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begin smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.8> to <35.4>." + } + ] + }, + { + "video_id": "B6ZIW", + "question_id": "B6ZIW", + "video_file": "Charades/videos/B6ZIW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is throwing food up the stairs' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.6>." + } + ] + }, + { + "video_id": "B6ZIW", + "question_id": "B6ZIW", + "video_file": "Charades/videos/B6ZIW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person throws three bags on the stairs' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.1>." + } + ] + }, + { + "video_id": "B6ZIW", + "question_id": "B6ZIW", + "video_file": "Charades/videos/B6ZIW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person holding a bag of groceries opens it up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <8.9>." + } + ] + }, + { + "video_id": "B6ZIW", + "question_id": "B6ZIW", + "video_file": "Charades/videos/B6ZIW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws food onto the stairs' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.6>." + } + ] + }, + { + "video_id": "8ML3S", + "question_id": "8ML3S", + "video_file": "Charades/videos/8ML3S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is running around their closet'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "21E40", + "question_id": "21E40", + "video_file": "Charades/videos/21E40.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person drinks from a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.0> to <32.0>." + } + ] + }, + { + "video_id": "A3MOW", + "question_id": "A3MOW", + "video_file": "Charades/videos/A3MOW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person sat in a chair eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <16.9>." + } + ] + }, + { + "video_id": "A3MOW", + "question_id": "A3MOW", + "video_file": "Charades/videos/A3MOW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is sitting on a chair eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <16.9>." + } + ] + }, + { + "video_id": "A3MOW", + "question_id": "A3MOW", + "video_file": "Charades/videos/A3MOW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person 1 is eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <16.9>." + } + ] + }, + { + "video_id": "A3MOW", + "question_id": "A3MOW", + "video_file": "Charades/videos/A3MOW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looking at a framed picture on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <21.6>." + } + ] + }, + { + "video_id": "6E4L5", + "question_id": "6E4L5", + "video_file": "Charades/videos/6E4L5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes pictures' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.8> to <24.6>." + } + ] + }, + { + "video_id": "6E4L5", + "question_id": "6E4L5", + "video_file": "Charades/videos/6E4L5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a picture of what he is eating' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.4> to <15.8>." + } + ] + }, + { + "video_id": "6E4L5", + "question_id": "6E4L5", + "video_file": "Charades/videos/6E4L5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes out what appears to be a phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.8> to <16.8>." + } + ] + }, + { + "video_id": "6E4L5", + "question_id": "6E4L5", + "video_file": "Charades/videos/6E4L5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a picture of their half-eaten sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "6E4L5", + "question_id": "6E4L5", + "video_file": "Charades/videos/6E4L5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes out a camera'." + }, + { + "from": "gpt", + "value": "From <11.8> to <16.8>." + } + ] + }, + { + "video_id": "6E4L5", + "question_id": "6E4L5", + "video_file": "Charades/videos/6E4L5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "6E4L5", + "question_id": "6E4L5", + "video_file": "Charades/videos/6E4L5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person stands in a garage eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "6E4L5", + "question_id": "6E4L5", + "video_file": "Charades/videos/6E4L5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <15.8>." + } + ] + }, + { + "video_id": "IDDTF", + "question_id": "IDDTF", + "video_file": "Charades/videos/IDDTF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating something'." + }, + { + "from": "gpt", + "value": "From <3.9> to <11.1>." + } + ] + }, + { + "video_id": "BCUTT", + "question_id": "BCUTT", + "video_file": "Charades/videos/BCUTT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they stand up'." + }, + { + "from": "gpt", + "value": "From <26.5> to <32.0>." + } + ] + }, + { + "video_id": "0A8ZT", + "question_id": "0A8ZT", + "video_file": "Charades/videos/0A8ZT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins to play on their phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.3> to <34.0>." + } + ] + }, + { + "video_id": "0A8ZT", + "question_id": "0A8ZT", + "video_file": "Charades/videos/0A8ZT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they begin playing on their phone'." + }, + { + "from": "gpt", + "value": "From <26.3> to <34.0>." + } + ] + }, + { + "video_id": "0A8ZT", + "question_id": "0A8ZT", + "video_file": "Charades/videos/0A8ZT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes their phone from their pocket'." + }, + { + "from": "gpt", + "value": "From <26.5> to <34.0>." + } + ] + }, + { + "video_id": "3CW6P", + "question_id": "3CW6P", + "video_file": "Charades/videos/3CW6P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person grab the doorknob open the door'." + }, + { + "from": "gpt", + "value": "From <23.9> to <30.0>." + } + ] + }, + { + "video_id": "3CW6P", + "question_id": "3CW6P", + "video_file": "Charades/videos/3CW6P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.9> to <30.0>." + } + ] + }, + { + "video_id": "IZ978", + "question_id": "IZ978", + "video_file": "Charades/videos/IZ978.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take a picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.6> to <35.9>." + } + ] + }, + { + "video_id": "IZ978", + "question_id": "IZ978", + "video_file": "Charades/videos/IZ978.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting on a toilet undressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.5> to <28.6>." + } + ] + }, + { + "video_id": "IZ978", + "question_id": "IZ978", + "video_file": "Charades/videos/IZ978.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person undressing in the bathroom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.5> to <28.6>." + } + ] + }, + { + "video_id": "RHYAJ", + "question_id": "RHYAJ", + "video_file": "Charades/videos/RHYAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is snuggling a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.4> to <30.0>." + } + ] + }, + { + "video_id": "RHYAJ", + "question_id": "RHYAJ", + "video_file": "Charades/videos/RHYAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person snuggling up with a pillow'." + }, + { + "from": "gpt", + "value": "From <18.4> to <30.0>." + } + ] + }, + { + "video_id": "K8AUX", + "question_id": "K8AUX", + "video_file": "Charades/videos/K8AUX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a laughing person is drinking some coffee from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.6>." + } + ] + }, + { + "video_id": "K8AUX", + "question_id": "K8AUX", + "video_file": "Charades/videos/K8AUX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walk out the room through an open doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.4> to <31.0>." + } + ] + }, + { + "video_id": "K8AUX", + "question_id": "K8AUX", + "video_file": "Charades/videos/K8AUX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is drinking from a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.6>." + } + ] + }, + { + "video_id": "U7Z37", + "question_id": "U7Z37", + "video_file": "Charades/videos/U7Z37.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the door on the stove'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <9.5>." + } + ] + }, + { + "video_id": "KI0Q7", + "question_id": "KI0Q7", + "video_file": "Charades/videos/KI0Q7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person gets dressed than sits down'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "KI0Q7", + "question_id": "KI0Q7", + "video_file": "Charades/videos/KI0Q7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person dresses in the doorway'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "7QLMY", + "question_id": "7QLMY", + "video_file": "Charades/videos/7QLMY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a woman runs into a room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <9.3>." + } + ] + }, + { + "video_id": "7QLMY", + "question_id": "7QLMY", + "video_file": "Charades/videos/7QLMY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs into the room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <9.3>." + } + ] + }, + { + "video_id": "D260T", + "question_id": "D260T", + "video_file": "Charades/videos/D260T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person tidying up a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.6> to <12.5>." + } + ] + }, + { + "video_id": "XR54W", + "question_id": "XR54W", + "video_file": "Charades/videos/XR54W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person takes a pillow from under the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "1PALI", + "question_id": "1PALI", + "video_file": "Charades/videos/1PALI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they open a window'." + }, + { + "from": "gpt", + "value": "From <9.3> to <17.3>." + } + ] + }, + { + "video_id": "1PALI", + "question_id": "1PALI", + "video_file": "Charades/videos/1PALI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opens the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.3> to <17.3>." + } + ] + }, + { + "video_id": "G1CM7", + "question_id": "G1CM7", + "video_file": "Charades/videos/G1CM7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts reading a book'." + }, + { + "from": "gpt", + "value": "From <24.6> to <37.0>." + } + ] + }, + { + "video_id": "G1CM7", + "question_id": "G1CM7", + "video_file": "Charades/videos/G1CM7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person read a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.6> to <37.0>." + } + ] + }, + { + "video_id": "8WQOP", + "question_id": "8WQOP", + "video_file": "Charades/videos/8WQOP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a sneezing person opens a basement door holding a box'." + }, + { + "from": "gpt", + "value": "From <15.5> to <26.1>." + } + ] + }, + { + "video_id": "8WQOP", + "question_id": "8WQOP", + "video_file": "Charades/videos/8WQOP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.5> to <26.1>." + } + ] + }, + { + "video_id": "8WQOP", + "question_id": "8WQOP", + "video_file": "Charades/videos/8WQOP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is standing holding a mirror in one hand'." + }, + { + "from": "gpt", + "value": "From <0.1> to <5.8>." + } + ] + }, + { + "video_id": "MZAN6", + "question_id": "MZAN6", + "video_file": "Charades/videos/MZAN6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turn off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <20.9>." + } + ] + }, + { + "video_id": "MZAN6", + "question_id": "MZAN6", + "video_file": "Charades/videos/MZAN6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is standing in the bathroom fixing their hair'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "MZAN6", + "question_id": "MZAN6", + "video_file": "Charades/videos/MZAN6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person turns off the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.0> to <20.9>." + } + ] + }, + { + "video_id": "MZAN6", + "question_id": "MZAN6", + "video_file": "Charades/videos/MZAN6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is fixing their hair'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "MZAN6", + "question_id": "MZAN6", + "video_file": "Charades/videos/MZAN6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <16.5> to <23.3>." + } + ] + }, + { + "video_id": "YGQ7E", + "question_id": "YGQ7E", + "video_file": "Charades/videos/YGQ7E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person wards open the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.9> to <22.0>." + } + ] + }, + { + "video_id": "YGQ7E", + "question_id": "YGQ7E", + "video_file": "Charades/videos/YGQ7E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.9> to <22.0>." + } + ] + }, + { + "video_id": "24XDE", + "question_id": "24XDE", + "video_file": "Charades/videos/24XDE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts a book on a shelf'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "24XDE", + "question_id": "24XDE", + "video_file": "Charades/videos/24XDE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a man putting a book on a shelf'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "24XDE", + "question_id": "24XDE", + "video_file": "Charades/videos/24XDE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins dressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.2> to <20.5>." + } + ] + }, + { + "video_id": "24XDE", + "question_id": "24XDE", + "video_file": "Charades/videos/24XDE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person smiling the whole time'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "L37YS", + "question_id": "L37YS", + "video_file": "Charades/videos/L37YS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is opening the door to a closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.9>." + } + ] + }, + { + "video_id": "E23CT", + "question_id": "E23CT", + "video_file": "Charades/videos/E23CT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person turns off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.6> to <30.5>." + } + ] + }, + { + "video_id": "E23CT", + "question_id": "E23CT", + "video_file": "Charades/videos/E23CT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes their broom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "UAIHK", + "question_id": "UAIHK", + "video_file": "Charades/videos/UAIHK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.8> to <35.0>." + } + ] + }, + { + "video_id": "UAIHK", + "question_id": "UAIHK", + "video_file": "Charades/videos/UAIHK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is awakening' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "UAIHK", + "question_id": "UAIHK", + "video_file": "Charades/videos/UAIHK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits in bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.9> to <38.0>." + } + ] + }, + { + "video_id": "QSSCH", + "question_id": "QSSCH", + "video_file": "Charades/videos/QSSCH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes their laptop from the arm of the sofa' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.7> to <19.8>." + } + ] + }, + { + "video_id": "6YTAJ", + "question_id": "6YTAJ", + "video_file": "Charades/videos/6YTAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is eating some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.5> to <31.5>." + } + ] + }, + { + "video_id": "6YTAJ", + "question_id": "6YTAJ", + "video_file": "Charades/videos/6YTAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'this person is eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <22.5> to <31.5>." + } + ] + }, + { + "video_id": "6YTAJ", + "question_id": "6YTAJ", + "video_file": "Charades/videos/6YTAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pouring over the words intensely as they eat' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.5> to <31.5>." + } + ] + }, + { + "video_id": "6YTAJ", + "question_id": "6YTAJ", + "video_file": "Charades/videos/6YTAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting at a desk eating' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.5> to <31.5>." + } + ] + }, + { + "video_id": "F8SNW", + "question_id": "F8SNW", + "video_file": "Charades/videos/F8SNW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put the blanket on the counter'." + }, + { + "from": "gpt", + "value": "From <39.0> to <44.1>." + } + ] + }, + { + "video_id": "F8SNW", + "question_id": "F8SNW", + "video_file": "Charades/videos/F8SNW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sitting on a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <47.1> to <53.8>." + } + ] + }, + { + "video_id": "L3Z0M", + "question_id": "L3Z0M", + "video_file": "Charades/videos/L3Z0M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turn off the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.9> to <16.0>." + } + ] + }, + { + "video_id": "L3Z0M", + "question_id": "L3Z0M", + "video_file": "Charades/videos/L3Z0M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person turns off the bathroom light'." + }, + { + "from": "gpt", + "value": "From <11.9> to <16.0>." + } + ] + }, + { + "video_id": "N93LD", + "question_id": "N93LD", + "video_file": "Charades/videos/N93LD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put the laptop down'." + }, + { + "from": "gpt", + "value": "From <16.6> to <27.9>." + } + ] + }, + { + "video_id": "N93LD", + "question_id": "N93LD", + "video_file": "Charades/videos/N93LD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person closes their laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <20.9>." + } + ] + }, + { + "video_id": "GNUZK", + "question_id": "GNUZK", + "video_file": "Charades/videos/GNUZK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes some medicine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <17.6>." + } + ] + }, + { + "video_id": "GNUZK", + "question_id": "GNUZK", + "video_file": "Charades/videos/GNUZK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is fixing a light'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "A58I2", + "question_id": "A58I2", + "video_file": "Charades/videos/A58I2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person sits on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.9> to <33.0>." + } + ] + }, + { + "video_id": "A58I2", + "question_id": "A58I2", + "video_file": "Charades/videos/A58I2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <14.9>." + } + ] + }, + { + "video_id": "UL9LN", + "question_id": "UL9LN", + "video_file": "Charades/videos/UL9LN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding clothes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <27.5>." + } + ] + }, + { + "video_id": "UL9LN", + "question_id": "UL9LN", + "video_file": "Charades/videos/UL9LN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens door into wardrobe gets clothes out' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "UL9LN", + "question_id": "UL9LN", + "video_file": "Charades/videos/UL9LN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is opening a closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "UL9LN", + "question_id": "UL9LN", + "video_file": "Charades/videos/UL9LN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes the clothes over' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.9> to <28.6>." + } + ] + }, + { + "video_id": "VSO80", + "question_id": "VSO80", + "video_file": "Charades/videos/VSO80.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits down in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.6> to <32.0>." + } + ] + }, + { + "video_id": "VSO80", + "question_id": "VSO80", + "video_file": "Charades/videos/VSO80.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sitting down in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.6> to <32.0>." + } + ] + }, + { + "video_id": "VSO80", + "question_id": "VSO80", + "video_file": "Charades/videos/VSO80.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes some food out of the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.1> to <12.5>." + } + ] + }, + { + "video_id": "VSO80", + "question_id": "VSO80", + "video_file": "Charades/videos/VSO80.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a refrigerator door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.0> to <13.3>." + } + ] + }, + { + "video_id": "VSO80", + "question_id": "VSO80", + "video_file": "Charades/videos/VSO80.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person close the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <12.9>." + } + ] + }, + { + "video_id": "57NFY", + "question_id": "57NFY", + "video_file": "Charades/videos/57NFY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.1> to <16.1>." + } + ] + }, + { + "video_id": "57NFY", + "question_id": "57NFY", + "video_file": "Charades/videos/57NFY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sneezing into a mirror'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.9> to <8.6>." + } + ] + }, + { + "video_id": "57NFY", + "question_id": "57NFY", + "video_file": "Charades/videos/57NFY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening a cabinet in a kitchen' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.1> to <16.1>." + } + ] + }, + { + "video_id": "57NFY", + "question_id": "57NFY", + "video_file": "Charades/videos/57NFY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they close the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.0> to <34.4>." + } + ] + }, + { + "video_id": "JSY2I", + "question_id": "JSY2I", + "video_file": "Charades/videos/JSY2I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closed a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <12.7>." + } + ] + }, + { + "video_id": "JSY2I", + "question_id": "JSY2I", + "video_file": "Charades/videos/JSY2I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sneezing into a bag of groceries in the entryway'." + }, + { + "from": "gpt", + "value": "From <9.6> to <16.4>." + } + ] + }, + { + "video_id": "DQ46N", + "question_id": "DQ46N", + "video_file": "Charades/videos/DQ46N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person sits in a chair in the kitchen'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "DQ46N", + "question_id": "DQ46N", + "video_file": "Charades/videos/DQ46N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person was sitting in a chair in the kitchen'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "QEXP4", + "question_id": "QEXP4", + "video_file": "Charades/videos/QEXP4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts washing a mirror'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <9.6>." + } + ] + }, + { + "video_id": "TOOYI", + "question_id": "TOOYI", + "video_file": "Charades/videos/TOOYI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person places the box in the closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.2> to <19.1>." + } + ] + }, + { + "video_id": "EN3QY", + "question_id": "EN3QY", + "video_file": "Charades/videos/EN3QY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person stops to look out the window'." + }, + { + "from": "gpt", + "value": "From <20.3> to <32.2>." + } + ] + }, + { + "video_id": "EN3QY", + "question_id": "EN3QY", + "video_file": "Charades/videos/EN3QY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person looks out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.3> to <32.2>." + } + ] + }, + { + "video_id": "EUJK0", + "question_id": "EUJK0", + "video_file": "Charades/videos/EUJK0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person start reading a book'." + }, + { + "from": "gpt", + "value": "From <20.9> to <32.0>." + } + ] + }, + { + "video_id": "EUJK0", + "question_id": "EUJK0", + "video_file": "Charades/videos/EUJK0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person close the laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <22.1>." + } + ] + }, + { + "video_id": "76DE2", + "question_id": "76DE2", + "video_file": "Charades/videos/76DE2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person begins laughing'." + }, + { + "from": "gpt", + "value": "From <9.9> to <17.3>." + } + ] + }, + { + "video_id": "PROQY", + "question_id": "PROQY", + "video_file": "Charades/videos/PROQY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person runs into the entryway with a towel' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.0>." + } + ] + }, + { + "video_id": "M1B7N", + "question_id": "M1B7N", + "video_file": "Charades/videos/M1B7N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person cooking at stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <14.2>." + } + ] + }, + { + "video_id": "M1B7N", + "question_id": "M1B7N", + "video_file": "Charades/videos/M1B7N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person cooks something in a pot on the stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.3> to <14.2>." + } + ] + }, + { + "video_id": "M1B7N", + "question_id": "M1B7N", + "video_file": "Charades/videos/M1B7N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person take a sip of coffee from a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.5> to <27.9>." + } + ] + }, + { + "video_id": "NXUA9", + "question_id": "NXUA9", + "video_file": "Charades/videos/NXUA9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person washes their hands' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.8> to <26.0>." + } + ] + }, + { + "video_id": "NXUA9", + "question_id": "NXUA9", + "video_file": "Charades/videos/NXUA9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is fixing the light on the wall' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.6> to <9.7>." + } + ] + }, + { + "video_id": "62VEF", + "question_id": "62VEF", + "video_file": "Charades/videos/62VEF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking from a coffee cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <10.7>." + } + ] + }, + { + "video_id": "62VEF", + "question_id": "62VEF", + "video_file": "Charades/videos/62VEF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.5> to <26.0>." + } + ] + }, + { + "video_id": "62VEF", + "question_id": "62VEF", + "video_file": "Charades/videos/62VEF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a sandwich at the same time'." + }, + { + "from": "gpt", + "value": "From <18.5> to <26.0>." + } + ] + }, + { + "video_id": "CBXI2", + "question_id": "CBXI2", + "video_file": "Charades/videos/CBXI2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts awakening due to loud noise from vacuum cleaner'." + }, + { + "from": "gpt", + "value": "From <19.5> to <25.0>." + } + ] + }, + { + "video_id": "23YS1", + "question_id": "23YS1", + "video_file": "Charades/videos/23YS1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts groceries away there'." + }, + { + "from": "gpt", + "value": "From <10.4> to <18.6>." + } + ] + }, + { + "video_id": "23YS1", + "question_id": "23YS1", + "video_file": "Charades/videos/23YS1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person moves a chair to stand on'." + }, + { + "from": "gpt", + "value": "From <29.1> to <43.0>." + } + ] + }, + { + "video_id": "THW67", + "question_id": "THW67", + "video_file": "Charades/videos/THW67.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.5> to <30.0>." + } + ] + }, + { + "video_id": "THW67", + "question_id": "THW67", + "video_file": "Charades/videos/THW67.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person begins eating'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.7> to <25.8>." + } + ] + }, + { + "video_id": "FU5BL", + "question_id": "FU5BL", + "video_file": "Charades/videos/FU5BL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the phone down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.0> to <25.0>." + } + ] + }, + { + "video_id": "FU5BL", + "question_id": "FU5BL", + "video_file": "Charades/videos/FU5BL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person also puts the phone down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.0> to <25.0>." + } + ] + }, + { + "video_id": "FU5BL", + "question_id": "FU5BL", + "video_file": "Charades/videos/FU5BL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person eats a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <23.5>." + } + ] + }, + { + "video_id": "TGKHG", + "question_id": "TGKHG", + "video_file": "Charades/videos/TGKHG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person is eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.7> to <35.4>." + } + ] + }, + { + "video_id": "TGKHG", + "question_id": "TGKHG", + "video_file": "Charades/videos/TGKHG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person walks through the doorway'." + }, + { + "from": "gpt", + "value": "From <1.6> to <8.5>." + } + ] + }, + { + "video_id": "TGKHG", + "question_id": "TGKHG", + "video_file": "Charades/videos/TGKHG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.4>." + } + ] + }, + { + "video_id": "XMJTG", + "question_id": "XMJTG", + "video_file": "Charades/videos/XMJTG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person is opening the door'." + }, + { + "from": "gpt", + "value": "From <1.8> to <8.3>." + } + ] + }, + { + "video_id": "1ESKX", + "question_id": "1ESKX", + "video_file": "Charades/videos/1ESKX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person smiles as they look out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <25.2>." + } + ] + }, + { + "video_id": "1ESKX", + "question_id": "1ESKX", + "video_file": "Charades/videos/1ESKX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a camera'." + }, + { + "from": "gpt", + "value": "From <14.5> to <33.1>." + } + ] + }, + { + "video_id": "1ESKX", + "question_id": "1ESKX", + "video_file": "Charades/videos/1ESKX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person looks out a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.6> to <25.2>." + } + ] + }, + { + "video_id": "1ESKX", + "question_id": "1ESKX", + "video_file": "Charades/videos/1ESKX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a picture with a camera'." + }, + { + "from": "gpt", + "value": "From <19.3> to <33.1>." + } + ] + }, + { + "video_id": "1ESKX", + "question_id": "1ESKX", + "video_file": "Charades/videos/1ESKX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.3> to <33.1>." + } + ] + }, + { + "video_id": "82XPD", + "question_id": "82XPD", + "video_file": "Charades/videos/82XPD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person laughing the whole' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.0> to <25.0>." + } + ] + }, + { + "video_id": "82XPD", + "question_id": "82XPD", + "video_file": "Charades/videos/82XPD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turning a light on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <22.1>." + } + ] + }, + { + "video_id": "82XPD", + "question_id": "82XPD", + "video_file": "Charades/videos/82XPD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turn on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <22.1>." + } + ] + }, + { + "video_id": "W3I8W", + "question_id": "W3I8W", + "video_file": "Charades/videos/W3I8W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes a towel off a bed'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "W3I8W", + "question_id": "W3I8W", + "video_file": "Charades/videos/W3I8W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person turns off a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.7> to <27.2>." + } + ] + }, + { + "video_id": "W3I8W", + "question_id": "W3I8W", + "video_file": "Charades/videos/W3I8W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.7> to <27.2>." + } + ] + }, + { + "video_id": "N3SBJ", + "question_id": "N3SBJ", + "video_file": "Charades/videos/N3SBJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'this person is sitting at a table eating a sandwhich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.1> to <22.0>." + } + ] + }, + { + "video_id": "6BXG4", + "question_id": "6BXG4", + "video_file": "Charades/videos/6BXG4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sitting in chair in kitchen holding mug'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "6BXG4", + "question_id": "6BXG4", + "video_file": "Charades/videos/6BXG4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sitting in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "H4KNQ", + "question_id": "H4KNQ", + "video_file": "Charades/videos/H4KNQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the dishes in the sink'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <20.3>." + } + ] + }, + { + "video_id": "ARESU", + "question_id": "ARESU", + "video_file": "Charades/videos/ARESU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a second person walks over to the table holding bags'." + }, + { + "from": "gpt", + "value": "From <0.5> to <5.6>." + } + ] + }, + { + "video_id": "PAXNL", + "question_id": "PAXNL", + "video_file": "Charades/videos/PAXNL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is eating food in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.7> to <29.0>." + } + ] + }, + { + "video_id": "3FMAS", + "question_id": "3FMAS", + "video_file": "Charades/videos/3FMAS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a picture of themself' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.6> to <38.0>." + } + ] + }, + { + "video_id": "3FMAS", + "question_id": "3FMAS", + "video_file": "Charades/videos/3FMAS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take a picture of them self' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.6> to <38.0>." + } + ] + }, + { + "video_id": "RQRR6", + "question_id": "RQRR6", + "video_file": "Charades/videos/RQRR6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person close the laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <24.5>." + } + ] + }, + { + "video_id": "9JI9W", + "question_id": "9JI9W", + "video_file": "Charades/videos/9JI9W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins sneezing'." + }, + { + "from": "gpt", + "value": "From <27.5> to <36.4>." + } + ] + }, + { + "video_id": "DVBVZ", + "question_id": "DVBVZ", + "video_file": "Charades/videos/DVBVZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a bag from a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <9.3>." + } + ] + }, + { + "video_id": "DVBVZ", + "question_id": "DVBVZ", + "video_file": "Charades/videos/DVBVZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person takes a bag from a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <9.3>." + } + ] + }, + { + "video_id": "J4TLX", + "question_id": "J4TLX", + "video_file": "Charades/videos/J4TLX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they begin laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.5> to <29.6>." + } + ] + }, + { + "video_id": "KFC28", + "question_id": "KFC28", + "video_file": "Charades/videos/KFC28.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting down on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "KFC28", + "question_id": "KFC28", + "video_file": "Charades/videos/KFC28.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes a cabinet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.2> to <35.0>." + } + ] + }, + { + "video_id": "KFC28", + "question_id": "KFC28", + "video_file": "Charades/videos/KFC28.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person sits down in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "KFC28", + "question_id": "KFC28", + "video_file": "Charades/videos/KFC28.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing a cabinet in the living room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <29.2> to <35.0>." + } + ] + }, + { + "video_id": "KFC28", + "question_id": "KFC28", + "video_file": "Charades/videos/KFC28.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person close an open cabinet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <29.2> to <35.0>." + } + ] + }, + { + "video_id": "EAGYP", + "question_id": "EAGYP", + "video_file": "Charades/videos/EAGYP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes a bite of some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.5> to <19.4>." + } + ] + }, + { + "video_id": "EAGYP", + "question_id": "EAGYP", + "video_file": "Charades/videos/EAGYP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person smiling as they take a bite of their sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <19.4>." + } + ] + }, + { + "video_id": "XRWBL", + "question_id": "XRWBL", + "video_file": "Charades/videos/XRWBL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pick up their phone to talk' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.1> to <31.0>." + } + ] + }, + { + "video_id": "ICXAE", + "question_id": "ICXAE", + "video_file": "Charades/videos/ICXAE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting in a chair with food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.4>." + } + ] + }, + { + "video_id": "ICXAE", + "question_id": "ICXAE", + "video_file": "Charades/videos/ICXAE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'this person is sitting in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.4>." + } + ] + }, + { + "video_id": "ECLIB", + "question_id": "ECLIB", + "video_file": "Charades/videos/ECLIB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they take a glass off of a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.8> to <14.3>." + } + ] + }, + { + "video_id": "VGPCG", + "question_id": "VGPCG", + "video_file": "Charades/videos/VGPCG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts on a pair of shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.1> to <22.7>." + } + ] + }, + { + "video_id": "VGPCG", + "question_id": "VGPCG", + "video_file": "Charades/videos/VGPCG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the door behind them'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.5> to <32.0>." + } + ] + }, + { + "video_id": "9HOV3", + "question_id": "9HOV3", + "video_file": "Charades/videos/9HOV3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours it into a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <16.9>." + } + ] + }, + { + "video_id": "9HOV3", + "question_id": "9HOV3", + "video_file": "Charades/videos/9HOV3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pours something into a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <16.9>." + } + ] + }, + { + "video_id": "JCKRL", + "question_id": "JCKRL", + "video_file": "Charades/videos/JCKRL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is walking through a doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "JCKRL", + "question_id": "JCKRL", + "video_file": "Charades/videos/JCKRL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person walks in through a doorway'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "QBOOH", + "question_id": "QBOOH", + "video_file": "Charades/videos/QBOOH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looking out of a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "3LLXX", + "question_id": "3LLXX", + "video_file": "Charades/videos/3LLXX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is holding a cup'." + }, + { + "from": "gpt", + "value": "From <12.5> to <23.7>." + } + ] + }, + { + "video_id": "3LLXX", + "question_id": "3LLXX", + "video_file": "Charades/videos/3LLXX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes some medicine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.1> to <39.6>." + } + ] + }, + { + "video_id": "W4C5O", + "question_id": "W4C5O", + "video_file": "Charades/videos/W4C5O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person drinks from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.2>." + } + ] + }, + { + "video_id": "W4C5O", + "question_id": "W4C5O", + "video_file": "Charades/videos/W4C5O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a towel'." + }, + { + "from": "gpt", + "value": "From <15.2> to <21.9>." + } + ] + }, + { + "video_id": "W4C5O", + "question_id": "W4C5O", + "video_file": "Charades/videos/W4C5O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holds up a towel'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <21.9>." + } + ] + }, + { + "video_id": "5B4HF", + "question_id": "5B4HF", + "video_file": "Charades/videos/5B4HF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks through the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.0>." + } + ] + }, + { + "video_id": "5B4HF", + "question_id": "5B4HF", + "video_file": "Charades/videos/5B4HF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking a cup of coffee' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "5B4HF", + "question_id": "5B4HF", + "video_file": "Charades/videos/5B4HF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person he walks through a door drinking from a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "INA5N", + "question_id": "INA5N", + "video_file": "Charades/videos/INA5N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person washes the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.9> to <12.6>." + } + ] + }, + { + "video_id": "1IJ7V", + "question_id": "1IJ7V", + "video_file": "Charades/videos/1IJ7V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put the cup on a desk' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.4> to <23.1>." + } + ] + }, + { + "video_id": "1IJ7V", + "question_id": "1IJ7V", + "video_file": "Charades/videos/1IJ7V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person is putting coffee on the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.9> to <23.3>." + } + ] + }, + { + "video_id": "JFU3J", + "question_id": "JFU3J", + "video_file": "Charades/videos/JFU3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns the light off' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.4> to <26.7>." + } + ] + }, + { + "video_id": "JFU3J", + "question_id": "JFU3J", + "video_file": "Charades/videos/JFU3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person lay on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.2> to <34.0>." + } + ] + }, + { + "video_id": "JFU3J", + "question_id": "JFU3J", + "video_file": "Charades/videos/JFU3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person lays in the floor'." + }, + { + "from": "gpt", + "value": "From <26.2> to <34.0>." + } + ] + }, + { + "video_id": "JFU3J", + "question_id": "JFU3J", + "video_file": "Charades/videos/JFU3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they turn off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.4> to <26.7>." + } + ] + }, + { + "video_id": "JFU3J", + "question_id": "JFU3J", + "video_file": "Charades/videos/JFU3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turn off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.4> to <26.7>." + } + ] + }, + { + "video_id": "C4OY0", + "question_id": "C4OY0", + "video_file": "Charades/videos/C4OY0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes out their phone to make a call'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.5> to <23.4>." + } + ] + }, + { + "video_id": "C4OY0", + "question_id": "C4OY0", + "video_file": "Charades/videos/C4OY0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person talking on their phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.7> to <32.0>." + } + ] + }, + { + "video_id": "P7TT1", + "question_id": "P7TT1", + "video_file": "Charades/videos/P7TT1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throwing them on the floor'." + }, + { + "from": "gpt", + "value": "From <1.4> to <6.6>." + } + ] + }, + { + "video_id": "P1LWC", + "question_id": "P1LWC", + "video_file": "Charades/videos/P1LWC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person drinks from a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.9>." + } + ] + }, + { + "video_id": "YDTH0", + "question_id": "YDTH0", + "video_file": "Charades/videos/YDTH0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put it on the table'." + }, + { + "from": "gpt", + "value": "From <29.5> to <34.8>." + } + ] + }, + { + "video_id": "0M9ZO", + "question_id": "0M9ZO", + "video_file": "Charades/videos/0M9ZO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "0M9ZO", + "question_id": "0M9ZO", + "video_file": "Charades/videos/0M9ZO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.6> to <21.1>." + } + ] + }, + { + "video_id": "Y8XYL", + "question_id": "Y8XYL", + "video_file": "Charades/videos/Y8XYL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person sits in a chair in the garage' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "8SVWO", + "question_id": "8SVWO", + "video_file": "Charades/videos/8SVWO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.1> to <34.0>." + } + ] + }, + { + "video_id": "8SVWO", + "question_id": "8SVWO", + "video_file": "Charades/videos/8SVWO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person only eats the cheese' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.6> to <34.0>." + } + ] + }, + { + "video_id": "8SVWO", + "question_id": "8SVWO", + "video_file": "Charades/videos/8SVWO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a smiling person playing a game takes medicine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.7> to <23.0>." + } + ] + }, + { + "video_id": "3LM6H", + "question_id": "3LM6H", + "video_file": "Charades/videos/3LM6H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a broom'." + }, + { + "from": "gpt", + "value": "From <11.6> to <22.8>." + } + ] + }, + { + "video_id": "3LM6H", + "question_id": "3LM6H", + "video_file": "Charades/videos/3LM6H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is holding a broom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <22.8>." + } + ] + }, + { + "video_id": "3LM6H", + "question_id": "3LM6H", + "video_file": "Charades/videos/3LM6H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person in the garage begins undressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <13.5>." + } + ] + }, + { + "video_id": "XXDHJ", + "question_id": "XXDHJ", + "video_file": "Charades/videos/XXDHJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person turns on a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <7.8>." + } + ] + }, + { + "video_id": "XXDHJ", + "question_id": "XXDHJ", + "video_file": "Charades/videos/XXDHJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin fixing a doorknob' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.3> to <17.3>." + } + ] + }, + { + "video_id": "XXDHJ", + "question_id": "XXDHJ", + "video_file": "Charades/videos/XXDHJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person opens the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <20.0>." + } + ] + }, + { + "video_id": "XXDHJ", + "question_id": "XXDHJ", + "video_file": "Charades/videos/XXDHJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person begins to work on a doorknob opening' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.6> to <20.0>." + } + ] + }, + { + "video_id": "4T97L", + "question_id": "4T97L", + "video_file": "Charades/videos/4T97L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a second person runs by' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.3> to <16.8>." + } + ] + }, + { + "video_id": "4T97L", + "question_id": "4T97L", + "video_file": "Charades/videos/4T97L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person another turns the light on' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.5> to <17.5>." + } + ] + }, + { + "video_id": "4T97L", + "question_id": "4T97L", + "video_file": "Charades/videos/4T97L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person drinks coffee as another person runs in' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.3> to <16.8>." + } + ] + }, + { + "video_id": "FRC6O", + "question_id": "FRC6O", + "video_file": "Charades/videos/FRC6O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <14.3>." + } + ] + }, + { + "video_id": "FRC6O", + "question_id": "FRC6O", + "video_file": "Charades/videos/FRC6O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens their laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.4> to <14.3>." + } + ] + }, + { + "video_id": "BRJ0H", + "question_id": "BRJ0H", + "video_file": "Charades/videos/BRJ0H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is walking through a doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "BRJ0H", + "question_id": "BRJ0H", + "video_file": "Charades/videos/BRJ0H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a dish' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.6> to <36.0>." + } + ] + }, + { + "video_id": "BRJ0H", + "question_id": "BRJ0H", + "video_file": "Charades/videos/BRJ0H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts some food on their dish' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <21.5>." + } + ] + }, + { + "video_id": "A84J7", + "question_id": "A84J7", + "video_file": "Charades/videos/A84J7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turn off the lights' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.4> to <19.2>." + } + ] + }, + { + "video_id": "V9W9C", + "question_id": "V9W9C", + "video_file": "Charades/videos/V9W9C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes a drink from a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.2> to <36.1>." + } + ] + }, + { + "video_id": "V9W9C", + "question_id": "V9W9C", + "video_file": "Charades/videos/V9W9C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens at a desk in the basement'." + }, + { + "from": "gpt", + "value": "From <14.1> to <27.4>." + } + ] + }, + { + "video_id": "ZU00I", + "question_id": "ZU00I", + "video_file": "Charades/videos/ZU00I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person seems to drink some coffee out of her cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.9> to <28.4>." + } + ] + }, + { + "video_id": "ZU00I", + "question_id": "ZU00I", + "video_file": "Charades/videos/ZU00I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person than drinking from a cup'." + }, + { + "from": "gpt", + "value": "From <20.9> to <28.4>." + } + ] + }, + { + "video_id": "OJFYD", + "question_id": "OJFYD", + "video_file": "Charades/videos/OJFYD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes off their shoes'." + }, + { + "from": "gpt", + "value": "From <11.4> to <20.0>." + } + ] + }, + { + "video_id": "OJFYD", + "question_id": "OJFYD", + "video_file": "Charades/videos/OJFYD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes their shoes off' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.4> to <20.0>." + } + ] + }, + { + "video_id": "W9HYI", + "question_id": "W9HYI", + "video_file": "Charades/videos/W9HYI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person start eating'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <20.5>." + } + ] + }, + { + "video_id": "W9HYI", + "question_id": "W9HYI", + "video_file": "Charades/videos/W9HYI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating it on the stairs' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <20.5>." + } + ] + }, + { + "video_id": "8F17S", + "question_id": "8F17S", + "video_file": "Charades/videos/8F17S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person #2 was sitting on a couch eating a snack' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.3>." + } + ] + }, + { + "video_id": "8F17S", + "question_id": "8F17S", + "video_file": "Charades/videos/8F17S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.3>." + } + ] + }, + { + "video_id": "8F17S", + "question_id": "8F17S", + "video_file": "Charades/videos/8F17S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking from a glass of water'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "8F17S", + "question_id": "8F17S", + "video_file": "Charades/videos/8F17S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person is smiling'." + }, + { + "from": "gpt", + "value": "From <2.9> to <8.8>." + } + ] + }, + { + "video_id": "8F17S", + "question_id": "8F17S", + "video_file": "Charades/videos/8F17S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person is standing drinking a glass of water'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "8F17S", + "question_id": "8F17S", + "video_file": "Charades/videos/8F17S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is drinking from a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "8F17S", + "question_id": "8F17S", + "video_file": "Charades/videos/8F17S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person is sitting on a couch eating a snack'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.3>." + } + ] + }, + { + "video_id": "8F17S", + "question_id": "8F17S", + "video_file": "Charades/videos/8F17S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <18.2>." + } + ] + }, + { + "video_id": "L7HA6", + "question_id": "L7HA6", + "video_file": "Charades/videos/L7HA6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is drinking from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.0> to <26.7>." + } + ] + }, + { + "video_id": "L7HA6", + "question_id": "L7HA6", + "video_file": "Charades/videos/L7HA6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person starts undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.6> to <43.0>." + } + ] + }, + { + "video_id": "L7HA6", + "question_id": "L7HA6", + "video_file": "Charades/videos/L7HA6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person talks on phone puts on shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.0> to <32.9>." + } + ] + }, + { + "video_id": "L7HA6", + "question_id": "L7HA6", + "video_file": "Charades/videos/L7HA6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <29.6> to <43.0>." + } + ] + }, + { + "video_id": "CNS9B", + "question_id": "CNS9B", + "video_file": "Charades/videos/CNS9B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person a standing in a garage laughing with person b' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.5> to <10.3>." + } + ] + }, + { + "video_id": "CNS9B", + "question_id": "CNS9B", + "video_file": "Charades/videos/CNS9B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person two people are talking laughing using hand gestures' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <10.3>." + } + ] + }, + { + "video_id": "BQKW5", + "question_id": "BQKW5", + "video_file": "Charades/videos/BQKW5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes towels out of the box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.0> to <16.2>." + } + ] + }, + { + "video_id": "BQKW5", + "question_id": "BQKW5", + "video_file": "Charades/videos/BQKW5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes out some towels'." + }, + { + "from": "gpt", + "value": "From <5.0> to <16.2>." + } + ] + }, + { + "video_id": "BQKW5", + "question_id": "BQKW5", + "video_file": "Charades/videos/BQKW5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks into a room holding a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "OWGPH", + "question_id": "OWGPH", + "video_file": "Charades/videos/OWGPH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throwing their clothes into the desk' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.4> to <15.3>." + } + ] + }, + { + "video_id": "OWGPH", + "question_id": "OWGPH", + "video_file": "Charades/videos/OWGPH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is undressing in the laundry room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.9> to <15.5>." + } + ] + }, + { + "video_id": "51ZY5", + "question_id": "51ZY5", + "video_file": "Charades/videos/51ZY5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person run out of the kitchen' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.4> to <17.3>." + } + ] + }, + { + "video_id": "51ZY5", + "question_id": "51ZY5", + "video_file": "Charades/videos/51ZY5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person laughing at their phone'." + }, + { + "from": "gpt", + "value": "From <7.7> to <15.8>." + } + ] + }, + { + "video_id": "S4UF8", + "question_id": "S4UF8", + "video_file": "Charades/videos/S4UF8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person open the closet door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.6> to <32.3>." + } + ] + }, + { + "video_id": "S4UF8", + "question_id": "S4UF8", + "video_file": "Charades/videos/S4UF8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking a cup of coffee'." + }, + { + "from": "gpt", + "value": "From <3.6> to <10.9>." + } + ] + }, + { + "video_id": "320VL", + "question_id": "320VL", + "video_file": "Charades/videos/320VL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is watching a television' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.0>." + } + ] + }, + { + "video_id": "313G8", + "question_id": "313G8", + "video_file": "Charades/videos/313G8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the phone down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.9> to <27.6>." + } + ] + }, + { + "video_id": "47ZXN", + "question_id": "47ZXN", + "video_file": "Charades/videos/47ZXN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person stands on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.5> to <31.6>." + } + ] + }, + { + "video_id": "47ZXN", + "question_id": "47ZXN", + "video_file": "Charades/videos/47ZXN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is tidying a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "47ZXN", + "question_id": "47ZXN", + "video_file": "Charades/videos/47ZXN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person try to fix a broken light'." + }, + { + "from": "gpt", + "value": "From <19.5> to <30.1>." + } + ] + }, + { + "video_id": "47ZXN", + "question_id": "47ZXN", + "video_file": "Charades/videos/47ZXN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person turns on the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <13.8>." + } + ] + }, + { + "video_id": "47ZXN", + "question_id": "47ZXN", + "video_file": "Charades/videos/47ZXN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person touches the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.5> to <13.8>." + } + ] + }, + { + "video_id": "ZSL58", + "question_id": "ZSL58", + "video_file": "Charades/videos/ZSL58.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closing the door in the doorway'." + }, + { + "from": "gpt", + "value": "From <16.9> to <30.4>." + } + ] + }, + { + "video_id": "ZSL58", + "question_id": "ZSL58", + "video_file": "Charades/videos/ZSL58.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes a wardrobe door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.9> to <30.4>." + } + ] + }, + { + "video_id": "ZSL58", + "question_id": "ZSL58", + "video_file": "Charades/videos/ZSL58.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is laughing in front of a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "ZSL58", + "question_id": "ZSL58", + "video_file": "Charades/videos/ZSL58.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes an open cabinet door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.9> to <30.4>." + } + ] + }, + { + "video_id": "013SD", + "question_id": "013SD", + "video_file": "Charades/videos/013SD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts a box on a bed'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "013SD", + "question_id": "013SD", + "video_file": "Charades/videos/013SD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turning off a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.8> to <10.0>." + } + ] + }, + { + "video_id": "013SD", + "question_id": "013SD", + "video_file": "Charades/videos/013SD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they turned around to turn the bedroom light off' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.8> to <10.0>." + } + ] + }, + { + "video_id": "L4FWN", + "question_id": "L4FWN", + "video_file": "Charades/videos/L4FWN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person laughing at the picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <17.1>." + } + ] + }, + { + "video_id": "L4FWN", + "question_id": "L4FWN", + "video_file": "Charades/videos/L4FWN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person playing on his cell phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.8> to <32.0>." + } + ] + }, + { + "video_id": "D5IDA", + "question_id": "D5IDA", + "video_file": "Charades/videos/D5IDA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pouring themselves a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.9> to <16.0>." + } + ] + }, + { + "video_id": "D5IDA", + "question_id": "D5IDA", + "video_file": "Charades/videos/D5IDA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes the lid off of a sandwich'." + }, + { + "from": "gpt", + "value": "From <15.2> to <22.8>." + } + ] + }, + { + "video_id": "NCLD1", + "question_id": "NCLD1", + "video_file": "Charades/videos/NCLD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "NCLD1", + "question_id": "NCLD1", + "video_file": "Charades/videos/NCLD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throwing a pillow in a box over'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "NCLD1", + "question_id": "NCLD1", + "video_file": "Charades/videos/NCLD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is standing drinking a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "FIA2R", + "question_id": "FIA2R", + "video_file": "Charades/videos/FIA2R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opened a box in the corner'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <25.5>." + } + ] + }, + { + "video_id": "74S1R", + "question_id": "74S1R", + "video_file": "Charades/videos/74S1R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks through the doorway of the bathroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.1> to <8.4>." + } + ] + }, + { + "video_id": "74S1R", + "question_id": "74S1R", + "video_file": "Charades/videos/74S1R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns on light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "74S1R", + "question_id": "74S1R", + "video_file": "Charades/videos/74S1R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens plastic bag containing sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.3> to <16.8>." + } + ] + }, + { + "video_id": "74S1R", + "question_id": "74S1R", + "video_file": "Charades/videos/74S1R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person as they eat from a dish of food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <19.2>." + } + ] + }, + { + "video_id": "74S1R", + "question_id": "74S1R", + "video_file": "Charades/videos/74S1R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks through a doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <8.4>." + } + ] + }, + { + "video_id": "74S1R", + "question_id": "74S1R", + "video_file": "Charades/videos/74S1R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes the sandwich out of the bag'." + }, + { + "from": "gpt", + "value": "From <9.3> to <20.3>." + } + ] + }, + { + "video_id": "74S1R", + "question_id": "74S1R", + "video_file": "Charades/videos/74S1R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turning on the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "L8FFR", + "question_id": "L8FFR", + "video_file": "Charades/videos/L8FFR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take a blanket from a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.5> to <31.1>." + } + ] + }, + { + "video_id": "472EJ", + "question_id": "472EJ", + "video_file": "Charades/videos/472EJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put food in there' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.1> to <21.3>." + } + ] + }, + { + "video_id": "0BZAD", + "question_id": "0BZAD", + "video_file": "Charades/videos/0BZAD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats the food in the bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.2> to <31.3>." + } + ] + }, + { + "video_id": "0BZAD", + "question_id": "0BZAD", + "video_file": "Charades/videos/0BZAD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating something'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.2> to <31.3>." + } + ] + }, + { + "video_id": "ZUKHZ", + "question_id": "ZUKHZ", + "video_file": "Charades/videos/ZUKHZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person another runs by eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <12.2> to <17.5>." + } + ] + }, + { + "video_id": "ZUKHZ", + "question_id": "ZUKHZ", + "video_file": "Charades/videos/ZUKHZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person running down the stairs'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <18.4>." + } + ] + }, + { + "video_id": "ZUKHZ", + "question_id": "ZUKHZ", + "video_file": "Charades/videos/ZUKHZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating something' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <17.7>." + } + ] + }, + { + "video_id": "K4X20", + "question_id": "K4X20", + "video_file": "Charades/videos/K4X20.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking a glass of water'." + }, + { + "from": "gpt", + "value": "From <15.4> to <27.8>." + } + ] + }, + { + "video_id": "VEHER", + "question_id": "VEHER", + "video_file": "Charades/videos/VEHER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs into the laundry room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.3> to <10.3>." + } + ] + }, + { + "video_id": "ZAJG5", + "question_id": "ZAJG5", + "video_file": "Charades/videos/ZAJG5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person undresses' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <11.6>." + } + ] + }, + { + "video_id": "ZAJG5", + "question_id": "ZAJG5", + "video_file": "Charades/videos/ZAJG5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person with a stack of clothes starts undressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <11.6>." + } + ] + }, + { + "video_id": "ZAJG5", + "question_id": "ZAJG5", + "video_file": "Charades/videos/ZAJG5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting clothes on the sink takes off a jacket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <6.2>." + } + ] + }, + { + "video_id": "KAKC6", + "question_id": "KAKC6", + "video_file": "Charades/videos/KAKC6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is working by reading a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <18.4>." + } + ] + }, + { + "video_id": "82B2V", + "question_id": "82B2V", + "video_file": "Charades/videos/82B2V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a camera' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <17.0>." + } + ] + }, + { + "video_id": "DW4YF", + "question_id": "DW4YF", + "video_file": "Charades/videos/DW4YF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person is sitting at desk drinking cup of coffee'." + }, + { + "from": "gpt", + "value": "From <2.1> to <8.1>." + } + ] + }, + { + "video_id": "DW4YF", + "question_id": "DW4YF", + "video_file": "Charades/videos/DW4YF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is drinking a cup of coffee'." + }, + { + "from": "gpt", + "value": "From <2.1> to <8.1>." + } + ] + }, + { + "video_id": "DVM8I", + "question_id": "DVM8I", + "video_file": "Charades/videos/DVM8I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <14.8>." + } + ] + }, + { + "video_id": "ZRXQ1", + "question_id": "ZRXQ1", + "video_file": "Charades/videos/ZRXQ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person a runs out of the room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <20.6>." + } + ] + }, + { + "video_id": "ZRXQ1", + "question_id": "ZRXQ1", + "video_file": "Charades/videos/ZRXQ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person runs out the room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.9> to <20.6>." + } + ] + }, + { + "video_id": "6Z9ZC", + "question_id": "6Z9ZC", + "video_file": "Charades/videos/6Z9ZC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts away some clothes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.2> to <18.1>." + } + ] + }, + { + "video_id": "6Z9ZC", + "question_id": "6Z9ZC", + "video_file": "Charades/videos/6Z9ZC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person someone putting clothes in a drawer' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.2> to <18.1>." + } + ] + }, + { + "video_id": "3FEYY", + "question_id": "3FEYY", + "video_file": "Charades/videos/3FEYY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting the towel on an ornament'." + }, + { + "from": "gpt", + "value": "From <12.3> to <17.3>." + } + ] + }, + { + "video_id": "3FEYY", + "question_id": "3FEYY", + "video_file": "Charades/videos/3FEYY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person cooks at the stove for a moment' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <26.0>." + } + ] + }, + { + "video_id": "UIPJB", + "question_id": "UIPJB", + "video_file": "Charades/videos/UIPJB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person is sneezing on the food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <8.7>." + } + ] + }, + { + "video_id": "BLFEV", + "question_id": "BLFEV", + "video_file": "Charades/videos/BLFEV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they pour water into a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.3> to <29.1>." + } + ] + }, + { + "video_id": "BLFEV", + "question_id": "BLFEV", + "video_file": "Charades/videos/BLFEV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they stand up'." + }, + { + "from": "gpt", + "value": "From <34.0> to <40.0>." + } + ] + }, + { + "video_id": "BLFEV", + "question_id": "BLFEV", + "video_file": "Charades/videos/BLFEV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pours a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.3> to <29.1>." + } + ] + }, + { + "video_id": "ERC0T", + "question_id": "ERC0T", + "video_file": "Charades/videos/ERC0T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person fixing a light'." + }, + { + "from": "gpt", + "value": "From <12.8> to <26.5>." + } + ] + }, + { + "video_id": "ERC0T", + "question_id": "ERC0T", + "video_file": "Charades/videos/ERC0T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person fixes a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <26.5>." + } + ] + }, + { + "video_id": "3QY79", + "question_id": "3QY79", + "video_file": "Charades/videos/3QY79.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes more clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.8> to <37.1>." + } + ] + }, + { + "video_id": "3QY79", + "question_id": "3QY79", + "video_file": "Charades/videos/3QY79.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take some new clothes from a cabinet'." + }, + { + "from": "gpt", + "value": "From <25.8> to <37.1>." + } + ] + }, + { + "video_id": "3QY79", + "question_id": "3QY79", + "video_file": "Charades/videos/3QY79.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts away the glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.9> to <15.6>." + } + ] + }, + { + "video_id": "3QY79", + "question_id": "3QY79", + "video_file": "Charades/videos/3QY79.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person so they begin undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.9> to <29.2>." + } + ] + }, + { + "video_id": "MDG2T", + "question_id": "MDG2T", + "video_file": "Charades/videos/MDG2T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws it on the floor'." + }, + { + "from": "gpt", + "value": "From <13.8> to <20.4>." + } + ] + }, + { + "video_id": "MDG2T", + "question_id": "MDG2T", + "video_file": "Charades/videos/MDG2T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person throws the towel on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.8> to <20.4>." + } + ] + }, + { + "video_id": "MX4FM", + "question_id": "MX4FM", + "video_file": "Charades/videos/MX4FM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they begin sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <20.8>." + } + ] + }, + { + "video_id": "B5R5B", + "question_id": "B5R5B", + "video_file": "Charades/videos/B5R5B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs in'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "B5R5B", + "question_id": "B5R5B", + "video_file": "Charades/videos/B5R5B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes out their phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.5>." + } + ] + }, + { + "video_id": "B5R5B", + "question_id": "B5R5B", + "video_file": "Charades/videos/B5R5B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs inside excitedly with a sandwich in hand' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "B5R5B", + "question_id": "B5R5B", + "video_file": "Charades/videos/B5R5B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs down the hallway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "3MIWR", + "question_id": "3MIWR", + "video_file": "Charades/videos/3MIWR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person start sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <13.0>." + } + ] + }, + { + "video_id": "3MIWR", + "question_id": "3MIWR", + "video_file": "Charades/videos/3MIWR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.3> to <13.0>." + } + ] + }, + { + "video_id": "0ZC7L", + "question_id": "0ZC7L", + "video_file": "Charades/videos/0ZC7L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person start playing on their phone'." + }, + { + "from": "gpt", + "value": "From <23.1> to <39.0>." + } + ] + }, + { + "video_id": "CA7T0", + "question_id": "CA7T0", + "video_file": "Charades/videos/CA7T0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts down the food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.5> to <30.2>." + } + ] + }, + { + "video_id": "CA7T0", + "question_id": "CA7T0", + "video_file": "Charades/videos/CA7T0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a box from on top of the dryer'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.4> to <40.0>." + } + ] + }, + { + "video_id": "UK5AT", + "question_id": "UK5AT", + "video_file": "Charades/videos/UK5AT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person talk on the phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.1> to <25.1>." + } + ] + }, + { + "video_id": "TSKZL", + "question_id": "TSKZL", + "video_file": "Charades/videos/TSKZL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person closes the dryer door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <20.2>." + } + ] + }, + { + "video_id": "4VYE5", + "question_id": "4VYE5", + "video_file": "Charades/videos/4VYE5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they stand up from the chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <91.7> to <102.0>." + } + ] + }, + { + "video_id": "QRULM", + "question_id": "QRULM", + "video_file": "Charades/videos/QRULM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins looking at a picture' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <28.5> to <36.0>." + } + ] + }, + { + "video_id": "QRULM", + "question_id": "QRULM", + "video_file": "Charades/videos/QRULM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looks at a picture' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.5> to <36.0>." + } + ] + }, + { + "video_id": "QRULM", + "question_id": "QRULM", + "video_file": "Charades/videos/QRULM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person smiling to themselves'." + }, + { + "from": "gpt", + "value": "From <30.8> to <36.0>." + } + ] + }, + { + "video_id": "0USFX", + "question_id": "0USFX", + "video_file": "Charades/videos/0USFX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "0USFX", + "question_id": "0USFX", + "video_file": "Charades/videos/0USFX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.5> to <20.0>." + } + ] + }, + { + "video_id": "0USFX", + "question_id": "0USFX", + "video_file": "Charades/videos/0USFX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes the vacuum' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.5> to <21.1>." + } + ] + }, + { + "video_id": "0USFX", + "question_id": "0USFX", + "video_file": "Charades/videos/0USFX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts groceries on the shelf close the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <19.5>." + } + ] + }, + { + "video_id": "0USFX", + "question_id": "0USFX", + "video_file": "Charades/videos/0USFX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes the vacuum out of the room'." + }, + { + "from": "gpt", + "value": "From <13.5> to <21.1>." + } + ] + }, + { + "video_id": "UJJ6Z", + "question_id": "UJJ6Z", + "video_file": "Charades/videos/UJJ6Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is putting the top on a medicine bottle'." + }, + { + "from": "gpt", + "value": "From <4.1> to <10.5>." + } + ] + }, + { + "video_id": "SRTW8", + "question_id": "SRTW8", + "video_file": "Charades/videos/SRTW8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person lies on a couch'." + }, + { + "from": "gpt", + "value": "From <8.2> to <15.3>." + } + ] + }, + { + "video_id": "SRTW8", + "question_id": "SRTW8", + "video_file": "Charades/videos/SRTW8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is smiling' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.1> to <18.1>." + } + ] + }, + { + "video_id": "SRTW8", + "question_id": "SRTW8", + "video_file": "Charades/videos/SRTW8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens on a fold-up bed'." + }, + { + "from": "gpt", + "value": "From <5.9> to <12.3>." + } + ] + }, + { + "video_id": "9HTQ1", + "question_id": "9HTQ1", + "video_file": "Charades/videos/9HTQ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put the pillow down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <22.2>." + } + ] + }, + { + "video_id": "9HTQ1", + "question_id": "9HTQ1", + "video_file": "Charades/videos/9HTQ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person stand up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.4> to <22.8>." + } + ] + }, + { + "video_id": "CB1Z2", + "question_id": "CB1Z2", + "video_file": "Charades/videos/CB1Z2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.9> to <22.0>." + } + ] + }, + { + "video_id": "CB1Z2", + "question_id": "CB1Z2", + "video_file": "Charades/videos/CB1Z2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a box'." + }, + { + "from": "gpt", + "value": "From <16.9> to <22.0>." + } + ] + }, + { + "video_id": "Y4X8J", + "question_id": "Y4X8J", + "video_file": "Charades/videos/Y4X8J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pull the window open' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.2> to <31.8>." + } + ] + }, + { + "video_id": "KB0Y8", + "question_id": "KB0Y8", + "video_file": "Charades/videos/KB0Y8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throw it on the floor'." + }, + { + "from": "gpt", + "value": "From <4.8> to <9.0>." + } + ] + }, + { + "video_id": "KB0Y8", + "question_id": "KB0Y8", + "video_file": "Charades/videos/KB0Y8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws it to the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.8> to <9.0>." + } + ] + }, + { + "video_id": "G4IV1", + "question_id": "G4IV1", + "video_file": "Charades/videos/G4IV1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.9> to <12.4>." + } + ] + }, + { + "video_id": "G4IV1", + "question_id": "G4IV1", + "video_file": "Charades/videos/G4IV1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs to the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.4> to <6.6>." + } + ] + }, + { + "video_id": "FU7HP", + "question_id": "FU7HP", + "video_file": "Charades/videos/FU7HP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.9>." + } + ] + }, + { + "video_id": "FU7HP", + "question_id": "FU7HP", + "video_file": "Charades/videos/FU7HP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the refrigerator door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.9>." + } + ] + }, + { + "video_id": "FU7HP", + "question_id": "FU7HP", + "video_file": "Charades/videos/FU7HP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door,walks over to the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <15.8>." + } + ] + }, + { + "video_id": "FU7HP", + "question_id": "FU7HP", + "video_file": "Charades/videos/FU7HP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the laptop on top of it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <20.5>." + } + ] + }, + { + "video_id": "GY9MZ", + "question_id": "GY9MZ", + "video_file": "Charades/videos/GY9MZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take off their shoes'." + }, + { + "from": "gpt", + "value": "From <26.7> to <36.0>." + } + ] + }, + { + "video_id": "3C9R2", + "question_id": "3C9R2", + "video_file": "Charades/videos/3C9R2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <23.4>." + } + ] + }, + { + "video_id": "GHJ8Z", + "question_id": "GHJ8Z", + "video_file": "Charades/videos/GHJ8Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a box of medicine on the counter' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.8> to <27.4>." + } + ] + }, + { + "video_id": "CSHUE", + "question_id": "CSHUE", + "video_file": "Charades/videos/CSHUE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is cooking some food at a stove'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.6>." + } + ] + }, + { + "video_id": "CSHUE", + "question_id": "CSHUE", + "video_file": "Charades/videos/CSHUE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they begin putting a blanket into their refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.1> to <17.9>." + } + ] + }, + { + "video_id": "CSHUE", + "question_id": "CSHUE", + "video_file": "Charades/videos/CSHUE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person trying to cook'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.6>." + } + ] + }, + { + "video_id": "61WBK", + "question_id": "61WBK", + "video_file": "Charades/videos/61WBK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is throwing clothes down the stairs' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "HXWI9", + "question_id": "HXWI9", + "video_file": "Charades/videos/HXWI9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person holding a picture twirls around the no purpose'." + }, + { + "from": "gpt", + "value": "From <18.3> to <28.5>." + } + ] + }, + { + "video_id": "HXWI9", + "question_id": "HXWI9", + "video_file": "Charades/videos/HXWI9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person 1 runs into the room holding a picture' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.3> to <28.5>." + } + ] + }, + { + "video_id": "81QUM", + "question_id": "81QUM", + "video_file": "Charades/videos/81QUM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts broom next to door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <19.7>." + } + ] + }, + { + "video_id": "81QUM", + "question_id": "81QUM", + "video_file": "Charades/videos/81QUM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts throwing shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.4> to <24.4>." + } + ] + }, + { + "video_id": "81QUM", + "question_id": "81QUM", + "video_file": "Charades/videos/81QUM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts laughing'." + }, + { + "from": "gpt", + "value": "From <22.5> to <29.0>." + } + ] + }, + { + "video_id": "FPJKR", + "question_id": "FPJKR", + "video_file": "Charades/videos/FPJKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a glass off of a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <9.4>." + } + ] + }, + { + "video_id": "FPJKR", + "question_id": "FPJKR", + "video_file": "Charades/videos/FPJKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs into their home office'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.5> to <15.5>." + } + ] + }, + { + "video_id": "FPJKR", + "question_id": "FPJKR", + "video_file": "Charades/videos/FPJKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a glass from the shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.2> to <9.4>." + } + ] + }, + { + "video_id": "FPJKR", + "question_id": "FPJKR", + "video_file": "Charades/videos/FPJKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs into a room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.5> to <15.5>." + } + ] + }, + { + "video_id": "9JPV3", + "question_id": "9JPV3", + "video_file": "Charades/videos/9JPV3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person closes the closet door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "9JPV3", + "question_id": "9JPV3", + "video_file": "Charades/videos/9JPV3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a man closes a closet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "LBGFT", + "question_id": "LBGFT", + "video_file": "Charades/videos/LBGFT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits on a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.6> to <9.0>." + } + ] + }, + { + "video_id": "LBGFT", + "question_id": "LBGFT", + "video_file": "Charades/videos/LBGFT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take off their shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.3> to <15.0>." + } + ] + }, + { + "video_id": "F1C8I", + "question_id": "F1C8I", + "video_file": "Charades/videos/F1C8I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding shoes'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "VOBBG", + "question_id": "VOBBG", + "video_file": "Charades/videos/VOBBG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes the dishes to the kitchen' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <27.9>." + } + ] + }, + { + "video_id": "VOBBG", + "question_id": "VOBBG", + "video_file": "Charades/videos/VOBBG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes the dish to near the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.8> to <27.9>." + } + ] + }, + { + "video_id": "VOBBG", + "question_id": "VOBBG", + "video_file": "Charades/videos/VOBBG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes a picture of their food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <28.0>." + } + ] + }, + { + "video_id": "A0PVR", + "question_id": "A0PVR", + "video_file": "Charades/videos/A0PVR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person closes a window beside a large mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <10.8>." + } + ] + }, + { + "video_id": "A0PVR", + "question_id": "A0PVR", + "video_file": "Charades/videos/A0PVR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <10.8>." + } + ] + }, + { + "video_id": "C1QDS", + "question_id": "C1QDS", + "video_file": "Charades/videos/C1QDS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person laughs at him also holding a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <18.7>." + } + ] + }, + { + "video_id": "C1QDS", + "question_id": "C1QDS", + "video_file": "Charades/videos/C1QDS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is pouring something back in forth in cups' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.1>." + } + ] + }, + { + "video_id": "TZSIE", + "question_id": "TZSIE", + "video_file": "Charades/videos/TZSIE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walks to the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <29.7>." + } + ] + }, + { + "video_id": "9207X", + "question_id": "9207X", + "video_file": "Charades/videos/9207X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening the cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <16.0>." + } + ] + }, + { + "video_id": "KRZW8", + "question_id": "KRZW8", + "video_file": "Charades/videos/KRZW8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person checks their cellphone smiling' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.0> to <33.0>." + } + ] + }, + { + "video_id": "KRZW8", + "question_id": "KRZW8", + "video_file": "Charades/videos/KRZW8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.0> to <32.0>." + } + ] + }, + { + "video_id": "D3N3P", + "question_id": "D3N3P", + "video_file": "Charades/videos/D3N3P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws it on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.5> to <18.0>." + } + ] + }, + { + "video_id": "CUZND", + "question_id": "CUZND", + "video_file": "Charades/videos/CUZND.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.9> to <24.1>." + } + ] + }, + { + "video_id": "CUZND", + "question_id": "CUZND", + "video_file": "Charades/videos/CUZND.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.2> to <17.0>." + } + ] + }, + { + "video_id": "CUZND", + "question_id": "CUZND", + "video_file": "Charades/videos/CUZND.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begin smiling in the mirror' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.2> to <17.0>." + } + ] + }, + { + "video_id": "CUZND", + "question_id": "CUZND", + "video_file": "Charades/videos/CUZND.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turning the light on'." + }, + { + "from": "gpt", + "value": "From <12.5> to <17.4>." + } + ] + }, + { + "video_id": "CUZND", + "question_id": "CUZND", + "video_file": "Charades/videos/CUZND.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person laughing at the bathroom mirror' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <29.7> to <36.7>." + } + ] + }, + { + "video_id": "CUZND", + "question_id": "CUZND", + "video_file": "Charades/videos/CUZND.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turn on the light'." + }, + { + "from": "gpt", + "value": "From <12.5> to <17.4>." + } + ] + }, + { + "video_id": "SJJ35", + "question_id": "SJJ35", + "video_file": "Charades/videos/SJJ35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a young person drinking coffee from a cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "SJJ35", + "question_id": "SJJ35", + "video_file": "Charades/videos/SJJ35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person was holding the dish with the pizza on it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "SJJ35", + "question_id": "SJJ35", + "video_file": "Charades/videos/SJJ35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they eat a sandwich'." + }, + { + "from": "gpt", + "value": "From <10.3> to <15.4>." + } + ] + }, + { + "video_id": "SJJ35", + "question_id": "SJJ35", + "video_file": "Charades/videos/SJJ35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person put the drink on the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <8.8>." + } + ] + }, + { + "video_id": "SJJ35", + "question_id": "SJJ35", + "video_file": "Charades/videos/SJJ35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is pouring a cup of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "SJJ35", + "question_id": "SJJ35", + "video_file": "Charades/videos/SJJ35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person pours coffee into a cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "TMA9G", + "question_id": "TMA9G", + "video_file": "Charades/videos/TMA9G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open a refrigerator door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.2> to <14.5>." + } + ] + }, + { + "video_id": "TMA9G", + "question_id": "TMA9G", + "video_file": "Charades/videos/TMA9G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.5> to <25.3>." + } + ] + }, + { + "video_id": "EC695", + "question_id": "EC695", + "video_file": "Charades/videos/EC695.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person another runs into the room to grab a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.6> to <32.0>." + } + ] + }, + { + "video_id": "OIRPJ", + "question_id": "OIRPJ", + "video_file": "Charades/videos/OIRPJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person are seen undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.0> to <14.6>." + } + ] + }, + { + "video_id": "VWAEL", + "question_id": "VWAEL", + "video_file": "Charades/videos/VWAEL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put it on a shelf'." + }, + { + "from": "gpt", + "value": "From <23.9> to <30.1>." + } + ] + }, + { + "video_id": "R9NRA", + "question_id": "R9NRA", + "video_file": "Charades/videos/R9NRA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person a is pouring a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.5> to <21.9>." + } + ] + }, + { + "video_id": "8DQ3V", + "question_id": "8DQ3V", + "video_file": "Charades/videos/8DQ3V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is smiling at a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "8DQ3V", + "question_id": "8DQ3V", + "video_file": "Charades/videos/8DQ3V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is looking out a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.5>." + } + ] + }, + { + "video_id": "8I352", + "question_id": "8I352", + "video_file": "Charades/videos/8I352.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person sits at the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <30.0>." + } + ] + }, + { + "video_id": "8NSDT", + "question_id": "8NSDT", + "video_file": "Charades/videos/8NSDT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a camera' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.9>." + } + ] + }, + { + "video_id": "8NSDT", + "question_id": "8NSDT", + "video_file": "Charades/videos/8NSDT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks into an entry way holding a camera' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.9>." + } + ] + }, + { + "video_id": "8NSDT", + "question_id": "8NSDT", + "video_file": "Charades/videos/8NSDT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the camera on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <12.4>." + } + ] + }, + { + "video_id": "KJ428", + "question_id": "KJ428", + "video_file": "Charades/videos/KJ428.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person closes a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.6>." + } + ] + }, + { + "video_id": "CSQLJ", + "question_id": "CSQLJ", + "video_file": "Charades/videos/CSQLJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sneezing on a dusty book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <7.2>." + } + ] + }, + { + "video_id": "57A1A", + "question_id": "57A1A", + "video_file": "Charades/videos/57A1A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a laundry closet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.0> to <9.0>." + } + ] + }, + { + "video_id": "57A1A", + "question_id": "57A1A", + "video_file": "Charades/videos/57A1A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws on the floor'." + }, + { + "from": "gpt", + "value": "From <11.0> to <16.0>." + } + ] + }, + { + "video_id": "57A1A", + "question_id": "57A1A", + "video_file": "Charades/videos/57A1A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person in a laundry room opens a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <9.0>." + } + ] + }, + { + "video_id": "Q7YXN", + "question_id": "Q7YXN", + "video_file": "Charades/videos/Q7YXN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting clothes in the washing machine'." + }, + { + "from": "gpt", + "value": "From <12.4> to <27.2>." + } + ] + }, + { + "video_id": "RVLLF", + "question_id": "RVLLF", + "video_file": "Charades/videos/RVLLF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts the towel on the shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <18.7>." + } + ] + }, + { + "video_id": "XBTN7", + "question_id": "XBTN7", + "video_file": "Charades/videos/XBTN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinks from a bottle of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <21.5>." + } + ] + }, + { + "video_id": "XBTN7", + "question_id": "XBTN7", + "video_file": "Charades/videos/XBTN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person throws a pile of clothes onto the sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.3> to <9.3>." + } + ] + }, + { + "video_id": "XBTN7", + "question_id": "XBTN7", + "video_file": "Charades/videos/XBTN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person drinks from a bottle of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <21.5>." + } + ] + }, + { + "video_id": "XBTN7", + "question_id": "XBTN7", + "video_file": "Charades/videos/XBTN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws clothes on the sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.3> to <9.3>." + } + ] + }, + { + "video_id": "6EU06", + "question_id": "6EU06", + "video_file": "Charades/videos/6EU06.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person can close a book in the hallway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <9.4>." + } + ] + }, + { + "video_id": "6EU06", + "question_id": "6EU06", + "video_file": "Charades/videos/6EU06.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person can wash a table in the hallway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.4> to <32.0>." + } + ] + }, + { + "video_id": "9SZG2", + "question_id": "9SZG2", + "video_file": "Charades/videos/9SZG2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is walking past the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "QODUB", + "question_id": "QODUB", + "video_file": "Charades/videos/QODUB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <27.0> to <33.1>." + } + ] + }, + { + "video_id": "R00L7", + "question_id": "R00L7", + "video_file": "Charades/videos/R00L7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looks at a recipe book'." + }, + { + "from": "gpt", + "value": "From <15.3> to <24.7>." + } + ] + }, + { + "video_id": "1HP1X", + "question_id": "1HP1X", + "video_file": "Charades/videos/1HP1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes a blanket off a chair'." + }, + { + "from": "gpt", + "value": "From <16.6> to <23.6>." + } + ] + }, + { + "video_id": "FSWT4", + "question_id": "FSWT4", + "video_file": "Charades/videos/FSWT4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is undressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <17.0>." + } + ] + }, + { + "video_id": "FSWT4", + "question_id": "FSWT4", + "video_file": "Charades/videos/FSWT4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting the clothes on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.0> to <23.2>." + } + ] + }, + { + "video_id": "3C9L1", + "question_id": "3C9L1", + "video_file": "Charades/videos/3C9L1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the laptop on the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.2> to <35.0>." + } + ] + }, + { + "video_id": "L9ANI", + "question_id": "L9ANI", + "video_file": "Charades/videos/L9ANI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks into a kitchen holding a grocery bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.0> to <10.5>." + } + ] + }, + { + "video_id": "RVV5Q", + "question_id": "RVV5Q", + "video_file": "Charades/videos/RVV5Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person turns on a light switch' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <6.6>." + } + ] + }, + { + "video_id": "RVV5Q", + "question_id": "RVV5Q", + "video_file": "Charades/videos/RVV5Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.9> to <6.6>." + } + ] + }, + { + "video_id": "EM0B6", + "question_id": "EM0B6", + "video_file": "Charades/videos/EM0B6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <30.0>." + } + ] + }, + { + "video_id": "8QTVQ", + "question_id": "8QTVQ", + "video_file": "Charades/videos/8QTVQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is holding clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <19.8>." + } + ] + }, + { + "video_id": "8QTVQ", + "question_id": "8QTVQ", + "video_file": "Charades/videos/8QTVQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a pile of clothes'." + }, + { + "from": "gpt", + "value": "From <0.0> to <19.8>." + } + ] + }, + { + "video_id": "8QTVQ", + "question_id": "8QTVQ", + "video_file": "Charades/videos/8QTVQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a man holding a bunch of clothes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <19.8>." + } + ] + }, + { + "video_id": "5VINT", + "question_id": "5VINT", + "video_file": "Charades/videos/5VINT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person walk through a doorway'." + }, + { + "from": "gpt", + "value": "From <23.7> to <29.7>." + } + ] + }, + { + "video_id": "5VINT", + "question_id": "5VINT", + "video_file": "Charades/videos/5VINT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closing the door behind them' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.4> to <32.0>." + } + ] + }, + { + "video_id": "YNIWQ", + "question_id": "YNIWQ", + "video_file": "Charades/videos/YNIWQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the book under their arm'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <23.0>." + } + ] + }, + { + "video_id": "YNIWQ", + "question_id": "YNIWQ", + "video_file": "Charades/videos/YNIWQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes out a piece of paper' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <23.3>." + } + ] + }, + { + "video_id": "LW51F", + "question_id": "LW51F", + "video_file": "Charades/videos/LW51F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person closes the book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <17.0>." + } + ] + }, + { + "video_id": "LW51F", + "question_id": "LW51F", + "video_file": "Charades/videos/LW51F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.3> to <17.0>." + } + ] + }, + { + "video_id": "22A2G", + "question_id": "22A2G", + "video_file": "Charades/videos/22A2G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they begin sneezing as they close the book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.7> to <37.0>." + } + ] + }, + { + "video_id": "UXET8", + "question_id": "UXET8", + "video_file": "Charades/videos/UXET8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person puts a towel in a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <9.8>." + } + ] + }, + { + "video_id": "UXET8", + "question_id": "UXET8", + "video_file": "Charades/videos/UXET8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put the bag on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <20.0>." + } + ] + }, + { + "video_id": "UXET8", + "question_id": "UXET8", + "video_file": "Charades/videos/UXET8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.3> to <15.4>." + } + ] + }, + { + "video_id": "B8JAF", + "question_id": "B8JAF", + "video_file": "Charades/videos/B8JAF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the cabinet door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <19.5>." + } + ] + }, + { + "video_id": "B8JAF", + "question_id": "B8JAF", + "video_file": "Charades/videos/B8JAF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins washing the dishes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.1> to <27.9>." + } + ] + }, + { + "video_id": "B8JAF", + "question_id": "B8JAF", + "video_file": "Charades/videos/B8JAF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is opening the wardrobe door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "B8JAF", + "question_id": "B8JAF", + "video_file": "Charades/videos/B8JAF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens up cabinet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "0D5JP", + "question_id": "0D5JP", + "video_file": "Charades/videos/0D5JP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs to table,grabs coffee drinks it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "0D5JP", + "question_id": "0D5JP", + "video_file": "Charades/videos/0D5JP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person runs to look outside window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.6> to <31.0>." + } + ] + }, + { + "video_id": "0M7YH", + "question_id": "0M7YH", + "video_file": "Charades/videos/0M7YH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <7.7>." + } + ] + }, + { + "video_id": "1EJKT", + "question_id": "1EJKT", + "video_file": "Charades/videos/1EJKT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is dressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "1EJKT", + "question_id": "1EJKT", + "video_file": "Charades/videos/1EJKT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they begin undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "PGKB4", + "question_id": "PGKB4", + "video_file": "Charades/videos/PGKB4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <14.2>." + } + ] + }, + { + "video_id": "PGKB4", + "question_id": "PGKB4", + "video_file": "Charades/videos/PGKB4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person quickly opens their laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.7> to <22.4>." + } + ] + }, + { + "video_id": "PGKB4", + "question_id": "PGKB4", + "video_file": "Charades/videos/PGKB4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open their laptop'." + }, + { + "from": "gpt", + "value": "From <14.7> to <22.4>." + } + ] + }, + { + "video_id": "PGKB4", + "question_id": "PGKB4", + "video_file": "Charades/videos/PGKB4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <3.2> to <14.2>." + } + ] + }, + { + "video_id": "PGKB4", + "question_id": "PGKB4", + "video_file": "Charades/videos/PGKB4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs into their recreation room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <14.2>." + } + ] + }, + { + "video_id": "T6OAB", + "question_id": "T6OAB", + "video_file": "Charades/videos/T6OAB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person decided to grasp the light switch' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.0> to <26.0>." + } + ] + }, + { + "video_id": "T6OAB", + "question_id": "T6OAB", + "video_file": "Charades/videos/T6OAB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens the medicine cabinet doors'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "T6OAB", + "question_id": "T6OAB", + "video_file": "Charades/videos/T6OAB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person flicks the light switch'." + }, + { + "from": "gpt", + "value": "From <21.0> to <26.0>." + } + ] + }, + { + "video_id": "CY6U7", + "question_id": "CY6U7", + "video_file": "Charades/videos/CY6U7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person sits in a chair'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.3>." + } + ] + }, + { + "video_id": "CY6U7", + "question_id": "CY6U7", + "video_file": "Charades/videos/CY6U7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person stand up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.5> to <12.7>." + } + ] + }, + { + "video_id": "SVA4D", + "question_id": "SVA4D", + "video_file": "Charades/videos/SVA4D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person is putting a laptop on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.3> to <24.6>." + } + ] + }, + { + "video_id": "SOR8E", + "question_id": "SOR8E", + "video_file": "Charades/videos/SOR8E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the refrigerator door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.6> to <10.6>." + } + ] + }, + { + "video_id": "SOR8E", + "question_id": "SOR8E", + "video_file": "Charades/videos/SOR8E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they open the door of the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <7.4>." + } + ] + }, + { + "video_id": "QHNTC", + "question_id": "QHNTC", + "video_file": "Charades/videos/QHNTC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the laptop on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <28.7>." + } + ] + }, + { + "video_id": "1T8RK", + "question_id": "1T8RK", + "video_file": "Charades/videos/1T8RK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person grabs bag from floor take towel out' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <7.1>." + } + ] + }, + { + "video_id": "1T8RK", + "question_id": "1T8RK", + "video_file": "Charades/videos/1T8RK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person opens a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.7> to <11.5>." + } + ] + }, + { + "video_id": "1T8RK", + "question_id": "1T8RK", + "video_file": "Charades/videos/1T8RK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes out a towel' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.2> to <12.3>." + } + ] + }, + { + "video_id": "1T8RK", + "question_id": "1T8RK", + "video_file": "Charades/videos/1T8RK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person is fixing hair in mirror' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "X0BEF", + "question_id": "X0BEF", + "video_file": "Charades/videos/X0BEF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is throwing shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <17.1>." + } + ] + }, + { + "video_id": "X0BEF", + "question_id": "X0BEF", + "video_file": "Charades/videos/X0BEF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the other person is drinking a cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.3> to <27.1>." + } + ] + }, + { + "video_id": "X0BEF", + "question_id": "X0BEF", + "video_file": "Charades/videos/X0BEF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is throwing shoes at another person' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <17.1>." + } + ] + }, + { + "video_id": "TW12W", + "question_id": "TW12W", + "video_file": "Charades/videos/TW12W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is throwing shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <25.8>." + } + ] + }, + { + "video_id": "TW12W", + "question_id": "TW12W", + "video_file": "Charades/videos/TW12W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person sits on a bed eating food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <17.3>." + } + ] + }, + { + "video_id": "TW12W", + "question_id": "TW12W", + "video_file": "Charades/videos/TW12W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <4.2> to <17.3>." + } + ] + }, + { + "video_id": "TW12W", + "question_id": "TW12W", + "video_file": "Charades/videos/TW12W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'clothes .the person starts throwing clothes on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.2> to <30.7>." + } + ] + }, + { + "video_id": "IQ4GU", + "question_id": "IQ4GU", + "video_file": "Charades/videos/IQ4GU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking from a cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.4> to <18.5>." + } + ] + }, + { + "video_id": "IQ4GU", + "question_id": "IQ4GU", + "video_file": "Charades/videos/IQ4GU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they drink from the cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <18.5>." + } + ] + }, + { + "video_id": "K7PVL", + "question_id": "K7PVL", + "video_file": "Charades/videos/K7PVL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is putting away the groceries' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <20.4>." + } + ] + }, + { + "video_id": "3TV9E", + "question_id": "3TV9E", + "video_file": "Charades/videos/3TV9E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drink a cup of coffee' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <11.0>." + } + ] + }, + { + "video_id": "HLMQ8", + "question_id": "HLMQ8", + "video_file": "Charades/videos/HLMQ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person laughing about the absurdity of the situation'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "HLMQ8", + "question_id": "HLMQ8", + "video_file": "Charades/videos/HLMQ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opened the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.2> to <14.7>." + } + ] + }, + { + "video_id": "CF1OQ", + "question_id": "CF1OQ", + "video_file": "Charades/videos/CF1OQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs around a sofa' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "CF1OQ", + "question_id": "CF1OQ", + "video_file": "Charades/videos/CF1OQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs in sits down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "CF1OQ", + "question_id": "CF1OQ", + "video_file": "Charades/videos/CF1OQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts on shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <8.8>." + } + ] + }, + { + "video_id": "XQSXK", + "question_id": "XQSXK", + "video_file": "Charades/videos/XQSXK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes a tissue from the tissue box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <18.8>." + } + ] + }, + { + "video_id": "08EPO", + "question_id": "08EPO", + "video_file": "Charades/videos/08EPO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they also eat food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.1> to <15.8>." + } + ] + }, + { + "video_id": "08EPO", + "question_id": "08EPO", + "video_file": "Charades/videos/08EPO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person is eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <15.7>." + } + ] + }, + { + "video_id": "08EPO", + "question_id": "08EPO", + "video_file": "Charades/videos/08EPO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they eat a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <15.7>." + } + ] + }, + { + "video_id": "RYHHP", + "question_id": "RYHHP", + "video_file": "Charades/videos/RYHHP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person fixing their hair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <16.2>." + } + ] + }, + { + "video_id": "WJ1E7", + "question_id": "WJ1E7", + "video_file": "Charades/videos/WJ1E7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts the shoes down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.7> to <14.1>." + } + ] + }, + { + "video_id": "GKH4A", + "question_id": "GKH4A", + "video_file": "Charades/videos/GKH4A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes food out of a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.8> to <33.0>." + } + ] + }, + { + "video_id": "FYH25", + "question_id": "FYH25", + "video_file": "Charades/videos/FYH25.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person closes a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "FYH25", + "question_id": "FYH25", + "video_file": "Charades/videos/FYH25.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes some medicine' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <30.2> to <38.0>." + } + ] + }, + { + "video_id": "G4JHX", + "question_id": "G4JHX", + "video_file": "Charades/videos/G4JHX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts it a computer bag on the couch'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <24.5>." + } + ] + }, + { + "video_id": "G4JHX", + "question_id": "G4JHX", + "video_file": "Charades/videos/G4JHX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts it in a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <24.5>." + } + ] + }, + { + "video_id": "G4JHX", + "question_id": "G4JHX", + "video_file": "Charades/videos/G4JHX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person takes a laptop off of the couch' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "L5TI4", + "question_id": "L5TI4", + "video_file": "Charades/videos/L5TI4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is putting food on the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.2> to <32.0>." + } + ] + }, + { + "video_id": "84893", + "question_id": "84893", + "video_file": "Charades/videos/84893.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person throws shoes on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <20.9>." + } + ] + }, + { + "video_id": "84893", + "question_id": "84893", + "video_file": "Charades/videos/84893.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person looks out a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "84893", + "question_id": "84893", + "video_file": "Charades/videos/84893.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws their shoes to the ground'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <20.5>." + } + ] + }, + { + "video_id": "84893", + "question_id": "84893", + "video_file": "Charades/videos/84893.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws some shoes on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <20.9>." + } + ] + }, + { + "video_id": "84893", + "question_id": "84893", + "video_file": "Charades/videos/84893.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is looking out a window in a bedroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "P03FP", + "question_id": "P03FP", + "video_file": "Charades/videos/P03FP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats it'." + }, + { + "from": "gpt", + "value": "From <26.3> to <31.7>." + } + ] + }, + { + "video_id": "P03FP", + "question_id": "P03FP", + "video_file": "Charades/videos/P03FP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes some food from a dish' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.0> to <29.1>." + } + ] + }, + { + "video_id": "CCN3E", + "question_id": "CCN3E", + "video_file": "Charades/videos/CCN3E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person but stops to watch themselves in the mirror' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.6> to <13.7>." + } + ] + }, + { + "video_id": "Q34XB", + "question_id": "Q34XB", + "video_file": "Charades/videos/Q34XB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person undresses in front of a mirror'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.3>." + } + ] + }, + { + "video_id": "Q34XB", + "question_id": "Q34XB", + "video_file": "Charades/videos/Q34XB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes some clothes from a nearby table'." + }, + { + "from": "gpt", + "value": "From <9.2> to <15.3>." + } + ] + }, + { + "video_id": "YJ06D", + "question_id": "YJ06D", + "video_file": "Charades/videos/YJ06D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person turns the light on'." + }, + { + "from": "gpt", + "value": "From <13.9> to <19.5>." + } + ] + }, + { + "video_id": "YJ06D", + "question_id": "YJ06D", + "video_file": "Charades/videos/YJ06D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <19.5>." + } + ] + }, + { + "video_id": "XGL67", + "question_id": "XGL67", + "video_file": "Charades/videos/XGL67.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts clothes from the bed into the dresser' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "XGL67", + "question_id": "XGL67", + "video_file": "Charades/videos/XGL67.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting the clothes in the drawer'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "D2VU8", + "question_id": "D2VU8", + "video_file": "Charades/videos/D2VU8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts down the book'." + }, + { + "from": "gpt", + "value": "From <16.0> to <21.0>." + } + ] + }, + { + "video_id": "D2VU8", + "question_id": "D2VU8", + "video_file": "Charades/videos/D2VU8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the phone down'." + }, + { + "from": "gpt", + "value": "From <2.8> to <8.2>." + } + ] + }, + { + "video_id": "D2VU8", + "question_id": "D2VU8", + "video_file": "Charades/videos/D2VU8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is putting a camera on a shelf'." + }, + { + "from": "gpt", + "value": "From <2.8> to <8.2>." + } + ] + }, + { + "video_id": "NPNB3", + "question_id": "NPNB3", + "video_file": "Charades/videos/NPNB3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts on shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <28.3>." + } + ] + }, + { + "video_id": "NPNB3", + "question_id": "NPNB3", + "video_file": "Charades/videos/NPNB3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts on a pair of shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <28.3>." + } + ] + }, + { + "video_id": "NPNB3", + "question_id": "NPNB3", + "video_file": "Charades/videos/NPNB3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a standing person dresses' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.6>." + } + ] + }, + { + "video_id": "NPNB3", + "question_id": "NPNB3", + "video_file": "Charades/videos/NPNB3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is dressing in the bathroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.6>." + } + ] + }, + { + "video_id": "C5LCU", + "question_id": "C5LCU", + "video_file": "Charades/videos/C5LCU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <7.3>." + } + ] + }, + { + "video_id": "C5LCU", + "question_id": "C5LCU", + "video_file": "Charades/videos/C5LCU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they open their laptop'." + }, + { + "from": "gpt", + "value": "From <0.6> to <7.3>." + } + ] + }, + { + "video_id": "DMANO", + "question_id": "DMANO", + "video_file": "Charades/videos/DMANO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person dresses' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.9>." + } + ] + }, + { + "video_id": "DMANO", + "question_id": "DMANO", + "video_file": "Charades/videos/DMANO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person dressing by putting a jacket on'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.9>." + } + ] + }, + { + "video_id": "FL6DF", + "question_id": "FL6DF", + "video_file": "Charades/videos/FL6DF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <8.1>." + } + ] + }, + { + "video_id": "FL6DF", + "question_id": "FL6DF", + "video_file": "Charades/videos/FL6DF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throwing a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.3> to <15.9>." + } + ] + }, + { + "video_id": "FL6DF", + "question_id": "FL6DF", + "video_file": "Charades/videos/FL6DF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <8.1>." + } + ] + }, + { + "video_id": "FL6DF", + "question_id": "FL6DF", + "video_file": "Charades/videos/FL6DF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is opening a laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <8.1>." + } + ] + }, + { + "video_id": "FL6DF", + "question_id": "FL6DF", + "video_file": "Charades/videos/FL6DF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws a pillow on the floor'." + }, + { + "from": "gpt", + "value": "From <10.3> to <15.9>." + } + ] + }, + { + "video_id": "H603E", + "question_id": "H603E", + "video_file": "Charades/videos/H603E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person working on a laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <15.1>." + } + ] + }, + { + "video_id": "H603E", + "question_id": "H603E", + "video_file": "Charades/videos/H603E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put the paper on the ground'." + }, + { + "from": "gpt", + "value": "From <20.7> to <29.0>." + } + ] + }, + { + "video_id": "H603E", + "question_id": "H603E", + "video_file": "Charades/videos/H603E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts eating' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.5> to <33.0>." + } + ] + }, + { + "video_id": "H603E", + "question_id": "H603E", + "video_file": "Charades/videos/H603E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person working on their laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <15.1>." + } + ] + }, + { + "video_id": "H603E", + "question_id": "H603E", + "video_file": "Charades/videos/H603E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eat something' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.5> to <33.0>." + } + ] + }, + { + "video_id": "YPQ97", + "question_id": "YPQ97", + "video_file": "Charades/videos/YPQ97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the shoes on' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.3> to <6.1>." + } + ] + }, + { + "video_id": "5ZDNU", + "question_id": "5ZDNU", + "video_file": "Charades/videos/5ZDNU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person standing in their closet starts undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "QKIYJ", + "question_id": "QKIYJ", + "video_file": "Charades/videos/QKIYJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they fix their hair in the mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.7>." + } + ] + }, + { + "video_id": "QKIYJ", + "question_id": "QKIYJ", + "video_file": "Charades/videos/QKIYJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is in the entryway at their home laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <19.5>." + } + ] + }, + { + "video_id": "QKIYJ", + "question_id": "QKIYJ", + "video_file": "Charades/videos/QKIYJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person fixes their hair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.7>." + } + ] + }, + { + "video_id": "XQITP", + "question_id": "XQITP", + "video_file": "Charades/videos/XQITP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person sitting on a toilet puts shoes on' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.1> to <16.0>." + } + ] + }, + { + "video_id": "9UU4H", + "question_id": "9UU4H", + "video_file": "Charades/videos/9UU4H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person puts come clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.8>." + } + ] + }, + { + "video_id": "D2T7Z", + "question_id": "D2T7Z", + "video_file": "Charades/videos/D2T7Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws it on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.8> to <24.0>." + } + ] + }, + { + "video_id": "D2T7Z", + "question_id": "D2T7Z", + "video_file": "Charades/videos/D2T7Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws it onto the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.8> to <24.0>." + } + ] + }, + { + "video_id": "5U92L", + "question_id": "5U92L", + "video_file": "Charades/videos/5U92L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <12.0>." + } + ] + }, + { + "video_id": "5U92L", + "question_id": "5U92L", + "video_file": "Charades/videos/5U92L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person undresses'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <12.0>." + } + ] + }, + { + "video_id": "5U92L", + "question_id": "5U92L", + "video_file": "Charades/videos/5U92L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws clothes on a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.2> to <12.7>." + } + ] + }, + { + "video_id": "EOD1J", + "question_id": "EOD1J", + "video_file": "Charades/videos/EOD1J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put start to put the shoes on'." + }, + { + "from": "gpt", + "value": "From <20.4> to <25.4>." + } + ] + }, + { + "video_id": "ITODQ", + "question_id": "ITODQ", + "video_file": "Charades/videos/ITODQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins laughing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "ITODQ", + "question_id": "ITODQ", + "video_file": "Charades/videos/ITODQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person is laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "3EHAJ", + "question_id": "3EHAJ", + "video_file": "Charades/videos/3EHAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person laughing at the homework that needs to be done'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "83654", + "question_id": "83654", + "video_file": "Charades/videos/83654.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person puts a sandwich down onto a desk' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.1>." + } + ] + }, + { + "video_id": "83654", + "question_id": "83654", + "video_file": "Charades/videos/83654.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.7> to <27.3>." + } + ] + }, + { + "video_id": "E0IME", + "question_id": "E0IME", + "video_file": "Charades/videos/E0IME.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person runs away'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.7> to <35.0>." + } + ] + }, + { + "video_id": "E0IME", + "question_id": "E0IME", + "video_file": "Charades/videos/E0IME.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person runs towards the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.7> to <35.0>." + } + ] + }, + { + "video_id": "QAJQC", + "question_id": "QAJQC", + "video_file": "Charades/videos/QAJQC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts their laptop in a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.7> to <13.1>." + } + ] + }, + { + "video_id": "QAJQC", + "question_id": "QAJQC", + "video_file": "Charades/videos/QAJQC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens up a cabinet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.6> to <10.3>." + } + ] + }, + { + "video_id": "QAJQC", + "question_id": "QAJQC", + "video_file": "Charades/videos/QAJQC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns off the light'." + }, + { + "from": "gpt", + "value": "From <2.0> to <7.0>." + } + ] + }, + { + "video_id": "QAJQC", + "question_id": "QAJQC", + "video_file": "Charades/videos/QAJQC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts a laptop on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.7> to <13.1>." + } + ] + }, + { + "video_id": "QAJQC", + "question_id": "QAJQC", + "video_file": "Charades/videos/QAJQC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns off light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <7.0>." + } + ] + }, + { + "video_id": "QAJQC", + "question_id": "QAJQC", + "video_file": "Charades/videos/QAJQC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person turns on the light'." + }, + { + "from": "gpt", + "value": "From <2.0> to <7.0>." + } + ] + }, + { + "video_id": "QAJQC", + "question_id": "QAJQC", + "video_file": "Charades/videos/QAJQC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns on a light'." + }, + { + "from": "gpt", + "value": "From <2.0> to <7.0>." + } + ] + }, + { + "video_id": "QAJQC", + "question_id": "QAJQC", + "video_file": "Charades/videos/QAJQC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the doors'." + }, + { + "from": "gpt", + "value": "From <24.7> to <30.0>." + } + ] + }, + { + "video_id": "CR5G1", + "question_id": "CR5G1", + "video_file": "Charades/videos/CR5G1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person sits on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.9> to <26.9>." + } + ] + }, + { + "video_id": "DEZJ5", + "question_id": "DEZJ5", + "video_file": "Charades/videos/DEZJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a closet door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.2> to <20.0>." + } + ] + }, + { + "video_id": "DEZJ5", + "question_id": "DEZJ5", + "video_file": "Charades/videos/DEZJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens the door which is for a closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.2> to <20.0>." + } + ] + }, + { + "video_id": "DEZJ5", + "question_id": "DEZJ5", + "video_file": "Charades/videos/DEZJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person reacts by laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.0> to <16.0>." + } + ] + }, + { + "video_id": "DEZJ5", + "question_id": "DEZJ5", + "video_file": "Charades/videos/DEZJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is closing the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.2> to <20.0>." + } + ] + }, + { + "video_id": "DEZJ5", + "question_id": "DEZJ5", + "video_file": "Charades/videos/DEZJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person is opening door to get box of shoes'." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.8>." + } + ] + }, + { + "video_id": "DEZJ5", + "question_id": "DEZJ5", + "video_file": "Charades/videos/DEZJ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <16.0>." + } + ] + }, + { + "video_id": "UB00R", + "question_id": "UB00R", + "video_file": "Charades/videos/UB00R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person stand up' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.0> to <27.0>." + } + ] + }, + { + "video_id": "VJ3B5", + "question_id": "VJ3B5", + "video_file": "Charades/videos/VJ3B5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws it to the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <10.2>." + } + ] + }, + { + "video_id": "VJ3B5", + "question_id": "VJ3B5", + "video_file": "Charades/videos/VJ3B5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes a box from the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.2> to <10.1>." + } + ] + }, + { + "video_id": "5AYUN", + "question_id": "5AYUN", + "video_file": "Charades/videos/5AYUN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'so person gets a chair to stand on' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.6> to <39.1>." + } + ] + }, + { + "video_id": "U6GKR", + "question_id": "U6GKR", + "video_file": "Charades/videos/U6GKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinks from the cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.6> to <18.6>." + } + ] + }, + { + "video_id": "U6GKR", + "question_id": "U6GKR", + "video_file": "Charades/videos/U6GKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinks iced coffee from the glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <18.6>." + } + ] + }, + { + "video_id": "KS29V", + "question_id": "KS29V", + "video_file": "Charades/videos/KS29V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is holding a pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "KS29V", + "question_id": "KS29V", + "video_file": "Charades/videos/KS29V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a pillow'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "4L6GZ", + "question_id": "4L6GZ", + "video_file": "Charades/videos/4L6GZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening a window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <30.6>." + } + ] + }, + { + "video_id": "JI705", + "question_id": "JI705", + "video_file": "Charades/videos/JI705.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a bag of food from the nearby shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.6> to <26.6>." + } + ] + }, + { + "video_id": "PGGRZ", + "question_id": "PGGRZ", + "video_file": "Charades/videos/PGGRZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <5.5>." + } + ] + }, + { + "video_id": "PGGRZ", + "question_id": "PGGRZ", + "video_file": "Charades/videos/PGGRZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is throwing a pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <5.5>." + } + ] + }, + { + "video_id": "57QKQ", + "question_id": "57QKQ", + "video_file": "Charades/videos/57QKQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person opens a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.3> to <10.9>." + } + ] + }, + { + "video_id": "57QKQ", + "question_id": "57QKQ", + "video_file": "Charades/videos/57QKQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws the blanket toward the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <16.3>." + } + ] + }, + { + "video_id": "57QKQ", + "question_id": "57QKQ", + "video_file": "Charades/videos/57QKQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person looks out the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.0> to <12.5>." + } + ] + }, + { + "video_id": "Z9IF5", + "question_id": "Z9IF5", + "video_file": "Charades/videos/Z9IF5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <14.1>." + } + ] + }, + { + "video_id": "I5MCA", + "question_id": "I5MCA", + "video_file": "Charades/videos/I5MCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person washes their hands' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.2> to <17.2>." + } + ] + }, + { + "video_id": "I5MCA", + "question_id": "I5MCA", + "video_file": "Charades/videos/I5MCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'that person puts the sandwich down'." + }, + { + "from": "gpt", + "value": "From <5.3> to <10.8>." + } + ] + }, + { + "video_id": "5VDT6", + "question_id": "5VDT6", + "video_file": "Charades/videos/5VDT6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws clothes on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.3> to <23.1>." + } + ] + }, + { + "video_id": "UPWTC", + "question_id": "UPWTC", + "video_file": "Charades/videos/UPWTC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting at the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "UPWTC", + "question_id": "UPWTC", + "video_file": "Charades/videos/UPWTC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they stand up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <6.0>." + } + ] + }, + { + "video_id": "UPWTC", + "question_id": "UPWTC", + "video_file": "Charades/videos/UPWTC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person decide to put on some shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.0> to <26.9>." + } + ] + }, + { + "video_id": "UPWTC", + "question_id": "UPWTC", + "video_file": "Charades/videos/UPWTC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts on shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.0> to <26.9>." + } + ] + }, + { + "video_id": "9ONDD", + "question_id": "9ONDD", + "video_file": "Charades/videos/9ONDD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is cooking on a stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.1>." + } + ] + }, + { + "video_id": "9ONDD", + "question_id": "9ONDD", + "video_file": "Charades/videos/9ONDD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open the refrigerator to grab eggs' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.5> to <20.5>." + } + ] + }, + { + "video_id": "9ONDD", + "question_id": "9ONDD", + "video_file": "Charades/videos/9ONDD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is closing the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.8> to <24.5>." + } + ] + }, + { + "video_id": "9ONDD", + "question_id": "9ONDD", + "video_file": "Charades/videos/9ONDD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking some food on the stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.1>." + } + ] + }, + { + "video_id": "8EYST", + "question_id": "8EYST", + "video_file": "Charades/videos/8EYST.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening a book on the stairs' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.0> to <32.0>." + } + ] + }, + { + "video_id": "SZYJO", + "question_id": "SZYJO", + "video_file": "Charades/videos/SZYJO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts a few bowls on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.1> to <32.0>." + } + ] + }, + { + "video_id": "SZYJO", + "question_id": "SZYJO", + "video_file": "Charades/videos/SZYJO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person is sneezing in front of the laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <13.1>." + } + ] + }, + { + "video_id": "IOMVF", + "question_id": "IOMVF", + "video_file": "Charades/videos/IOMVF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person walks over to look out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <25.0>." + } + ] + }, + { + "video_id": "0C5IQ", + "question_id": "0C5IQ", + "video_file": "Charades/videos/0C5IQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks in through the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "I562C", + "question_id": "I562C", + "video_file": "Charades/videos/I562C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eat a small plate of food'." + }, + { + "from": "gpt", + "value": "From <15.2> to <25.7>." + } + ] + }, + { + "video_id": "I562C", + "question_id": "I562C", + "video_file": "Charades/videos/I562C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens the garage door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.6> to <5.7>." + } + ] + }, + { + "video_id": "I562C", + "question_id": "I562C", + "video_file": "Charades/videos/I562C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits down to eat'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <25.7>." + } + ] + }, + { + "video_id": "I562C", + "question_id": "I562C", + "video_file": "Charades/videos/I562C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a door than walks to a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.6> to <5.7>." + } + ] + }, + { + "video_id": "69EK6", + "question_id": "69EK6", + "video_file": "Charades/videos/69EK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <30.8> to <37.0>." + } + ] + }, + { + "video_id": "69EK6", + "question_id": "69EK6", + "video_file": "Charades/videos/69EK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put their phone on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.0> to <34.4>." + } + ] + }, + { + "video_id": "69EK6", + "question_id": "69EK6", + "video_file": "Charades/videos/69EK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts it down on the table at the bottom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.5> to <33.9>." + } + ] + }, + { + "video_id": "X75PJ", + "question_id": "X75PJ", + "video_file": "Charades/videos/X75PJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is running in place'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "X75PJ", + "question_id": "X75PJ", + "video_file": "Charades/videos/X75PJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person finally turning off a light in a hallway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.8> to <29.0>." + } + ] + }, + { + "video_id": "X75PJ", + "question_id": "X75PJ", + "video_file": "Charades/videos/X75PJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person running in place' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "X75PJ", + "question_id": "X75PJ", + "video_file": "Charades/videos/X75PJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sneezing'." + }, + { + "from": "gpt", + "value": "From <3.0> to <11.7>." + } + ] + }, + { + "video_id": "X75PJ", + "question_id": "X75PJ", + "video_file": "Charades/videos/X75PJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turn off the light'." + }, + { + "from": "gpt", + "value": "From <24.8> to <29.0>." + } + ] + }, + { + "video_id": "974TA", + "question_id": "974TA", + "video_file": "Charades/videos/974TA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.0> to <32.0>." + } + ] + }, + { + "video_id": "CTCO8", + "question_id": "CTCO8", + "video_file": "Charades/videos/CTCO8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person is throwing clothes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <23.6>." + } + ] + }, + { + "video_id": "CTCO8", + "question_id": "CTCO8", + "video_file": "Charades/videos/CTCO8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person is watching a laptop'." + }, + { + "from": "gpt", + "value": "From <27.8> to <31.0>." + } + ] + }, + { + "video_id": "AKKWU", + "question_id": "AKKWU", + "video_file": "Charades/videos/AKKWU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the cabinet'." + }, + { + "from": "gpt", + "value": "From <38.0> to <13.0>." + } + ] + }, + { + "video_id": "AKKWU", + "question_id": "AKKWU", + "video_file": "Charades/videos/AKKWU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person has opened a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <38.0> to <13.0>." + } + ] + }, + { + "video_id": "YBXCT", + "question_id": "YBXCT", + "video_file": "Charades/videos/YBXCT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open the closet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <8.4>." + } + ] + }, + { + "video_id": "YBXCT", + "question_id": "YBXCT", + "video_file": "Charades/videos/YBXCT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is near a closet holding a towel'." + }, + { + "from": "gpt", + "value": "From <6.7> to <15.0>." + } + ] + }, + { + "video_id": "OXEMQ", + "question_id": "OXEMQ", + "video_file": "Charades/videos/OXEMQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person undresses in the entryway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "OXEMQ", + "question_id": "OXEMQ", + "video_file": "Charades/videos/OXEMQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins dressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.0> to <30.0>." + } + ] + }, + { + "video_id": "OXEMQ", + "question_id": "OXEMQ", + "video_file": "Charades/videos/OXEMQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is dressing - takes off' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.0> to <30.0>." + } + ] + }, + { + "video_id": "G7GVD", + "question_id": "G7GVD", + "video_file": "Charades/videos/G7GVD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts their foot up on the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <9.7>." + } + ] + }, + { + "video_id": "G7GVD", + "question_id": "G7GVD", + "video_file": "Charades/videos/G7GVD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person was working on a laptop on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.7> to <23.9>." + } + ] + }, + { + "video_id": "G7GVD", + "question_id": "G7GVD", + "video_file": "Charades/videos/G7GVD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes off a shoe' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.9> to <27.7>." + } + ] + }, + { + "video_id": "Y5NDR", + "question_id": "Y5NDR", + "video_file": "Charades/videos/Y5NDR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person the begin running around undressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <12.2>." + } + ] + }, + { + "video_id": "Y5NDR", + "question_id": "Y5NDR", + "video_file": "Charades/videos/Y5NDR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.0> to <6.5>." + } + ] + }, + { + "video_id": "8X7D6", + "question_id": "8X7D6", + "video_file": "Charades/videos/8X7D6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person closes a laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.9> to <28.0>." + } + ] + }, + { + "video_id": "8X7D6", + "question_id": "8X7D6", + "video_file": "Charades/videos/8X7D6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person watches the television' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.2> to <27.0>." + } + ] + }, + { + "video_id": "8X7D6", + "question_id": "8X7D6", + "video_file": "Charades/videos/8X7D6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.9> to <28.0>." + } + ] + }, + { + "video_id": "YUEB9", + "question_id": "YUEB9", + "video_file": "Charades/videos/YUEB9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person get up to look out a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.0> to <29.4>." + } + ] + }, + { + "video_id": "YUEB9", + "question_id": "YUEB9", + "video_file": "Charades/videos/YUEB9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they start sneezing'." + }, + { + "from": "gpt", + "value": "From <11.6> to <21.9>." + } + ] + }, + { + "video_id": "YUEB9", + "question_id": "YUEB9", + "video_file": "Charades/videos/YUEB9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looks out the window'." + }, + { + "from": "gpt", + "value": "From <21.0> to <29.4>." + } + ] + }, + { + "video_id": "JTM7L", + "question_id": "JTM7L", + "video_file": "Charades/videos/JTM7L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person #1 opens the front door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.9> to <13.3>." + } + ] + }, + { + "video_id": "CQPZU", + "question_id": "CQPZU", + "video_file": "Charades/videos/CQPZU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person runs out' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.0> to <24.0>." + } + ] + }, + { + "video_id": "CQPZU", + "question_id": "CQPZU", + "video_file": "Charades/videos/CQPZU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person run through'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <24.0>." + } + ] + }, + { + "video_id": "CQPZU", + "question_id": "CQPZU", + "video_file": "Charades/videos/CQPZU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.9> to <21.1>." + } + ] + }, + { + "video_id": "CQPZU", + "question_id": "CQPZU", + "video_file": "Charades/videos/CQPZU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person runs through'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <24.0>." + } + ] + }, + { + "video_id": "CQPZU", + "question_id": "CQPZU", + "video_file": "Charades/videos/CQPZU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they open the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.9> to <21.1>." + } + ] + }, + { + "video_id": "YJ888", + "question_id": "YJ888", + "video_file": "Charades/videos/YJ888.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person while seated on a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <15.8>." + } + ] + }, + { + "video_id": "YJ888", + "question_id": "YJ888", + "video_file": "Charades/videos/YJ888.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take out something to eat they look around'." + }, + { + "from": "gpt", + "value": "From <10.4> to <15.8>." + } + ] + }, + { + "video_id": "YJ888", + "question_id": "YJ888", + "video_file": "Charades/videos/YJ888.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.9>." + } + ] + }, + { + "video_id": "MUGVM", + "question_id": "MUGVM", + "video_file": "Charades/videos/MUGVM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is tidying up a table with a towel'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "MUGVM", + "question_id": "MUGVM", + "video_file": "Charades/videos/MUGVM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they fix their hair in a mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.0> to <26.1>." + } + ] + }, + { + "video_id": "MUGVM", + "question_id": "MUGVM", + "video_file": "Charades/videos/MUGVM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a drink of a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <25.3>." + } + ] + }, + { + "video_id": "MUGVM", + "question_id": "MUGVM", + "video_file": "Charades/videos/MUGVM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looking at a mirror to fix their hair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.0> to <26.1>." + } + ] + }, + { + "video_id": "W9XY8", + "question_id": "W9XY8", + "video_file": "Charades/videos/W9XY8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <8.4>." + } + ] + }, + { + "video_id": "2IC8O", + "question_id": "2IC8O", + "video_file": "Charades/videos/2IC8O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.1>." + } + ] + }, + { + "video_id": "2IC8O", + "question_id": "2IC8O", + "video_file": "Charades/videos/2IC8O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting in a chair watching television' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.8> to <25.0>." + } + ] + }, + { + "video_id": "OADOE", + "question_id": "OADOE", + "video_file": "Charades/videos/OADOE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the door behind him' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.1> to <25.0>." + } + ] + }, + { + "video_id": "OADOE", + "question_id": "OADOE", + "video_file": "Charades/videos/OADOE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a man is standing at a stove cooking something' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "FRLW2", + "question_id": "FRLW2", + "video_file": "Charades/videos/FRLW2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts away a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.6> to <15.4>." + } + ] + }, + { + "video_id": "25QPB", + "question_id": "25QPB", + "video_file": "Charades/videos/25QPB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a closet door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <25.0>." + } + ] + }, + { + "video_id": "25QPB", + "question_id": "25QPB", + "video_file": "Charades/videos/25QPB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person start smiling at what is inside' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <29.0>." + } + ] + }, + { + "video_id": "AHZWF", + "question_id": "AHZWF", + "video_file": "Charades/videos/AHZWF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they stand up' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.1> to <12.7>." + } + ] + }, + { + "video_id": "24KQ5", + "question_id": "24KQ5", + "video_file": "Charades/videos/24KQ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is watching television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "EMMYX", + "question_id": "EMMYX", + "video_file": "Charades/videos/EMMYX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put a broom at the entrance'." + }, + { + "from": "gpt", + "value": "From <23.6> to <32.0>." + } + ] + }, + { + "video_id": "EMMYX", + "question_id": "EMMYX", + "video_file": "Charades/videos/EMMYX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person smiling as they are closing the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.2> to <27.5>." + } + ] + }, + { + "video_id": "H3E1E", + "question_id": "H3E1E", + "video_file": "Charades/videos/H3E1E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking a cup of coffee' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "ENQUT", + "question_id": "ENQUT", + "video_file": "Charades/videos/ENQUT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes off their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.8> to <38.4>." + } + ] + }, + { + "video_id": "ENQUT", + "question_id": "ENQUT", + "video_file": "Charades/videos/ENQUT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is washing their hands'." + }, + { + "from": "gpt", + "value": "From <5.0> to <11.2>." + } + ] + }, + { + "video_id": "ENQUT", + "question_id": "ENQUT", + "video_file": "Charades/videos/ENQUT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take their shoes off'." + }, + { + "from": "gpt", + "value": "From <26.8> to <38.4>." + } + ] + }, + { + "video_id": "ENQUT", + "question_id": "ENQUT", + "video_file": "Charades/videos/ENQUT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws them on the floor underneath the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <36.8> to <41.0>." + } + ] + }, + { + "video_id": "10ND1", + "question_id": "10ND1", + "video_file": "Charades/videos/10ND1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is cooking something on the stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <30.1> to <35.0>." + } + ] + }, + { + "video_id": "10ND1", + "question_id": "10ND1", + "video_file": "Charades/videos/10ND1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the food on a plate'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.2> to <34.9>." + } + ] + }, + { + "video_id": "10ND1", + "question_id": "10ND1", + "video_file": "Charades/videos/10ND1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts them into the skillet/pot/pan to start cooking' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.1> to <35.0>." + } + ] + }, + { + "video_id": "10ND1", + "question_id": "10ND1", + "video_file": "Charades/videos/10ND1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put it on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "5XEOB", + "question_id": "5XEOB", + "video_file": "Charades/videos/5XEOB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is smiling'." + }, + { + "from": "gpt", + "value": "From <19.3> to <25.7>." + } + ] + }, + { + "video_id": "5XEOB", + "question_id": "5XEOB", + "video_file": "Charades/videos/5XEOB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closing a door in a pantry' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.5> to <17.2>." + } + ] + }, + { + "video_id": "5XEOB", + "question_id": "5XEOB", + "video_file": "Charades/videos/5XEOB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they close the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <17.2>." + } + ] + }, + { + "video_id": "NL7BY", + "question_id": "NL7BY", + "video_file": "Charades/videos/NL7BY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person is sitting down looking through a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "4G00A", + "question_id": "4G00A", + "video_file": "Charades/videos/4G00A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pouring some water into a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "4G00A", + "question_id": "4G00A", + "video_file": "Charades/videos/4G00A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is pouring themselves a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "573OJ", + "question_id": "573OJ", + "video_file": "Charades/videos/573OJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person turns off a light'." + }, + { + "from": "gpt", + "value": "From <26.6> to <31.0>." + } + ] + }, + { + "video_id": "573OJ", + "question_id": "573OJ", + "video_file": "Charades/videos/573OJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.6> to <31.0>." + } + ] + }, + { + "video_id": "L1Z8A", + "question_id": "L1Z8A", + "video_file": "Charades/videos/L1Z8A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person start holding up a mirror to their face' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.4> to <51.6>." + } + ] + }, + { + "video_id": "L1Z8A", + "question_id": "L1Z8A", + "video_file": "Charades/videos/L1Z8A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting on a couch'." + }, + { + "from": "gpt", + "value": "From <0.4> to <23.3>." + } + ] + }, + { + "video_id": "WN38A", + "question_id": "WN38A", + "video_file": "Charades/videos/WN38A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put the detergent box on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.6> to <18.7>." + } + ] + }, + { + "video_id": "3G4PN", + "question_id": "3G4PN", + "video_file": "Charades/videos/3G4PN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a smiling person walks into their bedroom'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.1> to <29.2>." + } + ] + }, + { + "video_id": "T1XP2", + "question_id": "T1XP2", + "video_file": "Charades/videos/T1XP2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person reading a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.2> to <9.6>." + } + ] + }, + { + "video_id": "T1XP2", + "question_id": "T1XP2", + "video_file": "Charades/videos/T1XP2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <31.1>." + } + ] + }, + { + "video_id": "3CIVD", + "question_id": "3CIVD", + "video_file": "Charades/videos/3CIVD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating some food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.8> to <21.1>." + } + ] + }, + { + "video_id": "7HV9Q", + "question_id": "7HV9Q", + "video_file": "Charades/videos/7HV9Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person is opening the laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.8>." + } + ] + }, + { + "video_id": "7HV9Q", + "question_id": "7HV9Q", + "video_file": "Charades/videos/7HV9Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a laptop to work on it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.8>." + } + ] + }, + { + "video_id": "G64VQ", + "question_id": "G64VQ", + "video_file": "Charades/videos/G64VQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person holding a phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "PURYC", + "question_id": "PURYC", + "video_file": "Charades/videos/PURYC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opened the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.5>." + } + ] + }, + { + "video_id": "FDVBR", + "question_id": "FDVBR", + "video_file": "Charades/videos/FDVBR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the picture on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.2> to <27.0>." + } + ] + }, + { + "video_id": "MNS8T", + "question_id": "MNS8T", + "video_file": "Charades/videos/MNS8T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person laughs at the food cooking on the stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.2>." + } + ] + }, + { + "video_id": "MNS8T", + "question_id": "MNS8T", + "video_file": "Charades/videos/MNS8T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is cooking on a stove using a pot'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.2>." + } + ] + }, + { + "video_id": "MNS8T", + "question_id": "MNS8T", + "video_file": "Charades/videos/MNS8T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is cooking some food on a stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.2>." + } + ] + }, + { + "video_id": "S9P0W", + "question_id": "S9P0W", + "video_file": "Charades/videos/S9P0W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turning off the light'." + }, + { + "from": "gpt", + "value": "From <27.1> to <32.6>." + } + ] + }, + { + "video_id": "S9P0W", + "question_id": "S9P0W", + "video_file": "Charades/videos/S9P0W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns off the light'." + }, + { + "from": "gpt", + "value": "From <27.1> to <32.6>." + } + ] + }, + { + "video_id": "CRNL5", + "question_id": "CRNL5", + "video_file": "Charades/videos/CRNL5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting them on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.4> to <30.2>." + } + ] + }, + { + "video_id": "QXBTG", + "question_id": "QXBTG", + "video_file": "Charades/videos/QXBTG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take a towel from the box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <9.8>." + } + ] + }, + { + "video_id": "QXBTG", + "question_id": "QXBTG", + "video_file": "Charades/videos/QXBTG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opened a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.9>." + } + ] + }, + { + "video_id": "QXBTG", + "question_id": "QXBTG", + "video_file": "Charades/videos/QXBTG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person start throwing shoes at the desk' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.4> to <21.0>." + } + ] + }, + { + "video_id": "X1R6D", + "question_id": "X1R6D", + "video_file": "Charades/videos/X1R6D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <8.9>." + } + ] + }, + { + "video_id": "X1R6D", + "question_id": "X1R6D", + "video_file": "Charades/videos/X1R6D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person closes the door behind them'." + }, + { + "from": "gpt", + "value": "From <2.2> to <8.9>." + } + ] + }, + { + "video_id": "RKIC6", + "question_id": "RKIC6", + "video_file": "Charades/videos/RKIC6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person runs down the stairs'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "VF3Y0", + "question_id": "VF3Y0", + "video_file": "Charades/videos/VF3Y0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts putting food from the pot onto a plate' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.0> to <10.0>." + } + ] + }, + { + "video_id": "GOWDG", + "question_id": "GOWDG", + "video_file": "Charades/videos/GOWDG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the broom away' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.6> to <20.4>." + } + ] + }, + { + "video_id": "GOWDG", + "question_id": "GOWDG", + "video_file": "Charades/videos/GOWDG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts a broom next to a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <9.6>." + } + ] + }, + { + "video_id": "GOWDG", + "question_id": "GOWDG", + "video_file": "Charades/videos/GOWDG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is opening a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "GOWDG", + "question_id": "GOWDG", + "video_file": "Charades/videos/GOWDG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person opens the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "FKAGD", + "question_id": "FKAGD", + "video_file": "Charades/videos/FKAGD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looking at a book / magazine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.8> to <25.2>." + } + ] + }, + { + "video_id": "FKAGD", + "question_id": "FKAGD", + "video_file": "Charades/videos/FKAGD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating food with a book open' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.8> to <21.0>." + } + ] + }, + { + "video_id": "FKAGD", + "question_id": "FKAGD", + "video_file": "Charades/videos/FKAGD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person the other is eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "5657O", + "question_id": "5657O", + "video_file": "Charades/videos/5657O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is undressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.4>." + } + ] + }, + { + "video_id": "5657O", + "question_id": "5657O", + "video_file": "Charades/videos/5657O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take a picture of them self' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.9> to <32.0>." + } + ] + }, + { + "video_id": "5657O", + "question_id": "5657O", + "video_file": "Charades/videos/5657O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.2> to <18.2>." + } + ] + }, + { + "video_id": "8BNUT", + "question_id": "8BNUT", + "video_file": "Charades/videos/8BNUT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins pouring soda into the glass on the desk' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.2> to <30.9>." + } + ] + }, + { + "video_id": "8BNUT", + "question_id": "8BNUT", + "video_file": "Charades/videos/8BNUT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pours some water into a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.2> to <30.9>." + } + ] + }, + { + "video_id": "8BNUT", + "question_id": "8BNUT", + "video_file": "Charades/videos/8BNUT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turning on a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <17.3>." + } + ] + }, + { + "video_id": "8BNUT", + "question_id": "8BNUT", + "video_file": "Charades/videos/8BNUT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns light on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <17.3>." + } + ] + }, + { + "video_id": "8BNUT", + "question_id": "8BNUT", + "video_file": "Charades/videos/8BNUT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person stands next to the light switch for a moment'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <17.3>." + } + ] + }, + { + "video_id": "8BNUT", + "question_id": "8BNUT", + "video_file": "Charades/videos/8BNUT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turning the lights on' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.3> to <17.3>." + } + ] + }, + { + "video_id": "4CAHN", + "question_id": "4CAHN", + "video_file": "Charades/videos/4CAHN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person close the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.5> to <24.8>." + } + ] + }, + { + "video_id": "4CAHN", + "question_id": "4CAHN", + "video_file": "Charades/videos/4CAHN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take a cup from the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.9> to <23.9>." + } + ] + }, + { + "video_id": "4CAHN", + "question_id": "4CAHN", + "video_file": "Charades/videos/4CAHN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take out a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.9> to <23.9>." + } + ] + }, + { + "video_id": "4CAHN", + "question_id": "4CAHN", + "video_file": "Charades/videos/4CAHN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person open the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.4> to <23.4>." + } + ] + }, + { + "video_id": "JIS2Q", + "question_id": "JIS2Q", + "video_file": "Charades/videos/JIS2Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opened the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <12.0>." + } + ] + }, + { + "video_id": "JIS2Q", + "question_id": "JIS2Q", + "video_file": "Charades/videos/JIS2Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the bag down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <7.4>." + } + ] + }, + { + "video_id": "TYHA8", + "question_id": "TYHA8", + "video_file": "Charades/videos/TYHA8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts down the phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <7.9>." + } + ] + }, + { + "video_id": "LCEM0", + "question_id": "LCEM0", + "video_file": "Charades/videos/LCEM0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they eat a grape' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.2> to <12.6>." + } + ] + }, + { + "video_id": "LCEM0", + "question_id": "LCEM0", + "video_file": "Charades/videos/LCEM0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put the plate on the table,'." + }, + { + "from": "gpt", + "value": "From <23.8> to <29.2>." + } + ] + }, + { + "video_id": "LCEM0", + "question_id": "LCEM0", + "video_file": "Charades/videos/LCEM0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating some groceries that the just bought' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.2> to <12.6>." + } + ] + }, + { + "video_id": "LKYCT", + "question_id": "LKYCT", + "video_file": "Charades/videos/LKYCT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.3> to <24.8>." + } + ] + }, + { + "video_id": "LKYCT", + "question_id": "LKYCT", + "video_file": "Charades/videos/LKYCT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drink a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.6> to <31.0>." + } + ] + }, + { + "video_id": "LKYCT", + "question_id": "LKYCT", + "video_file": "Charades/videos/LKYCT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking a glass of milk'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.6> to <31.0>." + } + ] + }, + { + "video_id": "47D1Y", + "question_id": "47D1Y", + "video_file": "Charades/videos/47D1Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens the bathroom cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.3> to <6.3>." + } + ] + }, + { + "video_id": "3HF45", + "question_id": "3HF45", + "video_file": "Charades/videos/3HF45.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes clothes out of a dryer'." + }, + { + "from": "gpt", + "value": "From <3.4> to <9.9>." + } + ] + }, + { + "video_id": "35W7G", + "question_id": "35W7G", + "video_file": "Charades/videos/35W7G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.8>." + } + ] + }, + { + "video_id": "35W7G", + "question_id": "35W7G", + "video_file": "Charades/videos/35W7G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'sneezing many times'." + }, + { + "from": "gpt", + "value": "From <14.1> to <23.6>." + } + ] + }, + { + "video_id": "35W7G", + "question_id": "35W7G", + "video_file": "Charades/videos/35W7G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person stands by a door sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <23.6>." + } + ] + }, + { + "video_id": "35W7G", + "question_id": "35W7G", + "video_file": "Charades/videos/35W7G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person is sneezing by the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.1> to <23.6>." + } + ] + }, + { + "video_id": "KRRBE", + "question_id": "KRRBE", + "video_file": "Charades/videos/KRRBE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they stand up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <11.1>." + } + ] + }, + { + "video_id": "KRRBE", + "question_id": "KRRBE", + "video_file": "Charades/videos/KRRBE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walk through a doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.6> to <18.5>." + } + ] + }, + { + "video_id": "KRRBE", + "question_id": "KRRBE", + "video_file": "Charades/videos/KRRBE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting on a chair using a tablet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "KRRBE", + "question_id": "KRRBE", + "video_file": "Charades/videos/KRRBE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is looking at a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.3>." + } + ] + }, + { + "video_id": "KRRBE", + "question_id": "KRRBE", + "video_file": "Charades/videos/KRRBE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person looking at a picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.9>." + } + ] + }, + { + "video_id": "UKGO4", + "question_id": "UKGO4", + "video_file": "Charades/videos/UKGO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person cup they take them away'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "UKGO4", + "question_id": "UKGO4", + "video_file": "Charades/videos/UKGO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turn off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.8> to <31.1>." + } + ] + }, + { + "video_id": "UKGO4", + "question_id": "UKGO4", + "video_file": "Charades/videos/UKGO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turning off a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.8> to <31.1>." + } + ] + }, + { + "video_id": "UKGO4", + "question_id": "UKGO4", + "video_file": "Charades/videos/UKGO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is dressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <20.2>." + } + ] + }, + { + "video_id": "KD2H4", + "question_id": "KD2H4", + "video_file": "Charades/videos/KD2H4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is washing dishes in the sink' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "KD2H4", + "question_id": "KD2H4", + "video_file": "Charades/videos/KD2H4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open the oven door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.2> to <25.5>." + } + ] + }, + { + "video_id": "KD2H4", + "question_id": "KD2H4", + "video_file": "Charades/videos/KD2H4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the oven door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.2> to <25.5>." + } + ] + }, + { + "video_id": "KD2H4", + "question_id": "KD2H4", + "video_file": "Charades/videos/KD2H4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is opening the oven door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.2> to <25.5>." + } + ] + }, + { + "video_id": "KD2H4", + "question_id": "KD2H4", + "video_file": "Charades/videos/KD2H4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is washing dishes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.6>." + } + ] + }, + { + "video_id": "5G9SV", + "question_id": "5G9SV", + "video_file": "Charades/videos/5G9SV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts a picture on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <15.0>." + } + ] + }, + { + "video_id": "LPPFL", + "question_id": "LPPFL", + "video_file": "Charades/videos/LPPFL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a man is drinking a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.3> to <9.8>." + } + ] + }, + { + "video_id": "NIG0Z", + "question_id": "NIG0Z", + "video_file": "Charades/videos/NIG0Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'there is a person sitting on a sofa drinking coffee'." + }, + { + "from": "gpt", + "value": "From <0.4> to <5.0>." + } + ] + }, + { + "video_id": "RRNDQ", + "question_id": "RRNDQ", + "video_file": "Charades/videos/RRNDQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sitting in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "RRNDQ", + "question_id": "RRNDQ", + "video_file": "Charades/videos/RRNDQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking out of a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <14.4>." + } + ] + }, + { + "video_id": "63WHZ", + "question_id": "63WHZ", + "video_file": "Charades/videos/63WHZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sitting in front of a computer folding a towel'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <21.5>." + } + ] + }, + { + "video_id": "L2DR0", + "question_id": "L2DR0", + "video_file": "Charades/videos/L2DR0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws the pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <20.0>." + } + ] + }, + { + "video_id": "L2DR0", + "question_id": "L2DR0", + "video_file": "Charades/videos/L2DR0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person in a garage throws a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <20.0>." + } + ] + }, + { + "video_id": "L2DR0", + "question_id": "L2DR0", + "video_file": "Charades/videos/L2DR0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a vacuum' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <9.8>." + } + ] + }, + { + "video_id": "3D81G", + "question_id": "3D81G", + "video_file": "Charades/videos/3D81G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "S2L5Z", + "question_id": "S2L5Z", + "video_file": "Charades/videos/S2L5Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts a camera on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <11.3>." + } + ] + }, + { + "video_id": "S2L5Z", + "question_id": "S2L5Z", + "video_file": "Charades/videos/S2L5Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a stack of dishes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <18.4>." + } + ] + }, + { + "video_id": "UJO36", + "question_id": "UJO36", + "video_file": "Charades/videos/UJO36.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pours some into the glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <9.3>." + } + ] + }, + { + "video_id": "UJO36", + "question_id": "UJO36", + "video_file": "Charades/videos/UJO36.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the glass down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.0> to <16.7>." + } + ] + }, + { + "video_id": "UJO36", + "question_id": "UJO36", + "video_file": "Charades/videos/UJO36.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is pouring soda into a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.9> to <9.3>." + } + ] + }, + { + "video_id": "YMGGV", + "question_id": "YMGGV", + "video_file": "Charades/videos/YMGGV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sitting on couch turns on the tv' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "K1D2G", + "question_id": "K1D2G", + "video_file": "Charades/videos/K1D2G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a towel'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.4> to <19.8>." + } + ] + }, + { + "video_id": "K1D2G", + "question_id": "K1D2G", + "video_file": "Charades/videos/K1D2G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is running into a bedroom holding a towel' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.4> to <19.8>." + } + ] + }, + { + "video_id": "K1D2G", + "question_id": "K1D2G", + "video_file": "Charades/videos/K1D2G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks through a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "TYYZK", + "question_id": "TYYZK", + "video_file": "Charades/videos/TYYZK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding a laptop bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.5> to <42.0>." + } + ] + }, + { + "video_id": "I5LAC", + "question_id": "I5LAC", + "video_file": "Charades/videos/I5LAC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put the bag on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <7.7>." + } + ] + }, + { + "video_id": "I5LAC", + "question_id": "I5LAC", + "video_file": "Charades/videos/I5LAC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks into the basement holding a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "0DBQD", + "question_id": "0DBQD", + "video_file": "Charades/videos/0DBQD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person one is eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "0DBQD", + "question_id": "0DBQD", + "video_file": "Charades/videos/0DBQD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "0DBQD", + "question_id": "0DBQD", + "video_file": "Charades/videos/0DBQD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they watched television' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "0DBQD", + "question_id": "0DBQD", + "video_file": "Charades/videos/0DBQD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person was also drinking a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.6> to <17.8>." + } + ] + }, + { + "video_id": "0DBQD", + "question_id": "0DBQD", + "video_file": "Charades/videos/0DBQD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person watching television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "0DBQD", + "question_id": "0DBQD", + "video_file": "Charades/videos/0DBQD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person was eating a sandwich on the couch' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "NLM0B", + "question_id": "NLM0B", + "video_file": "Charades/videos/NLM0B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they close the box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.6> to <15.6>." + } + ] + }, + { + "video_id": "R5IQ5", + "question_id": "R5IQ5", + "video_file": "Charades/videos/R5IQ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person is smiling' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.6> to <27.3>." + } + ] + }, + { + "video_id": "3O6KD", + "question_id": "3O6KD", + "video_file": "Charades/videos/3O6KD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits down on a chair'." + }, + { + "from": "gpt", + "value": "From <16.7> to <31.0>." + } + ] + }, + { + "video_id": "3O6KD", + "question_id": "3O6KD", + "video_file": "Charades/videos/3O6KD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting on a bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.0>." + } + ] + }, + { + "video_id": "3O6KD", + "question_id": "3O6KD", + "video_file": "Charades/videos/3O6KD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person sits on the bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.0>." + } + ] + }, + { + "video_id": "Z3DBQ", + "question_id": "Z3DBQ", + "video_file": "Charades/videos/Z3DBQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws down the bags of groceries'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.1> to <22.8>." + } + ] + }, + { + "video_id": "Z3DBQ", + "question_id": "Z3DBQ", + "video_file": "Charades/videos/Z3DBQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a man opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.4> to <8.7>." + } + ] + }, + { + "video_id": "Z3DBQ", + "question_id": "Z3DBQ", + "video_file": "Charades/videos/Z3DBQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws the bags over the railing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.1> to <22.8>." + } + ] + }, + { + "video_id": "Z3DBQ", + "question_id": "Z3DBQ", + "video_file": "Charades/videos/Z3DBQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens door to apartment' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.4> to <8.7>." + } + ] + }, + { + "video_id": "BLBJY", + "question_id": "BLBJY", + "video_file": "Charades/videos/BLBJY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.0>." + } + ] + }, + { + "video_id": "TX020", + "question_id": "TX020", + "video_file": "Charades/videos/TX020.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs into the garage'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "TX020", + "question_id": "TX020", + "video_file": "Charades/videos/TX020.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs into a garage' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "TX020", + "question_id": "TX020", + "video_file": "Charades/videos/TX020.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs out of garage' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "TX020", + "question_id": "TX020", + "video_file": "Charades/videos/TX020.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person runs into garage' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "C6F8V", + "question_id": "C6F8V", + "video_file": "Charades/videos/C6F8V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is standing in front of a mirror laughing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <19.0>." + } + ] + }, + { + "video_id": "NH3LX", + "question_id": "NH3LX", + "video_file": "Charades/videos/NH3LX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.8> to <31.7>." + } + ] + }, + { + "video_id": "VS5IY", + "question_id": "VS5IY", + "video_file": "Charades/videos/VS5IY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts a glass of milk on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <7.5>." + } + ] + }, + { + "video_id": "VS5IY", + "question_id": "VS5IY", + "video_file": "Charades/videos/VS5IY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person getting dressed'." + }, + { + "from": "gpt", + "value": "From <14.1> to <26.0>." + } + ] + }, + { + "video_id": "VS5IY", + "question_id": "VS5IY", + "video_file": "Charades/videos/VS5IY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts getting dressed for work'." + }, + { + "from": "gpt", + "value": "From <14.1> to <26.0>." + } + ] + }, + { + "video_id": "B8DL6", + "question_id": "B8DL6", + "video_file": "Charades/videos/B8DL6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they tidy up the table'." + }, + { + "from": "gpt", + "value": "From <14.1> to <22.0>." + } + ] + }, + { + "video_id": "B8DL6", + "question_id": "B8DL6", + "video_file": "Charades/videos/B8DL6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts tidying up a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <22.0>." + } + ] + }, + { + "video_id": "B8DL6", + "question_id": "B8DL6", + "video_file": "Charades/videos/B8DL6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person begins tidying the area around the table'." + }, + { + "from": "gpt", + "value": "From <14.1> to <22.0>." + } + ] + }, + { + "video_id": "3U1SS", + "question_id": "3U1SS", + "video_file": "Charades/videos/3U1SS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person fixes their hair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.7>." + } + ] + }, + { + "video_id": "3U1SS", + "question_id": "3U1SS", + "video_file": "Charades/videos/3U1SS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'personlaying on floor gets up'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.7>." + } + ] + }, + { + "video_id": "O8S0J", + "question_id": "O8S0J", + "video_file": "Charades/videos/O8S0J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting at a desk as they eat' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.4>." + } + ] + }, + { + "video_id": "TOE9V", + "question_id": "TOE9V", + "video_file": "Charades/videos/TOE9V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person throws a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <5.7>." + } + ] + }, + { + "video_id": "TOE9V", + "question_id": "TOE9V", + "video_file": "Charades/videos/TOE9V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts a towel over top the book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.9> to <9.3>." + } + ] + }, + { + "video_id": "ORAT0", + "question_id": "ORAT0", + "video_file": "Charades/videos/ORAT0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.0> to <16.1>." + } + ] + }, + { + "video_id": "ORAT0", + "question_id": "ORAT0", + "video_file": "Charades/videos/ORAT0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits at the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <12.0>." + } + ] + }, + { + "video_id": "ORAT0", + "question_id": "ORAT0", + "video_file": "Charades/videos/ORAT0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats something'." + }, + { + "from": "gpt", + "value": "From <8.0> to <16.1>." + } + ] + }, + { + "video_id": "W1QG3", + "question_id": "W1QG3", + "video_file": "Charades/videos/W1QG3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a sneezing person is fixing a shelf in their bedroom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <17.6>." + } + ] + }, + { + "video_id": "ORPD5", + "question_id": "ORPD5", + "video_file": "Charades/videos/ORPD5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is tidying up some clothes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.1> to <30.0>." + } + ] + }, + { + "video_id": "ORPD5", + "question_id": "ORPD5", + "video_file": "Charades/videos/ORPD5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is putting towels on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.3> to <10.5>." + } + ] + }, + { + "video_id": "MWUIN", + "question_id": "MWUIN", + "video_file": "Charades/videos/MWUIN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts the pillow on a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.3> to <8.0>." + } + ] + }, + { + "video_id": "GSQX3", + "question_id": "GSQX3", + "video_file": "Charades/videos/GSQX3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they start sneezing a lot'." + }, + { + "from": "gpt", + "value": "From <9.0> to <14.6>." + } + ] + }, + { + "video_id": "HJ8U7", + "question_id": "HJ8U7", + "video_file": "Charades/videos/HJ8U7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts drinking a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.7> to <28.3>." + } + ] + }, + { + "video_id": "HJ8U7", + "question_id": "HJ8U7", + "video_file": "Charades/videos/HJ8U7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is dressing in the entryway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.4>." + } + ] + }, + { + "video_id": "UECH4", + "question_id": "UECH4", + "video_file": "Charades/videos/UECH4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sneezing into a mirror in a hallway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.8> to <31.0>." + } + ] + }, + { + "video_id": "01K8X", + "question_id": "01K8X", + "video_file": "Charades/videos/01K8X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.6> to <32.0>." + } + ] + }, + { + "video_id": "P68QV", + "question_id": "P68QV", + "video_file": "Charades/videos/P68QV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person smiling at it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "P68QV", + "question_id": "P68QV", + "video_file": "Charades/videos/P68QV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is seen smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "ET224", + "question_id": "ET224", + "video_file": "Charades/videos/ET224.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closing the door behind him' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <50.3> to <56.1>." + } + ] + }, + { + "video_id": "ET224", + "question_id": "ET224", + "video_file": "Charades/videos/ET224.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opened a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <48.2> to <54.0>." + } + ] + }, + { + "video_id": "BWKJB", + "question_id": "BWKJB", + "video_file": "Charades/videos/BWKJB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes a blanket from a chair'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.4>." + } + ] + }, + { + "video_id": "469ZJ", + "question_id": "469ZJ", + "video_file": "Charades/videos/469ZJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turning on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "469ZJ", + "question_id": "469ZJ", + "video_file": "Charades/videos/469ZJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person turns on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "1L7XE", + "question_id": "1L7XE", + "video_file": "Charades/videos/1L7XE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.2> to <4.9>." + } + ] + }, + { + "video_id": "1L7XE", + "question_id": "1L7XE", + "video_file": "Charades/videos/1L7XE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opening a cabinet'." + }, + { + "from": "gpt", + "value": "From <0.2> to <4.9>." + } + ] + }, + { + "video_id": "FXBGY", + "question_id": "FXBGY", + "video_file": "Charades/videos/FXBGY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person laughs each time they sneeze'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <6.8>." + } + ] + }, + { + "video_id": "FXBGY", + "question_id": "FXBGY", + "video_file": "Charades/videos/FXBGY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person laughs'." + }, + { + "from": "gpt", + "value": "From <0.5> to <6.8>." + } + ] + }, + { + "video_id": "TZF0X", + "question_id": "TZF0X", + "video_file": "Charades/videos/TZF0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts down her bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <9.0>." + } + ] + }, + { + "video_id": "TZF0X", + "question_id": "TZF0X", + "video_file": "Charades/videos/TZF0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes her shoes off'." + }, + { + "from": "gpt", + "value": "From <9.1> to <13.7>." + } + ] + }, + { + "video_id": "TZF0X", + "question_id": "TZF0X", + "video_file": "Charades/videos/TZF0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a bag of clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <9.1>." + } + ] + }, + { + "video_id": "B6LRC", + "question_id": "B6LRC", + "video_file": "Charades/videos/B6LRC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes a pair of shoes from the shelf'." + }, + { + "from": "gpt", + "value": "From <17.7> to <23.7>." + } + ] + }, + { + "video_id": "B6LRC", + "question_id": "B6LRC", + "video_file": "Charades/videos/B6LRC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding the shoes'." + }, + { + "from": "gpt", + "value": "From <18.3> to <31.3>." + } + ] + }, + { + "video_id": "BZ171", + "question_id": "BZ171", + "video_file": "Charades/videos/BZ171.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is awakening'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "BZ171", + "question_id": "BZ171", + "video_file": "Charades/videos/BZ171.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they sit up on their bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.8> to <30.0>." + } + ] + }, + { + "video_id": "4Q3N9", + "question_id": "4Q3N9", + "video_file": "Charades/videos/4Q3N9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts laughing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.2> to <11.6>." + } + ] + }, + { + "video_id": "4Q3N9", + "question_id": "4Q3N9", + "video_file": "Charades/videos/4Q3N9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'as the person walks into the closet smiling'." + }, + { + "from": "gpt", + "value": "From <6.0> to <11.0>." + } + ] + }, + { + "video_id": "JH47V", + "question_id": "JH47V", + "video_file": "Charades/videos/JH47V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the man pours mountain dew into a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <17.0>." + } + ] + }, + { + "video_id": "IQ0ZX", + "question_id": "IQ0ZX", + "video_file": "Charades/videos/IQ0ZX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person watching a video on a laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "IQ0ZX", + "question_id": "IQ0ZX", + "video_file": "Charades/videos/IQ0ZX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking from a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.8> to <26.3>." + } + ] + }, + { + "video_id": "JSBMM", + "question_id": "JSBMM", + "video_file": "Charades/videos/JSBMM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes medicine from the cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.8> to <25.9>." + } + ] + }, + { + "video_id": "C1F1D", + "question_id": "C1F1D", + "video_file": "Charades/videos/C1F1D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person talk on a phone'." + }, + { + "from": "gpt", + "value": "From <16.8> to <30.0>." + } + ] + }, + { + "video_id": "C1F1D", + "question_id": "C1F1D", + "video_file": "Charades/videos/C1F1D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person the put the cup down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.8> to <17.3>." + } + ] + }, + { + "video_id": "C1F1D", + "question_id": "C1F1D", + "video_file": "Charades/videos/C1F1D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking a cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "C1F1D", + "question_id": "C1F1D", + "video_file": "Charades/videos/C1F1D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts talking on phone'." + }, + { + "from": "gpt", + "value": "From <16.8> to <30.0>." + } + ] + }, + { + "video_id": "J0SZG", + "question_id": "J0SZG", + "video_file": "Charades/videos/J0SZG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is putting a blanket into a washing machine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "J0SZG", + "question_id": "J0SZG", + "video_file": "Charades/videos/J0SZG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts a blanket in the washing machine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "J0SZG", + "question_id": "J0SZG", + "video_file": "Charades/videos/J0SZG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <14.4>." + } + ] + }, + { + "video_id": "J0SZG", + "question_id": "J0SZG", + "video_file": "Charades/videos/J0SZG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person loads their washer with clothes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.4> to <31.0>." + } + ] + }, + { + "video_id": "J0SZG", + "question_id": "J0SZG", + "video_file": "Charades/videos/J0SZG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person occasionally tapping the door to the laundry room open' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <14.4>." + } + ] + }, + { + "video_id": "JPEM7", + "question_id": "JPEM7", + "video_file": "Charades/videos/JPEM7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating some food'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.0>." + } + ] + }, + { + "video_id": "JPEM7", + "question_id": "JPEM7", + "video_file": "Charades/videos/JPEM7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits down to have a bite to eat'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.0>." + } + ] + }, + { + "video_id": "JPEM7", + "question_id": "JPEM7", + "video_file": "Charades/videos/JPEM7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins to eat'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.0>." + } + ] + }, + { + "video_id": "L5G6N", + "question_id": "L5G6N", + "video_file": "Charades/videos/L5G6N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is closing the medicine cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <9.1>." + } + ] + }, + { + "video_id": "L5G6N", + "question_id": "L5G6N", + "video_file": "Charades/videos/L5G6N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person closes cabinet drawers' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.1> to <9.1>." + } + ] + }, + { + "video_id": "L5G6N", + "question_id": "L5G6N", + "video_file": "Charades/videos/L5G6N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns off the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.4> to <30.0>." + } + ] + }, + { + "video_id": "L5G6N", + "question_id": "L5G6N", + "video_file": "Charades/videos/L5G6N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turning off the light in the bathroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.4> to <30.0>." + } + ] + }, + { + "video_id": "T1CQE", + "question_id": "T1CQE", + "video_file": "Charades/videos/T1CQE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes out a laptop computer' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.4> to <29.9>." + } + ] + }, + { + "video_id": "T1CQE", + "question_id": "T1CQE", + "video_file": "Charades/videos/T1CQE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes their laptop out of a bag'." + }, + { + "from": "gpt", + "value": "From <8.0> to <14.9>." + } + ] + }, + { + "video_id": "T1CQE", + "question_id": "T1CQE", + "video_file": "Charades/videos/T1CQE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes out a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.4> to <29.9>." + } + ] + }, + { + "video_id": "T1CQE", + "question_id": "T1CQE", + "video_file": "Charades/videos/T1CQE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.9> to <27.4>." + } + ] + }, + { + "video_id": "KWQKH", + "question_id": "KWQKH", + "video_file": "Charades/videos/KWQKH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drink coffee out of a glass open closet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.8> to <12.3>." + } + ] + }, + { + "video_id": "KWQKH", + "question_id": "KWQKH", + "video_file": "Charades/videos/KWQKH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they begin opening the closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.2> to <13.7>." + } + ] + }, + { + "video_id": "KWQKH", + "question_id": "KWQKH", + "video_file": "Charades/videos/KWQKH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a vacuum out the closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.0> to <17.0>." + } + ] + }, + { + "video_id": "KWQKH", + "question_id": "KWQKH", + "video_file": "Charades/videos/KWQKH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <21.5>." + } + ] + }, + { + "video_id": "KWQKH", + "question_id": "KWQKH", + "video_file": "Charades/videos/KWQKH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take out a vacuum'." + }, + { + "from": "gpt", + "value": "From <10.0> to <17.0>." + } + ] + }, + { + "video_id": "KWQKH", + "question_id": "KWQKH", + "video_file": "Charades/videos/KWQKH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <20.0>." + } + ] + }, + { + "video_id": "2QYZA", + "question_id": "2QYZA", + "video_file": "Charades/videos/2QYZA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats some of the food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.9> to <31.0>." + } + ] + }, + { + "video_id": "2QYZA", + "question_id": "2QYZA", + "video_file": "Charades/videos/2QYZA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts eating' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.9> to <31.0>." + } + ] + }, + { + "video_id": "2QYZA", + "question_id": "2QYZA", + "video_file": "Charades/videos/2QYZA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person fixes something to eat'." + }, + { + "from": "gpt", + "value": "From <21.9> to <31.0>." + } + ] + }, + { + "video_id": "2QYZA", + "question_id": "2QYZA", + "video_file": "Charades/videos/2QYZA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person opens a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.2>." + } + ] + }, + { + "video_id": "2QYZA", + "question_id": "2QYZA", + "video_file": "Charades/videos/2QYZA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the food on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <12.9>." + } + ] + }, + { + "video_id": "758T4", + "question_id": "758T4", + "video_file": "Charades/videos/758T4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take their laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.4> to <18.6>." + } + ] + }, + { + "video_id": "UVN96", + "question_id": "UVN96", + "video_file": "Charades/videos/UVN96.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a phone from a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.3> to <15.7>." + } + ] + }, + { + "video_id": "UVN96", + "question_id": "UVN96", + "video_file": "Charades/videos/UVN96.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person opens a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <16.5>." + } + ] + }, + { + "video_id": "YIM12", + "question_id": "YIM12", + "video_file": "Charades/videos/YIM12.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the smiling person enters' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.9> to <32.0>." + } + ] + }, + { + "video_id": "F5OZ5", + "question_id": "F5OZ5", + "video_file": "Charades/videos/F5OZ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is in the basement throwing a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.5> to <25.0>." + } + ] + }, + { + "video_id": "F5OZ5", + "question_id": "F5OZ5", + "video_file": "Charades/videos/F5OZ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they throw the pillow on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.5> to <25.0>." + } + ] + }, + { + "video_id": "97JCX", + "question_id": "97JCX", + "video_file": "Charades/videos/97JCX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinks a cup of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <8.1>." + } + ] + }, + { + "video_id": "97JCX", + "question_id": "97JCX", + "video_file": "Charades/videos/97JCX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is dressing in the bathroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.9> to <12.4>." + } + ] + }, + { + "video_id": "XB4HN", + "question_id": "XB4HN", + "video_file": "Charades/videos/XB4HN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.2> to <13.3>." + } + ] + }, + { + "video_id": "7JZGN", + "question_id": "7JZGN", + "video_file": "Charades/videos/7JZGN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person closes a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.9> to <13.1>." + } + ] + }, + { + "video_id": "5L04F", + "question_id": "5L04F", + "video_file": "Charades/videos/5L04F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person sitting on stairs laughs' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.3> to <16.7>." + } + ] + }, + { + "video_id": "SP6M4", + "question_id": "SP6M4", + "video_file": "Charades/videos/SP6M4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <7.5>." + } + ] + }, + { + "video_id": "SP6M4", + "question_id": "SP6M4", + "video_file": "Charades/videos/SP6M4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a sandwich out of the bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.4> to <12.4>." + } + ] + }, + { + "video_id": "SP6M4", + "question_id": "SP6M4", + "video_file": "Charades/videos/SP6M4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes some food out of a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.7> to <13.0>." + } + ] + }, + { + "video_id": "SP6M4", + "question_id": "SP6M4", + "video_file": "Charades/videos/SP6M4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.2> to <11.4>." + } + ] + }, + { + "video_id": "SP6M4", + "question_id": "SP6M4", + "video_file": "Charades/videos/SP6M4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats the sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.8> to <14.0>." + } + ] + }, + { + "video_id": "SP6M4", + "question_id": "SP6M4", + "video_file": "Charades/videos/SP6M4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins eating it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.8> to <14.0>." + } + ] + }, + { + "video_id": "SP6M4", + "question_id": "SP6M4", + "video_file": "Charades/videos/SP6M4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats food'." + }, + { + "from": "gpt", + "value": "From <9.8> to <14.0>." + } + ] + }, + { + "video_id": "C4E9X", + "question_id": "C4E9X", + "video_file": "Charades/videos/C4E9X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.9> to <23.8>." + } + ] + }, + { + "video_id": "07RDG", + "question_id": "07RDG", + "video_file": "Charades/videos/07RDG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts down some dishes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.7> to <7.5>." + } + ] + }, + { + "video_id": "JL6MG", + "question_id": "JL6MG", + "video_file": "Charades/videos/JL6MG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts eating'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.4> to <36.2>." + } + ] + }, + { + "video_id": "JL6MG", + "question_id": "JL6MG", + "video_file": "Charades/videos/JL6MG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.6> to <41.0>." + } + ] + }, + { + "video_id": "83KZR", + "question_id": "83KZR", + "video_file": "Charades/videos/83KZR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person is closing the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.2> to <22.9>." + } + ] + }, + { + "video_id": "U2FA5", + "question_id": "U2FA5", + "video_file": "Charades/videos/U2FA5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begin working on their laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <31.0>." + } + ] + }, + { + "video_id": "U2FA5", + "question_id": "U2FA5", + "video_file": "Charades/videos/U2FA5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits on chair'." + }, + { + "from": "gpt", + "value": "From <16.4> to <21.6>." + } + ] + }, + { + "video_id": "U2FA5", + "question_id": "U2FA5", + "video_file": "Charades/videos/U2FA5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sit in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.4> to <21.6>." + } + ] + }, + { + "video_id": "U2FA5", + "question_id": "U2FA5", + "video_file": "Charades/videos/U2FA5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts working on a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.1> to <31.0>." + } + ] + }, + { + "video_id": "KVONB", + "question_id": "KVONB", + "video_file": "Charades/videos/KVONB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens up a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "KVONB", + "question_id": "KVONB", + "video_file": "Charades/videos/KVONB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts clothes on the vanity'." + }, + { + "from": "gpt", + "value": "From <21.5> to <27.8>." + } + ] + }, + { + "video_id": "7EHQU", + "question_id": "7EHQU", + "video_file": "Charades/videos/7EHQU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.1> to <12.6>." + } + ] + }, + { + "video_id": "7EHQU", + "question_id": "7EHQU", + "video_file": "Charades/videos/7EHQU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a sip from the cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.5> to <15.5>." + } + ] + }, + { + "video_id": "7EHQU", + "question_id": "7EHQU", + "video_file": "Charades/videos/7EHQU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts down a blanket by the tub' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "7EHQU", + "question_id": "7EHQU", + "video_file": "Charades/videos/7EHQU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pours liquid into a small plastic cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.0> to <21.4>." + } + ] + }, + { + "video_id": "QRLOG", + "question_id": "QRLOG", + "video_file": "Charades/videos/QRLOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.3> to <11.3>." + } + ] + }, + { + "video_id": "QRLOG", + "question_id": "QRLOG", + "video_file": "Charades/videos/QRLOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put the phone down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.2>." + } + ] + }, + { + "video_id": "QRLOG", + "question_id": "QRLOG", + "video_file": "Charades/videos/QRLOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person when they stand up they grab a broom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.3> to <11.3>." + } + ] + }, + { + "video_id": "SE3WY", + "question_id": "SE3WY", + "video_file": "Charades/videos/SE3WY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <25.5>." + } + ] + }, + { + "video_id": "SE3WY", + "question_id": "SE3WY", + "video_file": "Charades/videos/SE3WY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <25.5>." + } + ] + }, + { + "video_id": "YJFWW", + "question_id": "YJFWW", + "video_file": "Charades/videos/YJFWW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person runs into the garage with a phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "YJFWW", + "question_id": "YJFWW", + "video_file": "Charades/videos/YJFWW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs into a garage holding a phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "EI5M3", + "question_id": "EI5M3", + "video_file": "Charades/videos/EI5M3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a closet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.2>." + } + ] + }, + { + "video_id": "EI5M3", + "question_id": "EI5M3", + "video_file": "Charades/videos/EI5M3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts in a clean towel' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.1> to <14.9>." + } + ] + }, + { + "video_id": "EI5M3", + "question_id": "EI5M3", + "video_file": "Charades/videos/EI5M3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person open a closet door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.2>." + } + ] + }, + { + "video_id": "EI5M3", + "question_id": "EI5M3", + "video_file": "Charades/videos/EI5M3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens the wardrobe in a closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <7.0>." + } + ] + }, + { + "video_id": "EI5M3", + "question_id": "EI5M3", + "video_file": "Charades/videos/EI5M3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting a towel on the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.1> to <14.9>." + } + ] + }, + { + "video_id": "EAVWL", + "question_id": "EAVWL", + "video_file": "Charades/videos/EAVWL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts a cup on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.7> to <13.2>." + } + ] + }, + { + "video_id": "297S4", + "question_id": "297S4", + "video_file": "Charades/videos/297S4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes door behind them' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <27.5> to <34.7>." + } + ] + }, + { + "video_id": "297S4", + "question_id": "297S4", + "video_file": "Charades/videos/297S4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person stand on one leg on chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.2> to <26.5>." + } + ] + }, + { + "video_id": "297S4", + "question_id": "297S4", + "video_file": "Charades/videos/297S4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the door behind them' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.5> to <34.7>." + } + ] + }, + { + "video_id": "297S4", + "question_id": "297S4", + "video_file": "Charades/videos/297S4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person stands on a chair in the closet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.2> to <26.5>." + } + ] + }, + { + "video_id": "UXEL9", + "question_id": "UXEL9", + "video_file": "Charades/videos/UXEL9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person started opening the boxes'." + }, + { + "from": "gpt", + "value": "From <14.2> to <26.8>." + } + ] + }, + { + "video_id": "UXEL9", + "question_id": "UXEL9", + "video_file": "Charades/videos/UXEL9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts it down on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.4> to <18.7>." + } + ] + }, + { + "video_id": "2JP64", + "question_id": "2JP64", + "video_file": "Charades/videos/2JP64.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person decide to eat a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <21.9>." + } + ] + }, + { + "video_id": "CBW6S", + "question_id": "CBW6S", + "video_file": "Charades/videos/CBW6S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they laugh'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.9> to <17.5>." + } + ] + }, + { + "video_id": "CBW6S", + "question_id": "CBW6S", + "video_file": "Charades/videos/CBW6S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person looks out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <11.0>." + } + ] + }, + { + "video_id": "7UQFM", + "question_id": "7UQFM", + "video_file": "Charades/videos/7UQFM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person wards they start drinking from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.8> to <30.6>." + } + ] + }, + { + "video_id": "7UQFM", + "question_id": "7UQFM", + "video_file": "Charades/videos/7UQFM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'that same person puts their phone in their pocket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <18.9>." + } + ] + }, + { + "video_id": "7UQFM", + "question_id": "7UQFM", + "video_file": "Charades/videos/7UQFM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts their phone in their pocket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.5> to <18.9>." + } + ] + }, + { + "video_id": "7UQFM", + "question_id": "7UQFM", + "video_file": "Charades/videos/7UQFM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is playing with their phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <11.0>." + } + ] + }, + { + "video_id": "7UQFM", + "question_id": "7UQFM", + "video_file": "Charades/videos/7UQFM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put the phone in their pocket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <18.9>." + } + ] + }, + { + "video_id": "IF4Y9", + "question_id": "IF4Y9", + "video_file": "Charades/videos/IF4Y9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person closes the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.9> to <30.7>." + } + ] + }, + { + "video_id": "IF4Y9", + "question_id": "IF4Y9", + "video_file": "Charades/videos/IF4Y9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.9> to <30.7>." + } + ] + }, + { + "video_id": "PS29P", + "question_id": "PS29P", + "video_file": "Charades/videos/PS29P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is running down the stairs holding a broom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.9>." + } + ] + }, + { + "video_id": "LLOGB", + "question_id": "LLOGB", + "video_file": "Charades/videos/LLOGB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a glass of water'." + }, + { + "from": "gpt", + "value": "From <0.6> to <9.0>." + } + ] + }, + { + "video_id": "K9V2R", + "question_id": "K9V2R", + "video_file": "Charades/videos/K9V2R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holds a box'." + }, + { + "from": "gpt", + "value": "From <32.6> to <35.0>." + } + ] + }, + { + "video_id": "KNGUT", + "question_id": "KNGUT", + "video_file": "Charades/videos/KNGUT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating something from a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.0> to <31.0>." + } + ] + }, + { + "video_id": "AXZVQ", + "question_id": "AXZVQ", + "video_file": "Charades/videos/AXZVQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <10.1>." + } + ] + }, + { + "video_id": "DSHVL", + "question_id": "DSHVL", + "video_file": "Charades/videos/DSHVL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person walks out through the doorway'." + }, + { + "from": "gpt", + "value": "From <17.3> to <23.0>." + } + ] + }, + { + "video_id": "WA7WD", + "question_id": "WA7WD", + "video_file": "Charades/videos/WA7WD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating it'." + }, + { + "from": "gpt", + "value": "From <24.2> to <32.0>." + } + ] + }, + { + "video_id": "WA7WD", + "question_id": "WA7WD", + "video_file": "Charades/videos/WA7WD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins to eat it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.2> to <32.0>." + } + ] + }, + { + "video_id": "WA7WD", + "question_id": "WA7WD", + "video_file": "Charades/videos/WA7WD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they open a lunch bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <20.9>." + } + ] + }, + { + "video_id": "I5F23", + "question_id": "I5F23", + "video_file": "Charades/videos/I5F23.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is walking to turn the light on' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.2> to <5.2>." + } + ] + }, + { + "video_id": "X4YHQ", + "question_id": "X4YHQ", + "video_file": "Charades/videos/X4YHQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sitting down in a chair'." + }, + { + "from": "gpt", + "value": "From <10.4> to <28.9>." + } + ] + }, + { + "video_id": "CTFYL", + "question_id": "CTFYL", + "video_file": "Charades/videos/CTFYL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is putting food in a cabinet'." + }, + { + "from": "gpt", + "value": "From <2.8> to <19.4>." + } + ] + }, + { + "video_id": "CTFYL", + "question_id": "CTFYL", + "video_file": "Charades/videos/CTFYL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts food in a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <19.4>." + } + ] + }, + { + "video_id": "C6LW1", + "question_id": "C6LW1", + "video_file": "Charades/videos/C6LW1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person looks out the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "C6LW1", + "question_id": "C6LW1", + "video_file": "Charades/videos/C6LW1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a picture of the outside' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <21.4>." + } + ] + }, + { + "video_id": "C6LW1", + "question_id": "C6LW1", + "video_file": "Charades/videos/C6LW1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is looking out the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "C6LW1", + "question_id": "C6LW1", + "video_file": "Charades/videos/C6LW1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking a bottle of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.0> to <30.0>." + } + ] + }, + { + "video_id": "7SKSD", + "question_id": "7SKSD", + "video_file": "Charades/videos/7SKSD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is watching themself in a mirror'." + }, + { + "from": "gpt", + "value": "From <1.0> to <12.1>." + } + ] + }, + { + "video_id": "PVB4S", + "question_id": "PVB4S", + "video_file": "Charades/videos/PVB4S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they throw the bag on to the floor'." + }, + { + "from": "gpt", + "value": "From <16.7> to <25.4>." + } + ] + }, + { + "video_id": "PVB4S", + "question_id": "PVB4S", + "video_file": "Charades/videos/PVB4S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person throws the bag next to the sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <25.4>." + } + ] + }, + { + "video_id": "PVB4S", + "question_id": "PVB4S", + "video_file": "Charades/videos/PVB4S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks down some stairs holding a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <18.7>." + } + ] + }, + { + "video_id": "2IS9G", + "question_id": "2IS9G", + "video_file": "Charades/videos/2IS9G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns on the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.2> to <14.8>." + } + ] + }, + { + "video_id": "2IS9G", + "question_id": "2IS9G", + "video_file": "Charades/videos/2IS9G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person awakens' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <9.2>." + } + ] + }, + { + "video_id": "2IS9G", + "question_id": "2IS9G", + "video_file": "Charades/videos/2IS9G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turn on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.2> to <14.8>." + } + ] + }, + { + "video_id": "2IS9G", + "question_id": "2IS9G", + "video_file": "Charades/videos/2IS9G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns on a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.2> to <14.8>." + } + ] + }, + { + "video_id": "V7PGV", + "question_id": "V7PGV", + "video_file": "Charades/videos/V7PGV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the towel down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.5> to <25.1>." + } + ] + }, + { + "video_id": "V7PGV", + "question_id": "V7PGV", + "video_file": "Charades/videos/V7PGV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns off the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.9> to <26.8>." + } + ] + }, + { + "video_id": "V7PGV", + "question_id": "V7PGV", + "video_file": "Charades/videos/V7PGV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs out of the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.8> to <27.0>." + } + ] + }, + { + "video_id": "58R8A", + "question_id": "58R8A", + "video_file": "Charades/videos/58R8A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.5> to <17.9>." + } + ] + }, + { + "video_id": "O1YWR", + "question_id": "O1YWR", + "video_file": "Charades/videos/O1YWR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting the clothes in a bag on their bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.6> to <31.4>." + } + ] + }, + { + "video_id": "O1YWR", + "question_id": "O1YWR", + "video_file": "Charades/videos/O1YWR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing by grasping' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.5>." + } + ] + }, + { + "video_id": "CM293", + "question_id": "CM293", + "video_file": "Charades/videos/CM293.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <16.0>." + } + ] + }, + { + "video_id": "CM293", + "question_id": "CM293", + "video_file": "Charades/videos/CM293.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is walking down the hallway eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.5> to <9.4>." + } + ] + }, + { + "video_id": "CM293", + "question_id": "CM293", + "video_file": "Charades/videos/CM293.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.5> to <9.4>." + } + ] + }, + { + "video_id": "CM293", + "question_id": "CM293", + "video_file": "Charades/videos/CM293.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats a bite of sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.5> to <9.4>." + } + ] + }, + { + "video_id": "CM293", + "question_id": "CM293", + "video_file": "Charades/videos/CM293.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they open a closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <16.0>." + } + ] + }, + { + "video_id": "29E4A", + "question_id": "29E4A", + "video_file": "Charades/videos/29E4A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person is standing in hallway putting on shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <19.9>." + } + ] + }, + { + "video_id": "29E4A", + "question_id": "29E4A", + "video_file": "Charades/videos/29E4A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts dressing by mirror to go out'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <36.3>." + } + ] + }, + { + "video_id": "29E4A", + "question_id": "29E4A", + "video_file": "Charades/videos/29E4A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is putting on shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.8> to <19.9>." + } + ] + }, + { + "video_id": "29E4A", + "question_id": "29E4A", + "video_file": "Charades/videos/29E4A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting on shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <19.9>." + } + ] + }, + { + "video_id": "XBPEJ", + "question_id": "XBPEJ", + "video_file": "Charades/videos/XBPEJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person he takes a box off the washing machine' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.1> to <9.8>." + } + ] + }, + { + "video_id": "XW1LX", + "question_id": "XW1LX", + "video_file": "Charades/videos/XW1LX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns on the light next to the bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.5> to <21.1>." + } + ] + }, + { + "video_id": "XW1LX", + "question_id": "XW1LX", + "video_file": "Charades/videos/XW1LX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person awakens in bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.2>." + } + ] + }, + { + "video_id": "9BCZU", + "question_id": "9BCZU", + "video_file": "Charades/videos/9BCZU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they undress'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.6> to <26.6>." + } + ] + }, + { + "video_id": "47I6H", + "question_id": "47I6H", + "video_file": "Charades/videos/47I6H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person reaches to turn the light on' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <17.9>." + } + ] + }, + { + "video_id": "47I6H", + "question_id": "47I6H", + "video_file": "Charades/videos/47I6H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turned on the light'." + }, + { + "from": "gpt", + "value": "From <12.0> to <17.9>." + } + ] + }, + { + "video_id": "HONRP", + "question_id": "HONRP", + "video_file": "Charades/videos/HONRP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.8> to <24.2>." + } + ] + }, + { + "video_id": "LWEPB", + "question_id": "LWEPB", + "video_file": "Charades/videos/LWEPB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <16.8>." + } + ] + }, + { + "video_id": "LWEPB", + "question_id": "LWEPB", + "video_file": "Charades/videos/LWEPB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person stand up' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.9> to <16.8>." + } + ] + }, + { + "video_id": "LWEPB", + "question_id": "LWEPB", + "video_file": "Charades/videos/LWEPB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.6> to <26.1>." + } + ] + }, + { + "video_id": "LWEPB", + "question_id": "LWEPB", + "video_file": "Charades/videos/LWEPB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.4> to <15.4>." + } + ] + }, + { + "video_id": "LWEPB", + "question_id": "LWEPB", + "video_file": "Charades/videos/LWEPB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person playing on their phone'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.3>." + } + ] + }, + { + "video_id": "LWEPB", + "question_id": "LWEPB", + "video_file": "Charades/videos/LWEPB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting on a couch' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.9>." + } + ] + }, + { + "video_id": "LWEPB", + "question_id": "LWEPB", + "video_file": "Charades/videos/LWEPB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put the pillow on a bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.1> to <15.3>." + } + ] + }, + { + "video_id": "LWEPB", + "question_id": "LWEPB", + "video_file": "Charades/videos/LWEPB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting on a chair with a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.9>." + } + ] + }, + { + "video_id": "LWEPB", + "question_id": "LWEPB", + "video_file": "Charades/videos/LWEPB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open a door'." + }, + { + "from": "gpt", + "value": "From <18.7> to <26.3>." + } + ] + }, + { + "video_id": "CI6H7", + "question_id": "CI6H7", + "video_file": "Charades/videos/CI6H7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.6> to <7.7>." + } + ] + }, + { + "video_id": "CI6H7", + "question_id": "CI6H7", + "video_file": "Charades/videos/CI6H7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looking at a picture'." + }, + { + "from": "gpt", + "value": "From <2.4> to <6.8>." + } + ] + }, + { + "video_id": "LHMMJ", + "question_id": "LHMMJ", + "video_file": "Charades/videos/LHMMJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person sits in a chair in the kitchen'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.3>." + } + ] + }, + { + "video_id": "HHZUI", + "question_id": "HHZUI", + "video_file": "Charades/videos/HHZUI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a book from it'." + }, + { + "from": "gpt", + "value": "From <24.2> to <34.0>." + } + ] + }, + { + "video_id": "808H5", + "question_id": "808H5", + "video_file": "Charades/videos/808H5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening the closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.0> to <18.7>." + } + ] + }, + { + "video_id": "808H5", + "question_id": "808H5", + "video_file": "Charades/videos/808H5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <19.0>." + } + ] + }, + { + "video_id": "JUI9U", + "question_id": "JUI9U", + "video_file": "Charades/videos/JUI9U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they stand up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <27.2>." + } + ] + }, + { + "video_id": "JUI9U", + "question_id": "JUI9U", + "video_file": "Charades/videos/JUI9U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a refrigerator door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.4>." + } + ] + }, + { + "video_id": "JUI9U", + "question_id": "JUI9U", + "video_file": "Charades/videos/JUI9U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person smiling as he/she opens the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.4>." + } + ] + }, + { + "video_id": "GYRQH", + "question_id": "GYRQH", + "video_file": "Charades/videos/GYRQH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts undressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <20.1>." + } + ] + }, + { + "video_id": "GYRQH", + "question_id": "GYRQH", + "video_file": "Charades/videos/GYRQH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is undressing in front of a mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.6> to <20.1>." + } + ] + }, + { + "video_id": "GYRQH", + "question_id": "GYRQH", + "video_file": "Charades/videos/GYRQH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'throws a pillow to person a' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.6> to <26.1>." + } + ] + }, + { + "video_id": "ZBDB5", + "question_id": "ZBDB5", + "video_file": "Charades/videos/ZBDB5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they close the laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.4> to <14.6>." + } + ] + }, + { + "video_id": "ZBDB5", + "question_id": "ZBDB5", + "video_file": "Charades/videos/ZBDB5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person closes a laptop in the bedroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.4> to <14.6>." + } + ] + }, + { + "video_id": "16S3C", + "question_id": "16S3C", + "video_file": "Charades/videos/16S3C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looks at a picture on it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <28.0>." + } + ] + }, + { + "video_id": "16S3C", + "question_id": "16S3C", + "video_file": "Charades/videos/16S3C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person leaning on a doorway eats a sandwich'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.6>." + } + ] + }, + { + "video_id": "EQO9N", + "question_id": "EQO9N", + "video_file": "Charades/videos/EQO9N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person b is putting clothes on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.5>." + } + ] + }, + { + "video_id": "5NF0R", + "question_id": "5NF0R", + "video_file": "Charades/videos/5NF0R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person sits down on a couch' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <7.0>." + } + ] + }, + { + "video_id": "5NF0R", + "question_id": "5NF0R", + "video_file": "Charades/videos/5NF0R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a woman puts a box on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.9> to <13.0>." + } + ] + }, + { + "video_id": "5NF0R", + "question_id": "5NF0R", + "video_file": "Charades/videos/5NF0R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person plays on her phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.4> to <31.9>." + } + ] + }, + { + "video_id": "5NF0R", + "question_id": "5NF0R", + "video_file": "Charades/videos/5NF0R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins to play with their phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.4> to <31.9>." + } + ] + }, + { + "video_id": "5NF0R", + "question_id": "5NF0R", + "video_file": "Charades/videos/5NF0R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts a box on a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.9> to <13.0>." + } + ] + }, + { + "video_id": "MOKPE", + "question_id": "MOKPE", + "video_file": "Charades/videos/MOKPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the box back up on the ledge' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.6> to <21.2>." + } + ] + }, + { + "video_id": "MOKPE", + "question_id": "MOKPE", + "video_file": "Charades/videos/MOKPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes medicine out of the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.4>." + } + ] + }, + { + "video_id": "4WX0H", + "question_id": "4WX0H", + "video_file": "Charades/videos/4WX0H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a other person is laughing into a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.5> to <27.2>." + } + ] + }, + { + "video_id": "0BXRP", + "question_id": "0BXRP", + "video_file": "Charades/videos/0BXRP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'this person starts cooking at the stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.4> to <28.0>." + } + ] + }, + { + "video_id": "OHNF4", + "question_id": "OHNF4", + "video_file": "Charades/videos/OHNF4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they take a glass of water from the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <23.8>." + } + ] + }, + { + "video_id": "AB06W", + "question_id": "AB06W", + "video_file": "Charades/videos/AB06W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person lies down to watch television from bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.1> to <30.0>." + } + ] + }, + { + "video_id": "2XT4N", + "question_id": "2XT4N", + "video_file": "Charades/videos/2XT4N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person runs out of the room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.6> to <18.9>." + } + ] + }, + { + "video_id": "OTLX9", + "question_id": "OTLX9", + "video_file": "Charades/videos/OTLX9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is dressing in the living room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.4>." + } + ] + }, + { + "video_id": "PB0JI", + "question_id": "PB0JI", + "video_file": "Charades/videos/PB0JI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the laptop on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.3> to <18.3>." + } + ] + }, + { + "video_id": "SFHY2", + "question_id": "SFHY2", + "video_file": "Charades/videos/SFHY2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throwing food towards the table'." + }, + { + "from": "gpt", + "value": "From <4.1> to <9.4>." + } + ] + }, + { + "video_id": "EW3VV", + "question_id": "EW3VV", + "video_file": "Charades/videos/EW3VV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits on the sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.7> to <9.3>." + } + ] + }, + { + "video_id": "EW3VV", + "question_id": "EW3VV", + "video_file": "Charades/videos/EW3VV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begins drinking from the glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <16.3>." + } + ] + }, + { + "video_id": "YRPAR", + "question_id": "YRPAR", + "video_file": "Charades/videos/YRPAR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is holding a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.7> to <9.0>." + } + ] + }, + { + "video_id": "JL5O0", + "question_id": "JL5O0", + "video_file": "Charades/videos/JL5O0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put something on a table'." + }, + { + "from": "gpt", + "value": "From <6.3> to <11.8>." + } + ] + }, + { + "video_id": "JL5O0", + "question_id": "JL5O0", + "video_file": "Charades/videos/JL5O0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting the phone on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.3> to <11.8>." + } + ] + }, + { + "video_id": "NMRF0", + "question_id": "NMRF0", + "video_file": "Charades/videos/NMRF0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes a picture of the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.6> to <22.6>." + } + ] + }, + { + "video_id": "NMRF0", + "question_id": "NMRF0", + "video_file": "Charades/videos/NMRF0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person leaves the room smiling'." + }, + { + "from": "gpt", + "value": "From <27.6> to <32.0>." + } + ] + }, + { + "video_id": "05ZNF", + "question_id": "05ZNF", + "video_file": "Charades/videos/05ZNF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person watching a laptop screen' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.1> to <13.0>." + } + ] + }, + { + "video_id": "05ZNF", + "question_id": "05ZNF", + "video_file": "Charades/videos/05ZNF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats some food'." + }, + { + "from": "gpt", + "value": "From <23.1> to <32.0>." + } + ] + }, + { + "video_id": "05ZNF", + "question_id": "05ZNF", + "video_file": "Charades/videos/05ZNF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats them'." + }, + { + "from": "gpt", + "value": "From <23.1> to <32.0>." + } + ] + }, + { + "video_id": "Q366H", + "question_id": "Q366H", + "video_file": "Charades/videos/Q366H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a third person is sitting in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.3> to <32.0>." + } + ] + }, + { + "video_id": "Q366H", + "question_id": "Q366H", + "video_file": "Charades/videos/Q366H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting on the floor reading a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.4>." + } + ] + }, + { + "video_id": "Q366H", + "question_id": "Q366H", + "video_file": "Charades/videos/Q366H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sitting on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.4>." + } + ] + }, + { + "video_id": "4JOAD", + "question_id": "4JOAD", + "video_file": "Charades/videos/4JOAD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts groceries down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "4JOAD", + "question_id": "4JOAD", + "video_file": "Charades/videos/4JOAD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts tidying the counter with a towel' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.6> to <17.6>." + } + ] + }, + { + "video_id": "8ECJ1", + "question_id": "8ECJ1", + "video_file": "Charades/videos/8ECJ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'seated man reads book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.7>." + } + ] + }, + { + "video_id": "06EDS", + "question_id": "06EDS", + "video_file": "Charades/videos/06EDS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is cooking on the stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.9> to <36.0>." + } + ] + }, + { + "video_id": "06EDS", + "question_id": "06EDS", + "video_file": "Charades/videos/06EDS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they're cooking a sandwich in their kitchen'." + }, + { + "from": "gpt", + "value": "From <23.9> to <36.0>." + } + ] + }, + { + "video_id": "0I0FX", + "question_id": "0I0FX", + "video_file": "Charades/videos/0I0FX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person finally eating a sandwich in a garage'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.0> to <31.0>." + } + ] + }, + { + "video_id": "0I0FX", + "question_id": "0I0FX", + "video_file": "Charades/videos/0I0FX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person start eating it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.0> to <31.0>." + } + ] + }, + { + "video_id": "RGY38", + "question_id": "RGY38", + "video_file": "Charades/videos/RGY38.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the bottle down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.0> to <25.7>." + } + ] + }, + { + "video_id": "6VUIO", + "question_id": "6VUIO", + "video_file": "Charades/videos/6VUIO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <14.6>." + } + ] + }, + { + "video_id": "6VUIO", + "question_id": "6VUIO", + "video_file": "Charades/videos/6VUIO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking from a glass in the pantry' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <14.6>." + } + ] + }, + { + "video_id": "6VUIO", + "question_id": "6VUIO", + "video_file": "Charades/videos/6VUIO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.1> to <11.9>." + } + ] + }, + { + "video_id": "WFVD3", + "question_id": "WFVD3", + "video_file": "Charades/videos/WFVD3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws a book inside near a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.2> to <18.4>." + } + ] + }, + { + "video_id": "WFVD3", + "question_id": "WFVD3", + "video_file": "Charades/videos/WFVD3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they turn the light off'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.7> to <14.3>." + } + ] + }, + { + "video_id": "WFVD3", + "question_id": "WFVD3", + "video_file": "Charades/videos/WFVD3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws the book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.2> to <18.4>." + } + ] + }, + { + "video_id": "WFVD3", + "question_id": "WFVD3", + "video_file": "Charades/videos/WFVD3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throw the book in a room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.2> to <18.4>." + } + ] + }, + { + "video_id": "QO49P", + "question_id": "QO49P", + "video_file": "Charades/videos/QO49P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws a pillow on the ground'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <7.8>." + } + ] + }, + { + "video_id": "J27HW", + "question_id": "J27HW", + "video_file": "Charades/videos/J27HW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts it down to eat some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.7> to <32.0>." + } + ] + }, + { + "video_id": "J27HW", + "question_id": "J27HW", + "video_file": "Charades/videos/J27HW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person laughs at something on their phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.8> to <29.1>." + } + ] + }, + { + "video_id": "BZWSJ", + "question_id": "BZWSJ", + "video_file": "Charades/videos/BZWSJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting the cup on a counter'." + }, + { + "from": "gpt", + "value": "From <10.5> to <16.3>." + } + ] + }, + { + "video_id": "BZWSJ", + "question_id": "BZWSJ", + "video_file": "Charades/videos/BZWSJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person drinks from a cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.8> to <8.6>." + } + ] + }, + { + "video_id": "E6389", + "question_id": "E6389", + "video_file": "Charades/videos/E6389.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the laptop onto a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.0> to <16.0>." + } + ] + }, + { + "video_id": "E6389", + "question_id": "E6389", + "video_file": "Charades/videos/E6389.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person holds their laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.0>." + } + ] + }, + { + "video_id": "E6389", + "question_id": "E6389", + "video_file": "Charades/videos/E6389.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the laptop on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <16.0>." + } + ] + }, + { + "video_id": "E6389", + "question_id": "E6389", + "video_file": "Charades/videos/E6389.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take a camera from a counter' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.6> to <20.6>." + } + ] + }, + { + "video_id": "E6389", + "question_id": "E6389", + "video_file": "Charades/videos/E6389.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks through a doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "G1BSD", + "question_id": "G1BSD", + "video_file": "Charades/videos/G1BSD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they both begin eating'." + }, + { + "from": "gpt", + "value": "From <13.9> to <19.5>." + } + ] + }, + { + "video_id": "G1BSD", + "question_id": "G1BSD", + "video_file": "Charades/videos/G1BSD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts eating' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <19.5>." + } + ] + }, + { + "video_id": "G1BSD", + "question_id": "G1BSD", + "video_file": "Charades/videos/G1BSD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is pouring a cup of water'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.0>." + } + ] + }, + { + "video_id": "G1BSD", + "question_id": "G1BSD", + "video_file": "Charades/videos/G1BSD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person pours water into 2 cups'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.0>." + } + ] + }, + { + "video_id": "G1BSD", + "question_id": "G1BSD", + "video_file": "Charades/videos/G1BSD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person both are eating chips or something off of plate' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.9> to <19.5>." + } + ] + }, + { + "video_id": "LEG82", + "question_id": "LEG82", + "video_file": "Charades/videos/LEG82.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <18.3>." + } + ] + }, + { + "video_id": "LEG82", + "question_id": "LEG82", + "video_file": "Charades/videos/LEG82.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eat it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <18.6>." + } + ] + }, + { + "video_id": "LEG82", + "question_id": "LEG82", + "video_file": "Charades/videos/LEG82.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eat again' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.0> to <18.6>." + } + ] + }, + { + "video_id": "LEG82", + "question_id": "LEG82", + "video_file": "Charades/videos/LEG82.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the sandwich down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.2> to <20.7>." + } + ] + }, + { + "video_id": "RUMS5", + "question_id": "RUMS5", + "video_file": "Charades/videos/RUMS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person switches on the light to find some homework'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.1> to <11.5>." + } + ] + }, + { + "video_id": "RUMS5", + "question_id": "RUMS5", + "video_file": "Charades/videos/RUMS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "RUMS5", + "question_id": "RUMS5", + "video_file": "Charades/videos/RUMS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person runs out' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.3> to <29.6>." + } + ] + }, + { + "video_id": "RUMS5", + "question_id": "RUMS5", + "video_file": "Charades/videos/RUMS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person grabs a flashlight'." + }, + { + "from": "gpt", + "value": "From <5.1> to <11.5>." + } + ] + }, + { + "video_id": "RUMS5", + "question_id": "RUMS5", + "video_file": "Charades/videos/RUMS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person is laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.3> to <29.2>." + } + ] + }, + { + "video_id": "RUMS5", + "question_id": "RUMS5", + "video_file": "Charades/videos/RUMS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person runs into the closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.3> to <29.6>." + } + ] + }, + { + "video_id": "RUMS5", + "question_id": "RUMS5", + "video_file": "Charades/videos/RUMS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person runs in'." + }, + { + "from": "gpt", + "value": "From <24.3> to <29.6>." + } + ] + }, + { + "video_id": "RUMS5", + "question_id": "RUMS5", + "video_file": "Charades/videos/RUMS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <27.9>." + } + ] + }, + { + "video_id": "RUMS5", + "question_id": "RUMS5", + "video_file": "Charades/videos/RUMS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door to the closet'." + }, + { + "from": "gpt", + "value": "From <21.6> to <27.9>." + } + ] + }, + { + "video_id": "BVS2Z", + "question_id": "BVS2Z", + "video_file": "Charades/videos/BVS2Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sneezing on a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "MZK2X", + "question_id": "MZK2X", + "video_file": "Charades/videos/MZK2X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a picture with the camera' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.6> to <27.8>." + } + ] + }, + { + "video_id": "MZK2X", + "question_id": "MZK2X", + "video_file": "Charades/videos/MZK2X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a picture'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.6> to <27.8>." + } + ] + }, + { + "video_id": "BTN7H", + "question_id": "BTN7H", + "video_file": "Charades/videos/BTN7H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes their door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.9> to <11.5>." + } + ] + }, + { + "video_id": "BTN7H", + "question_id": "BTN7H", + "video_file": "Charades/videos/BTN7H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begin to take selfies with their phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.1> to <27.8>." + } + ] + }, + { + "video_id": "BTN7H", + "question_id": "BTN7H", + "video_file": "Charades/videos/BTN7H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <11.5>." + } + ] + }, + { + "video_id": "BTN7H", + "question_id": "BTN7H", + "video_file": "Charades/videos/BTN7H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a smiling person begins to undress their clothes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.1> to <25.5>." + } + ] + }, + { + "video_id": "BTN7H", + "question_id": "BTN7H", + "video_file": "Charades/videos/BTN7H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person after closing the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.9> to <11.5>." + } + ] + }, + { + "video_id": "BTN7H", + "question_id": "BTN7H", + "video_file": "Charades/videos/BTN7H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person grasping their phone to take selfies' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.1> to <27.8>." + } + ] + }, + { + "video_id": "P2HZG", + "question_id": "P2HZG", + "video_file": "Charades/videos/P2HZG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they close the laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <10.2>." + } + ] + }, + { + "video_id": "P2HZG", + "question_id": "P2HZG", + "video_file": "Charades/videos/P2HZG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person an opens a cabinet'." + }, + { + "from": "gpt", + "value": "From <13.0> to <18.4>." + } + ] + }, + { + "video_id": "P2HZG", + "question_id": "P2HZG", + "video_file": "Charades/videos/P2HZG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting at the dining room table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.3>." + } + ] + }, + { + "video_id": "RRS54", + "question_id": "RRS54", + "video_file": "Charades/videos/RRS54.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person snuggling under a blanket'." + }, + { + "from": "gpt", + "value": "From <7.5> to <14.2>." + } + ] + }, + { + "video_id": "48BUM", + "question_id": "48BUM", + "video_file": "Charades/videos/48BUM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person fixes their hair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <19.6>." + } + ] + }, + { + "video_id": "48BUM", + "question_id": "48BUM", + "video_file": "Charades/videos/48BUM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting at a table doing homework' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.3>." + } + ] + }, + { + "video_id": "2SYBS", + "question_id": "2SYBS", + "video_file": "Charades/videos/2SYBS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws the pillow back on the bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.4> to <32.1>." + } + ] + }, + { + "video_id": "9O0HH", + "question_id": "9O0HH", + "video_file": "Charades/videos/9O0HH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts eating them in the kitchen' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <25.5>." + } + ] + }, + { + "video_id": "9O0HH", + "question_id": "9O0HH", + "video_file": "Charades/videos/9O0HH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'eating the chips the person pulls out some food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <25.5>." + } + ] + }, + { + "video_id": "9O0HH", + "question_id": "9O0HH", + "video_file": "Charades/videos/9O0HH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begins to cook'." + }, + { + "from": "gpt", + "value": "From <32.4> to <37.0>." + } + ] + }, + { + "video_id": "9O0HH", + "question_id": "9O0HH", + "video_file": "Charades/videos/9O0HH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts cooking it in a pan on the stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <32.4> to <37.0>." + } + ] + }, + { + "video_id": "H378S", + "question_id": "H378S", + "video_file": "Charades/videos/H378S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting away a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.2> to <28.2>." + } + ] + }, + { + "video_id": "TZEJS", + "question_id": "TZEJS", + "video_file": "Charades/videos/TZEJS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person pours himself a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <9.3>." + } + ] + }, + { + "video_id": "TZEJS", + "question_id": "TZEJS", + "video_file": "Charades/videos/TZEJS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts eating' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <17.0>." + } + ] + }, + { + "video_id": "TZEJS", + "question_id": "TZEJS", + "video_file": "Charades/videos/TZEJS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats something' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <17.0>." + } + ] + }, + { + "video_id": "SQ3C9", + "question_id": "SQ3C9", + "video_file": "Charades/videos/SQ3C9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "SA9PB", + "question_id": "SA9PB", + "video_file": "Charades/videos/SA9PB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is eating a snack'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.3> to <9.0>." + } + ] + }, + { + "video_id": "SA9PB", + "question_id": "SA9PB", + "video_file": "Charades/videos/SA9PB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person underneath a blanket' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.3> to <9.0>." + } + ] + }, + { + "video_id": "SA9PB", + "question_id": "SA9PB", + "video_file": "Charades/videos/SA9PB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person eats food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.3> to <9.0>." + } + ] + }, + { + "video_id": "WL1DJ", + "question_id": "WL1DJ", + "video_file": "Charades/videos/WL1DJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person began to take a picture' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.9> to <34.6>." + } + ] + }, + { + "video_id": "K1M15", + "question_id": "K1M15", + "video_file": "Charades/videos/K1M15.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person playing with their phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.4> to <28.9>." + } + ] + }, + { + "video_id": "RK4U5", + "question_id": "RK4U5", + "video_file": "Charades/videos/RK4U5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws it on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.0> to <30.0>." + } + ] + }, + { + "video_id": "AFY5E", + "question_id": "AFY5E", + "video_file": "Charades/videos/AFY5E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking from the cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.1> to <14.6>." + } + ] + }, + { + "video_id": "AFY5E", + "question_id": "AFY5E", + "video_file": "Charades/videos/AFY5E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is opening the door to a bedroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "0BNML", + "question_id": "0BNML", + "video_file": "Charades/videos/0BNML.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person sitting on the bed opens a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <9.6>." + } + ] + }, + { + "video_id": "0BNML", + "question_id": "0BNML", + "video_file": "Charades/videos/0BNML.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put the book on the bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.4> to <24.5>." + } + ] + }, + { + "video_id": "BPN95", + "question_id": "BPN95", + "video_file": "Charades/videos/BPN95.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turn off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.9> to <17.3>." + } + ] + }, + { + "video_id": "K8DDR", + "question_id": "K8DDR", + "video_file": "Charades/videos/K8DDR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.9> to <6.2>." + } + ] + }, + { + "video_id": "ML12Q", + "question_id": "ML12Q", + "video_file": "Charades/videos/ML12Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person begins laughing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.6> to <18.2>." + } + ] + }, + { + "video_id": "NTW37", + "question_id": "NTW37", + "video_file": "Charades/videos/NTW37.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.1> to <24.0>." + } + ] + }, + { + "video_id": "NTW37", + "question_id": "NTW37", + "video_file": "Charades/videos/NTW37.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person attempt to start closing the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <24.0>." + } + ] + }, + { + "video_id": "NTW37", + "question_id": "NTW37", + "video_file": "Charades/videos/NTW37.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they stand on top of a chair'." + }, + { + "from": "gpt", + "value": "From <14.5> to <23.1>." + } + ] + }, + { + "video_id": "NTW37", + "question_id": "NTW37", + "video_file": "Charades/videos/NTW37.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person stands up on a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.5> to <23.1>." + } + ] + }, + { + "video_id": "MCHRP", + "question_id": "MCHRP", + "video_file": "Charades/videos/MCHRP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.5> to <12.2>." + } + ] + }, + { + "video_id": "MCHRP", + "question_id": "MCHRP", + "video_file": "Charades/videos/MCHRP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a door'." + }, + { + "from": "gpt", + "value": "From <1.5> to <12.2>." + } + ] + }, + { + "video_id": "MCHRP", + "question_id": "MCHRP", + "video_file": "Charades/videos/MCHRP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts their phone on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.4> to <17.6>." + } + ] + }, + { + "video_id": "E6P07", + "question_id": "E6P07", + "video_file": "Charades/videos/E6P07.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is reading a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <13.3>." + } + ] + }, + { + "video_id": "E6P07", + "question_id": "E6P07", + "video_file": "Charades/videos/E6P07.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begin laughing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.2> to <18.9>." + } + ] + }, + { + "video_id": "NUW2S", + "question_id": "NUW2S", + "video_file": "Charades/videos/NUW2S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is awakening on a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.4> to <24.9>." + } + ] + }, + { + "video_id": "NUW2S", + "question_id": "NUW2S", + "video_file": "Charades/videos/NUW2S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person stand up'." + }, + { + "from": "gpt", + "value": "From <31.1> to <37.8>." + } + ] + }, + { + "video_id": "JDDD8", + "question_id": "JDDD8", + "video_file": "Charades/videos/JDDD8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person gently throws some food into the pantry'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.8> to <11.8>." + } + ] + }, + { + "video_id": "JDDD8", + "question_id": "JDDD8", + "video_file": "Charades/videos/JDDD8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a man putting things on a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <11.8>." + } + ] + }, + { + "video_id": "BYDSE", + "question_id": "BYDSE", + "video_file": "Charades/videos/BYDSE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins undressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.1> to <32.0>." + } + ] + }, + { + "video_id": "X1XOL", + "question_id": "X1XOL", + "video_file": "Charades/videos/X1XOL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person closes the laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.9> to <26.9>." + } + ] + }, + { + "video_id": "X1XOL", + "question_id": "X1XOL", + "video_file": "Charades/videos/X1XOL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person closes their laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.9> to <26.9>." + } + ] + }, + { + "video_id": "ZOTR8", + "question_id": "ZOTR8", + "video_file": "Charades/videos/ZOTR8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put the shoe down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.7> to <15.0>." + } + ] + }, + { + "video_id": "ZOTR8", + "question_id": "ZOTR8", + "video_file": "Charades/videos/ZOTR8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <19.9>." + } + ] + }, + { + "video_id": "JG7BS", + "question_id": "JG7BS", + "video_file": "Charades/videos/JG7BS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking from the glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <20.4>." + } + ] + }, + { + "video_id": "W2NHA", + "question_id": "W2NHA", + "video_file": "Charades/videos/W2NHA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is laughing at a television' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.6>." + } + ] + }, + { + "video_id": "W2NHA", + "question_id": "W2NHA", + "video_file": "Charades/videos/W2NHA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person forth from laughter' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.6>." + } + ] + }, + { + "video_id": "W2NHA", + "question_id": "W2NHA", + "video_file": "Charades/videos/W2NHA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is seen laughing'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.6>." + } + ] + }, + { + "video_id": "7ZTGM", + "question_id": "7ZTGM", + "video_file": "Charades/videos/7ZTGM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person waiting for dinner to finish cooking on the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <7.6>." + } + ] + }, + { + "video_id": "7ZTGM", + "question_id": "7ZTGM", + "video_file": "Charades/videos/7ZTGM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person cooking' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <7.6>." + } + ] + }, + { + "video_id": "SBT2X", + "question_id": "SBT2X", + "video_file": "Charades/videos/SBT2X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes a vacuum through the door'." + }, + { + "from": "gpt", + "value": "From <24.5> to <34.0>." + } + ] + }, + { + "video_id": "SBT2X", + "question_id": "SBT2X", + "video_file": "Charades/videos/SBT2X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <7.0>." + } + ] + }, + { + "video_id": "SBT2X", + "question_id": "SBT2X", + "video_file": "Charades/videos/SBT2X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens the door to the garage'." + }, + { + "from": "gpt", + "value": "From <26.4> to <34.0>." + } + ] + }, + { + "video_id": "SBT2X", + "question_id": "SBT2X", + "video_file": "Charades/videos/SBT2X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes a pillow off a car' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.9> to <14.6>." + } + ] + }, + { + "video_id": "MN1MC", + "question_id": "MN1MC", + "video_file": "Charades/videos/MN1MC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats a sandwich'." + }, + { + "from": "gpt", + "value": "From <16.7> to <29.1>." + } + ] + }, + { + "video_id": "00X3U", + "question_id": "00X3U", + "video_file": "Charades/videos/00X3U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person shutting the closet door behind them'." + }, + { + "from": "gpt", + "value": "From <14.9> to <20.5>." + } + ] + }, + { + "video_id": "00X3U", + "question_id": "00X3U", + "video_file": "Charades/videos/00X3U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person grabbing a blanket leaving the closet shutting the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.9> to <20.5>." + } + ] + }, + { + "video_id": "00X3U", + "question_id": "00X3U", + "video_file": "Charades/videos/00X3U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person run away' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.1> to <22.0>." + } + ] + }, + { + "video_id": "RM1FR", + "question_id": "RM1FR", + "video_file": "Charades/videos/RM1FR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person opens a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.2> to <10.7>." + } + ] + }, + { + "video_id": "RM1FR", + "question_id": "RM1FR", + "video_file": "Charades/videos/RM1FR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a couple cabinets' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.2> to <10.7>." + } + ] + }, + { + "video_id": "RM1FR", + "question_id": "RM1FR", + "video_file": "Charades/videos/RM1FR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the cabinets' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.2> to <20.1>." + } + ] + }, + { + "video_id": "FNPEY", + "question_id": "FNPEY", + "video_file": "Charades/videos/FNPEY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person running through a hallway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "VG8QP", + "question_id": "VG8QP", + "video_file": "Charades/videos/VG8QP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pouring out the old water from the glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.3> to <27.0>." + } + ] + }, + { + "video_id": "5UNDJ", + "question_id": "5UNDJ", + "video_file": "Charades/videos/5UNDJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating some food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <98.1> to <178.9>." + } + ] + }, + { + "video_id": "5UNDJ", + "question_id": "5UNDJ", + "video_file": "Charades/videos/5UNDJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person took something to eat from out of the bowl'." + }, + { + "from": "gpt", + "value": "From <98.1> to <178.9>." + } + ] + }, + { + "video_id": "5UNDJ", + "question_id": "5UNDJ", + "video_file": "Charades/videos/5UNDJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person eats the apple'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <98.1> to <178.9>." + } + ] + }, + { + "video_id": "XDS14", + "question_id": "XDS14", + "video_file": "Charades/videos/XDS14.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes out food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.5>." + } + ] + }, + { + "video_id": "C3ZJU", + "question_id": "C3ZJU", + "video_file": "Charades/videos/C3ZJU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person sitting in a chair is eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <18.1>." + } + ] + }, + { + "video_id": "C3ZJU", + "question_id": "C3ZJU", + "video_file": "Charades/videos/C3ZJU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person is sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <9.7>." + } + ] + }, + { + "video_id": "C3ZJU", + "question_id": "C3ZJU", + "video_file": "Charades/videos/C3ZJU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a bottle' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <15.6>." + } + ] + }, + { + "video_id": "C3ZJU", + "question_id": "C3ZJU", + "video_file": "Charades/videos/C3ZJU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <18.1>." + } + ] + }, + { + "video_id": "C3ZJU", + "question_id": "C3ZJU", + "video_file": "Charades/videos/C3ZJU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person is sitting on a chair eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.5> to <18.1>." + } + ] + }, + { + "video_id": "ZL55N", + "question_id": "ZL55N", + "video_file": "Charades/videos/ZL55N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the camera down'." + }, + { + "from": "gpt", + "value": "From <5.1> to <10.3>." + } + ] + }, + { + "video_id": "ZL55N", + "question_id": "ZL55N", + "video_file": "Charades/videos/ZL55N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person close the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.3> to <13.5>." + } + ] + }, + { + "video_id": "ZL55N", + "question_id": "ZL55N", + "video_file": "Charades/videos/ZL55N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sneezing into a camera'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <9.0>." + } + ] + }, + { + "video_id": "ZL55N", + "question_id": "ZL55N", + "video_file": "Charades/videos/ZL55N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the open door behind them'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.3> to <13.5>." + } + ] + }, + { + "video_id": "ZL55N", + "question_id": "ZL55N", + "video_file": "Charades/videos/ZL55N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.3> to <13.5>." + } + ] + }, + { + "video_id": "ZL55N", + "question_id": "ZL55N", + "video_file": "Charades/videos/ZL55N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person takes a picture/video of themselves sneezing several times' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <9.0>." + } + ] + }, + { + "video_id": "GM4QR", + "question_id": "GM4QR", + "video_file": "Charades/videos/GM4QR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <9.3>." + } + ] + }, + { + "video_id": "NR94C", + "question_id": "NR94C", + "video_file": "Charades/videos/NR94C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is watching television through the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.4> to <33.0>." + } + ] + }, + { + "video_id": "MS58Y", + "question_id": "MS58Y", + "video_file": "Charades/videos/MS58Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is running through the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "MS58Y", + "question_id": "MS58Y", + "video_file": "Charades/videos/MS58Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walk through a second doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.8>." + } + ] + }, + { + "video_id": "00N38", + "question_id": "00N38", + "video_file": "Charades/videos/00N38.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <20.8>." + } + ] + }, + { + "video_id": "EBJN1", + "question_id": "EBJN1", + "video_file": "Charades/videos/EBJN1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they look out the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.7> to <36.0>." + } + ] + }, + { + "video_id": "EBJN1", + "question_id": "EBJN1", + "video_file": "Charades/videos/EBJN1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person looking out the window in the living room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.7> to <36.0>." + } + ] + }, + { + "video_id": "CYLQ0", + "question_id": "CYLQ0", + "video_file": "Charades/videos/CYLQ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person close the cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <21.1>." + } + ] + }, + { + "video_id": "CYLQ0", + "question_id": "CYLQ0", + "video_file": "Charades/videos/CYLQ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take some food from the cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.1> to <23.1>." + } + ] + }, + { + "video_id": "CYLQ0", + "question_id": "CYLQ0", + "video_file": "Charades/videos/CYLQ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <10.4>." + } + ] + }, + { + "video_id": "AWAFY", + "question_id": "AWAFY", + "video_file": "Charades/videos/AWAFY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts clothes in a bag'." + }, + { + "from": "gpt", + "value": "From <16.8> to <22.6>." + } + ] + }, + { + "video_id": "AWAFY", + "question_id": "AWAFY", + "video_file": "Charades/videos/AWAFY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening a bag of clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <10.9>." + } + ] + }, + { + "video_id": "DG33V", + "question_id": "DG33V", + "video_file": "Charades/videos/DG33V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts it on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.8> to <23.1>." + } + ] + }, + { + "video_id": "DG33V", + "question_id": "DG33V", + "video_file": "Charades/videos/DG33V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is closing a refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <19.6>." + } + ] + }, + { + "video_id": "6H4D1", + "question_id": "6H4D1", + "video_file": "Charades/videos/6H4D1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sit on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.1>." + } + ] + }, + { + "video_id": "6H4D1", + "question_id": "6H4D1", + "video_file": "Charades/videos/6H4D1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person stand up an left the room'." + }, + { + "from": "gpt", + "value": "From <44.3> to <51.0>." + } + ] + }, + { + "video_id": "6H4D1", + "question_id": "6H4D1", + "video_file": "Charades/videos/6H4D1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take off both shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.1> to <28.4>." + } + ] + }, + { + "video_id": "6H4D1", + "question_id": "6H4D1", + "video_file": "Charades/videos/6H4D1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes off their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <28.4>." + } + ] + }, + { + "video_id": "6H4D1", + "question_id": "6H4D1", + "video_file": "Charades/videos/6H4D1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person sits on a towel on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.1>." + } + ] + }, + { + "video_id": "N6R39", + "question_id": "N6R39", + "video_file": "Charades/videos/N6R39.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.0> to <27.0>." + } + ] + }, + { + "video_id": "N6R39", + "question_id": "N6R39", + "video_file": "Charades/videos/N6R39.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throwing food toward a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.0> to <29.0>." + } + ] + }, + { + "video_id": "Q3QC8", + "question_id": "Q3QC8", + "video_file": "Charades/videos/Q3QC8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts their laptop down on the sink'." + }, + { + "from": "gpt", + "value": "From <1.0> to <8.4>." + } + ] + }, + { + "video_id": "Q3QC8", + "question_id": "Q3QC8", + "video_file": "Charades/videos/Q3QC8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts a laptop down by the sink' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <8.4>." + } + ] + }, + { + "video_id": "Z04NK", + "question_id": "Z04NK", + "video_file": "Charades/videos/Z04NK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is putting a book next to a phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "Y3B48", + "question_id": "Y3B48", + "video_file": "Charades/videos/Y3B48.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes a laptop lid'." + }, + { + "from": "gpt", + "value": "From <14.1> to <26.1>." + } + ] + }, + { + "video_id": "Y3B48", + "question_id": "Y3B48", + "video_file": "Charades/videos/Y3B48.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sneezing on a mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.8>." + } + ] + }, + { + "video_id": "E27NK", + "question_id": "E27NK", + "video_file": "Charades/videos/E27NK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns the light off' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.4> to <28.3>." + } + ] + }, + { + "video_id": "E27NK", + "question_id": "E27NK", + "video_file": "Charades/videos/E27NK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person turns the light off' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.4> to <28.3>." + } + ] + }, + { + "video_id": "E27NK", + "question_id": "E27NK", + "video_file": "Charades/videos/E27NK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.2> to <29.5>." + } + ] + }, + { + "video_id": "E27NK", + "question_id": "E27NK", + "video_file": "Charades/videos/E27NK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.4> to <28.3>." + } + ] + }, + { + "video_id": "XJFMW", + "question_id": "XJFMW", + "video_file": "Charades/videos/XJFMW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person another sits with food at a small table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.6> to <32.0>." + } + ] + }, + { + "video_id": "0M1ZU", + "question_id": "0M1ZU", + "video_file": "Charades/videos/0M1ZU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a smiling person enters a bedroom sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.3> to <11.0>." + } + ] + }, + { + "video_id": "ZS2WD", + "question_id": "ZS2WD", + "video_file": "Charades/videos/ZS2WD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person throws a pillow on the couch' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <5.2>." + } + ] + }, + { + "video_id": "ZS2WD", + "question_id": "ZS2WD", + "video_file": "Charades/videos/ZS2WD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person throws a pillow towards the sofa'." + }, + { + "from": "gpt", + "value": "From <0.1> to <5.2>." + } + ] + }, + { + "video_id": "W0624", + "question_id": "W0624", + "video_file": "Charades/videos/W0624.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opening a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.1> to <30.0>." + } + ] + }, + { + "video_id": "W0624", + "question_id": "W0624", + "video_file": "Charades/videos/W0624.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.1> to <30.0>." + } + ] + }, + { + "video_id": "W0624", + "question_id": "W0624", + "video_file": "Charades/videos/W0624.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the cabinet'." + }, + { + "from": "gpt", + "value": "From <23.1> to <30.0>." + } + ] + }, + { + "video_id": "KA9T3", + "question_id": "KA9T3", + "video_file": "Charades/videos/KA9T3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person getting dressed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <19.5>." + } + ] + }, + { + "video_id": "KA9T3", + "question_id": "KA9T3", + "video_file": "Charades/videos/KA9T3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person is standing at a mirror fixing their hair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.3> to <29.6>." + } + ] + }, + { + "video_id": "JU74I", + "question_id": "JU74I", + "video_file": "Charades/videos/JU74I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person laughs at the television'." + }, + { + "from": "gpt", + "value": "From <9.8> to <19.3>." + } + ] + }, + { + "video_id": "9IO0Y", + "question_id": "9IO0Y", + "video_file": "Charades/videos/9IO0Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes their shoes off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <24.3>." + } + ] + }, + { + "video_id": "9IO0Y", + "question_id": "9IO0Y", + "video_file": "Charades/videos/9IO0Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts them on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.0> to <36.8>." + } + ] + }, + { + "video_id": "9IO0Y", + "question_id": "9IO0Y", + "video_file": "Charades/videos/9IO0Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts their shoes on the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.0> to <36.8>." + } + ] + }, + { + "video_id": "9IO0Y", + "question_id": "9IO0Y", + "video_file": "Charades/videos/9IO0Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <24.3>." + } + ] + }, + { + "video_id": "OM66H", + "question_id": "OM66H", + "video_file": "Charades/videos/OM66H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person but they start sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.3> to <28.5>." + } + ] + }, + { + "video_id": "ITJIY", + "question_id": "ITJIY", + "video_file": "Charades/videos/ITJIY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.9> to <14.3>." + } + ] + }, + { + "video_id": "GW2Y8", + "question_id": "GW2Y8", + "video_file": "Charades/videos/GW2Y8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking a glass of water'." + }, + { + "from": "gpt", + "value": "From <23.4> to <33.0>." + } + ] + }, + { + "video_id": "GW2Y8", + "question_id": "GW2Y8", + "video_file": "Charades/videos/GW2Y8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is playing on their laptop sitting down'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.9>." + } + ] + }, + { + "video_id": "GW2Y8", + "question_id": "GW2Y8", + "video_file": "Charades/videos/GW2Y8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinks from a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.4> to <33.0>." + } + ] + }, + { + "video_id": "GW2Y8", + "question_id": "GW2Y8", + "video_file": "Charades/videos/GW2Y8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes a sandwich from a dish' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.9> to <19.8>." + } + ] + }, + { + "video_id": "GW2Y8", + "question_id": "GW2Y8", + "video_file": "Charades/videos/GW2Y8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is playing a game on their laptop'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.9>." + } + ] + }, + { + "video_id": "1UHL7", + "question_id": "1UHL7", + "video_file": "Charades/videos/1UHL7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person undresses in front of a mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.2>." + } + ] + }, + { + "video_id": "1UHL7", + "question_id": "1UHL7", + "video_file": "Charades/videos/1UHL7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is undressing in front of a mirror'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.2>." + } + ] + }, + { + "video_id": "1UHL7", + "question_id": "1UHL7", + "video_file": "Charades/videos/1UHL7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening its door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <20.1>." + } + ] + }, + { + "video_id": "1UHL7", + "question_id": "1UHL7", + "video_file": "Charades/videos/1UHL7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person close the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.5> to <32.0>." + } + ] + }, + { + "video_id": "1UHL7", + "question_id": "1UHL7", + "video_file": "Charades/videos/1UHL7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put their shoes in a closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.6> to <22.0>." + } + ] + }, + { + "video_id": "D2ES2", + "question_id": "D2ES2", + "video_file": "Charades/videos/D2ES2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is walking through a doorway'." + }, + { + "from": "gpt", + "value": "From <10.7> to <15.6>." + } + ] + }, + { + "video_id": "IHSRC", + "question_id": "IHSRC", + "video_file": "Charades/videos/IHSRC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person laughs at the picture'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.1> to <32.0>." + } + ] + }, + { + "video_id": "IHSRC", + "question_id": "IHSRC", + "video_file": "Charades/videos/IHSRC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person looks at a picture on the wall' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <32.0>." + } + ] + }, + { + "video_id": "J1RLO", + "question_id": "J1RLO", + "video_file": "Charades/videos/J1RLO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they close the door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "NQPM5", + "question_id": "NQPM5", + "video_file": "Charades/videos/NQPM5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks out through the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <21.0>." + } + ] + }, + { + "video_id": "NQPM5", + "question_id": "NQPM5", + "video_file": "Charades/videos/NQPM5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walk through the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <21.0>." + } + ] + }, + { + "video_id": "RGE8C", + "question_id": "RGE8C", + "video_file": "Charades/videos/RGE8C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a look a clothes in a washer' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <17.1>." + } + ] + }, + { + "video_id": "RGE8C", + "question_id": "RGE8C", + "video_file": "Charades/videos/RGE8C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts back on their shoes'." + }, + { + "from": "gpt", + "value": "From <16.4> to <28.9>." + } + ] + }, + { + "video_id": "NALYZ", + "question_id": "NALYZ", + "video_file": "Charades/videos/NALYZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they stand up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.1> to <28.9>." + } + ] + }, + { + "video_id": "NALYZ", + "question_id": "NALYZ", + "video_file": "Charades/videos/NALYZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinks something form a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <26.2>." + } + ] + }, + { + "video_id": "P4MBS", + "question_id": "P4MBS", + "video_file": "Charades/videos/P4MBS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.0> to <27.2>." + } + ] + }, + { + "video_id": "P4MBS", + "question_id": "P4MBS", + "video_file": "Charades/videos/P4MBS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats something out of a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.6> to <6.2>." + } + ] + }, + { + "video_id": "P4MBS", + "question_id": "P4MBS", + "video_file": "Charades/videos/P4MBS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is wrapped in a blanket eating something' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <6.2>." + } + ] + }, + { + "video_id": "P4MBS", + "question_id": "P4MBS", + "video_file": "Charades/videos/P4MBS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they stand up abruptly' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.8> to <25.8>." + } + ] + }, + { + "video_id": "P4MBS", + "question_id": "P4MBS", + "video_file": "Charades/videos/P4MBS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a snack'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <6.2>." + } + ] + }, + { + "video_id": "P4MBS", + "question_id": "P4MBS", + "video_file": "Charades/videos/P4MBS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close the lid of the laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.0> to <27.2>." + } + ] + }, + { + "video_id": "FXC28", + "question_id": "FXC28", + "video_file": "Charades/videos/FXC28.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throw a blanket in the air repeatedly' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.3> to <30.1>." + } + ] + }, + { + "video_id": "924QD", + "question_id": "924QD", + "video_file": "Charades/videos/924QD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is running with a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.9> to <26.5>." + } + ] + }, + { + "video_id": "1H6PS", + "question_id": "1H6PS", + "video_file": "Charades/videos/1H6PS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person sneezes on the sandwich repeatedly'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "1H6PS", + "question_id": "1H6PS", + "video_file": "Charades/videos/1H6PS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person walks through the doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.2> to <26.0>." + } + ] + }, + { + "video_id": "1H6PS", + "question_id": "1H6PS", + "video_file": "Charades/videos/1H6PS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "1H6PS", + "question_id": "1H6PS", + "video_file": "Charades/videos/1H6PS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person eats some food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "QM4AD", + "question_id": "QM4AD", + "video_file": "Charades/videos/QM4AD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put the glass down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <23.4>." + } + ] + }, + { + "video_id": "O1LOW", + "question_id": "O1LOW", + "video_file": "Charades/videos/O1LOW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put objects on a table in an entryway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.1> to <28.1>." + } + ] + }, + { + "video_id": "SSH5I", + "question_id": "SSH5I", + "video_file": "Charades/videos/SSH5I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person is eating'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.3>." + } + ] + }, + { + "video_id": "HYR9Q", + "question_id": "HYR9Q", + "video_file": "Charades/videos/HYR9Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts their laptop on a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <9.1>." + } + ] + }, + { + "video_id": "HYR9Q", + "question_id": "HYR9Q", + "video_file": "Charades/videos/HYR9Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person puts a laptop on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <9.1>." + } + ] + }, + { + "video_id": "HYR9Q", + "question_id": "HYR9Q", + "video_file": "Charades/videos/HYR9Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.5> to <13.9>." + } + ] + }, + { + "video_id": "HYR9Q", + "question_id": "HYR9Q", + "video_file": "Charades/videos/HYR9Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take out some food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <19.6>." + } + ] + }, + { + "video_id": "75RPN", + "question_id": "75RPN", + "video_file": "Charades/videos/75RPN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.2> to <16.8>." + } + ] + }, + { + "video_id": "0ZOPH", + "question_id": "0ZOPH", + "video_file": "Charades/videos/0ZOPH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person throws a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.7> to <10.9>." + } + ] + }, + { + "video_id": "KTBKA", + "question_id": "KTBKA", + "video_file": "Charades/videos/KTBKA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opens the medicine cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.0> to <16.4>." + } + ] + }, + { + "video_id": "KTBKA", + "question_id": "KTBKA", + "video_file": "Charades/videos/KTBKA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the medicine cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <16.4>." + } + ] + }, + { + "video_id": "I164P", + "question_id": "I164P", + "video_file": "Charades/videos/I164P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person walks out of the house through the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.6> to <24.4>." + } + ] + }, + { + "video_id": "I164P", + "question_id": "I164P", + "video_file": "Charades/videos/I164P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <21.5>." + } + ] + }, + { + "video_id": "QK5KC", + "question_id": "QK5KC", + "video_file": "Charades/videos/QK5KC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <5.4>." + } + ] + }, + { + "video_id": "QK5KC", + "question_id": "QK5KC", + "video_file": "Charades/videos/QK5KC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person looking at a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <22.4>." + } + ] + }, + { + "video_id": "QK5KC", + "question_id": "QK5KC", + "video_file": "Charades/videos/QK5KC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting in a chair smiling' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.4> to <22.2>." + } + ] + }, + { + "video_id": "2LCLG", + "question_id": "2LCLG", + "video_file": "Charades/videos/2LCLG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throws the book from it on to a bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <17.2>." + } + ] + }, + { + "video_id": "2LCLG", + "question_id": "2LCLG", + "video_file": "Charades/videos/2LCLG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes out a book grabs it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.5> to <20.0>." + } + ] + }, + { + "video_id": "2VNXF", + "question_id": "2VNXF", + "video_file": "Charades/videos/2VNXF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts the dishes back on the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.6> to <22.2>." + } + ] + }, + { + "video_id": "2VNXF", + "question_id": "2VNXF", + "video_file": "Charades/videos/2VNXF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes their glass of coffee off the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "OFPAE", + "question_id": "OFPAE", + "video_file": "Charades/videos/OFPAE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens on a sofa'." + }, + { + "from": "gpt", + "value": "From <2.0> to <9.2>." + } + ] + }, + { + "video_id": "OWAA9", + "question_id": "OWAA9", + "video_file": "Charades/videos/OWAA9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.4> to <28.2>." + } + ] + }, + { + "video_id": "OWAA9", + "question_id": "OWAA9", + "video_file": "Charades/videos/OWAA9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a door walks out of the room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.3> to <26.8>." + } + ] + }, + { + "video_id": "OWAA9", + "question_id": "OWAA9", + "video_file": "Charades/videos/OWAA9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closing the door behind them' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.4> to <28.2>." + } + ] + }, + { + "video_id": "9AFQ7", + "question_id": "9AFQ7", + "video_file": "Charades/videos/9AFQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a closet'." + }, + { + "from": "gpt", + "value": "From <7.8> to <14.2>." + } + ] + }, + { + "video_id": "9AFQ7", + "question_id": "9AFQ7", + "video_file": "Charades/videos/9AFQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking a glass of water'." + }, + { + "from": "gpt", + "value": "From <24.8> to <33.2>." + } + ] + }, + { + "video_id": "9AFQ7", + "question_id": "9AFQ7", + "video_file": "Charades/videos/9AFQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'this person closes a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.1> to <10.5>." + } + ] + }, + { + "video_id": "1AR0K", + "question_id": "1AR0K", + "video_file": "Charades/videos/1AR0K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person walks into the bathroom holding a broom'." + }, + { + "from": "gpt", + "value": "From <11.1> to <18.6>." + } + ] + }, + { + "video_id": "1AR0K", + "question_id": "1AR0K", + "video_file": "Charades/videos/1AR0K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts the broom down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.6> to <20.7>." + } + ] + }, + { + "video_id": "5OLVC", + "question_id": "5OLVC", + "video_file": "Charades/videos/5OLVC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pours a glass of water from the sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.2> to <19.8>." + } + ] + }, + { + "video_id": "NO9MI", + "question_id": "NO9MI", + "video_file": "Charades/videos/NO9MI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.8> to <25.0>." + } + ] + }, + { + "video_id": "NO9MI", + "question_id": "NO9MI", + "video_file": "Charades/videos/NO9MI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.8> to <25.0>." + } + ] + }, + { + "video_id": "0OMFD", + "question_id": "0OMFD", + "video_file": "Charades/videos/0OMFD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person getting dressed'." + }, + { + "from": "gpt", + "value": "From <2.8> to <8.9>." + } + ] + }, + { + "video_id": "0OMFD", + "question_id": "0OMFD", + "video_file": "Charades/videos/0OMFD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they get dressed'." + }, + { + "from": "gpt", + "value": "From <2.8> to <8.9>." + } + ] + }, + { + "video_id": "IOBKP", + "question_id": "IOBKP", + "video_file": "Charades/videos/IOBKP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take a drink from a coffee cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.7> to <28.3>." + } + ] + }, + { + "video_id": "IOBKP", + "question_id": "IOBKP", + "video_file": "Charades/videos/IOBKP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is snuggling in bed with a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <13.5>." + } + ] + }, + { + "video_id": "EXG45", + "question_id": "EXG45", + "video_file": "Charades/videos/EXG45.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the book away'." + }, + { + "from": "gpt", + "value": "From <24.4> to <29.8>." + } + ] + }, + { + "video_id": "EXG45", + "question_id": "EXG45", + "video_file": "Charades/videos/EXG45.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person closes a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.2> to <28.8>." + } + ] + }, + { + "video_id": "EXG45", + "question_id": "EXG45", + "video_file": "Charades/videos/EXG45.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts away the book used for references' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.4> to <29.8>." + } + ] + }, + { + "video_id": "M0KKW", + "question_id": "M0KKW", + "video_file": "Charades/videos/M0KKW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is holding a broom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "M0KKW", + "question_id": "M0KKW", + "video_file": "Charades/videos/M0KKW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.6> to <12.2>." + } + ] + }, + { + "video_id": "M0KKW", + "question_id": "M0KKW", + "video_file": "Charades/videos/M0KKW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.1> to <15.5>." + } + ] + }, + { + "video_id": "RXBUJ", + "question_id": "RXBUJ", + "video_file": "Charades/videos/RXBUJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person starts closing a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <13.3>." + } + ] + }, + { + "video_id": "D9F5V", + "question_id": "D9F5V", + "video_file": "Charades/videos/D9F5V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a bottle'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.1> to <23.5>." + } + ] + }, + { + "video_id": "JYBGS", + "question_id": "JYBGS", + "video_file": "Charades/videos/JYBGS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs toward the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <13.3>." + } + ] + }, + { + "video_id": "JYBGS", + "question_id": "JYBGS", + "video_file": "Charades/videos/JYBGS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person runs out of the hallway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <13.3>." + } + ] + }, + { + "video_id": "JYBGS", + "question_id": "JYBGS", + "video_file": "Charades/videos/JYBGS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks into a hallway holding a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <9.1>." + } + ] + }, + { + "video_id": "5J91T", + "question_id": "5J91T", + "video_file": "Charades/videos/5J91T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a medicine cabinet in the laundry room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "5J91T", + "question_id": "5J91T", + "video_file": "Charades/videos/5J91T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a medicine cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "TC015", + "question_id": "TC015", + "video_file": "Charades/videos/TC015.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person puts blanket on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.4> to <9.0>." + } + ] + }, + { + "video_id": "D2ORL", + "question_id": "D2ORL", + "video_file": "Charades/videos/D2ORL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is on the phone laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <30.8> to <41.0>." + } + ] + }, + { + "video_id": "509W9", + "question_id": "509W9", + "video_file": "Charades/videos/509W9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person snuggles with a pillow in a very creepy way'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.8> to <25.4>." + } + ] + }, + { + "video_id": "509W9", + "question_id": "509W9", + "video_file": "Charades/videos/509W9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person snuggles up to a pillow on a sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.8> to <25.4>." + } + ] + }, + { + "video_id": "509W9", + "question_id": "509W9", + "video_file": "Charades/videos/509W9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person sits down on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <12.2>." + } + ] + }, + { + "video_id": "509W9", + "question_id": "509W9", + "video_file": "Charades/videos/509W9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sits in a chair snuggling with a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.8> to <25.4>." + } + ] + }, + { + "video_id": "4C1KY", + "question_id": "4C1KY", + "video_file": "Charades/videos/4C1KY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts down the book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <30.1> to <35.4>." + } + ] + }, + { + "video_id": "PYIA3", + "question_id": "PYIA3", + "video_file": "Charades/videos/PYIA3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person held the refrigerator door open' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <7.0>." + } + ] + }, + { + "video_id": "0YTSJ", + "question_id": "0YTSJ", + "video_file": "Charades/videos/0YTSJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes off their glasses' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.7> to <28.7>." + } + ] + }, + { + "video_id": "0YTSJ", + "question_id": "0YTSJ", + "video_file": "Charades/videos/0YTSJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person are drinking a glass of water'." + }, + { + "from": "gpt", + "value": "From <22.7> to <34.1>." + } + ] + }, + { + "video_id": "0YTSJ", + "question_id": "0YTSJ", + "video_file": "Charades/videos/0YTSJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they reach for a glass to take a drink'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.7> to <28.7>." + } + ] + }, + { + "video_id": "0YTSJ", + "question_id": "0YTSJ", + "video_file": "Charades/videos/0YTSJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.7> to <38.0>." + } + ] + }, + { + "video_id": "0YTSJ", + "question_id": "0YTSJ", + "video_file": "Charades/videos/0YTSJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes a bite from the sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.7> to <36.3>." + } + ] + }, + { + "video_id": "GFON7", + "question_id": "GFON7", + "video_file": "Charades/videos/GFON7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is laughing over the sink' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.5>." + } + ] + }, + { + "video_id": "GFON7", + "question_id": "GFON7", + "video_file": "Charades/videos/GFON7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person is laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.5>." + } + ] + }, + { + "video_id": "MH601", + "question_id": "MH601", + "video_file": "Charades/videos/MH601.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.1>." + } + ] + }, + { + "video_id": "MH601", + "question_id": "MH601", + "video_file": "Charades/videos/MH601.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "N7243", + "question_id": "N7243", + "video_file": "Charades/videos/N7243.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the picture in the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <11.8>." + } + ] + }, + { + "video_id": "Q18ZQ", + "question_id": "Q18ZQ", + "video_file": "Charades/videos/Q18ZQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.5> to <13.4>." + } + ] + }, + { + "video_id": "Q18ZQ", + "question_id": "Q18ZQ", + "video_file": "Charades/videos/Q18ZQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.3> to <41.4>." + } + ] + }, + { + "video_id": "Q18ZQ", + "question_id": "Q18ZQ", + "video_file": "Charades/videos/Q18ZQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is holding a towel' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.0>." + } + ] + }, + { + "video_id": "Q18ZQ", + "question_id": "Q18ZQ", + "video_file": "Charades/videos/Q18ZQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put the towel' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.3> to <19.6>." + } + ] + }, + { + "video_id": "GLIC0", + "question_id": "GLIC0", + "video_file": "Charades/videos/GLIC0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the bag down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <28.7>." + } + ] + }, + { + "video_id": "GLIC0", + "question_id": "GLIC0", + "video_file": "Charades/videos/GLIC0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is holding a bag of clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.1> to <26.5>." + } + ] + }, + { + "video_id": "9F691", + "question_id": "9F691", + "video_file": "Charades/videos/9F691.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a young woman sitting in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "HA02S", + "question_id": "HA02S", + "video_file": "Charades/videos/HA02S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "HA02S", + "question_id": "HA02S", + "video_file": "Charades/videos/HA02S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding the broom deciding where to put it'." + }, + { + "from": "gpt", + "value": "From <31.4> to <38.0>." + } + ] + }, + { + "video_id": "HA02S", + "question_id": "HA02S", + "video_file": "Charades/videos/HA02S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the broom inside' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.4> to <38.0>." + } + ] + }, + { + "video_id": "HA02S", + "question_id": "HA02S", + "video_file": "Charades/videos/HA02S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.7> to <38.0>." + } + ] + }, + { + "video_id": "6R7QW", + "question_id": "6R7QW", + "video_file": "Charades/videos/6R7QW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <18.1>." + } + ] + }, + { + "video_id": "6R7QW", + "question_id": "6R7QW", + "video_file": "Charades/videos/6R7QW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats something'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.5> to <37.0>." + } + ] + }, + { + "video_id": "6R7QW", + "question_id": "6R7QW", + "video_file": "Charades/videos/6R7QW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins to eat some food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.5> to <37.0>." + } + ] + }, + { + "video_id": "VLE11", + "question_id": "VLE11", + "video_file": "Charades/videos/VLE11.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open up their homework book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.2> to <29.7>." + } + ] + }, + { + "video_id": "Z1H81", + "question_id": "Z1H81", + "video_file": "Charades/videos/Z1H81.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person in their garage walks to the doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "Z1H81", + "question_id": "Z1H81", + "video_file": "Charades/videos/Z1H81.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns on the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.0> to <13.1>." + } + ] + }, + { + "video_id": "VUSFK", + "question_id": "VUSFK", + "video_file": "Charades/videos/VUSFK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.7> to <22.9>." + } + ] + }, + { + "video_id": "Q5Y02", + "question_id": "Q5Y02", + "video_file": "Charades/videos/Q5Y02.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking a glass of water'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "Q5Y02", + "question_id": "Q5Y02", + "video_file": "Charades/videos/Q5Y02.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person look out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.7> to <25.7>." + } + ] + }, + { + "video_id": "Q5Y02", + "question_id": "Q5Y02", + "video_file": "Charades/videos/Q5Y02.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put down a glass they were drinking from' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "Q5Y02", + "question_id": "Q5Y02", + "video_file": "Charades/videos/Q5Y02.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks through the doorway into their bedroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "TRFB0", + "question_id": "TRFB0", + "video_file": "Charades/videos/TRFB0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person washes their hands in a sink'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.1>." + } + ] + }, + { + "video_id": "TRFB0", + "question_id": "TRFB0", + "video_file": "Charades/videos/TRFB0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they open the medicine cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.7> to <17.8>." + } + ] + }, + { + "video_id": "TRFB0", + "question_id": "TRFB0", + "video_file": "Charades/videos/TRFB0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is washing their hands' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.1>." + } + ] + }, + { + "video_id": "TRFB0", + "question_id": "TRFB0", + "video_file": "Charades/videos/TRFB0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person proceeds to wash hands'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.1>." + } + ] + }, + { + "video_id": "TRFB0", + "question_id": "TRFB0", + "video_file": "Charades/videos/TRFB0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.7> to <17.8>." + } + ] + }, + { + "video_id": "36EK5", + "question_id": "36EK5", + "video_file": "Charades/videos/36EK5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person is walking through the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.7> to <10.6>." + } + ] + }, + { + "video_id": "36EK5", + "question_id": "36EK5", + "video_file": "Charades/videos/36EK5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is walking through a doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.7> to <10.6>." + } + ] + }, + { + "video_id": "QBXRP", + "question_id": "QBXRP", + "video_file": "Charades/videos/QBXRP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person pours a glass of water down the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.6> to <11.0>." + } + ] + }, + { + "video_id": "5QW1X", + "question_id": "5QW1X", + "video_file": "Charades/videos/5QW1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.0> to <30.0>." + } + ] + }, + { + "video_id": "5QW1X", + "question_id": "5QW1X", + "video_file": "Charades/videos/5QW1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is opening a book'." + }, + { + "from": "gpt", + "value": "From <7.0> to <13.4>." + } + ] + }, + { + "video_id": "5QW1X", + "question_id": "5QW1X", + "video_file": "Charades/videos/5QW1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they open a book'." + }, + { + "from": "gpt", + "value": "From <7.0> to <13.4>." + } + ] + }, + { + "video_id": "BJXRT", + "question_id": "BJXRT", + "video_file": "Charades/videos/BJXRT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits in a chair hugging the blanket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.1> to <36.0>." + } + ] + }, + { + "video_id": "BJXRT", + "question_id": "BJXRT", + "video_file": "Charades/videos/BJXRT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begin sitting in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.1> to <36.0>." + } + ] + }, + { + "video_id": "LRPBA", + "question_id": "LRPBA", + "video_file": "Charades/videos/LRPBA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a smiling person runs into their closet holding a mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.6>." + } + ] + }, + { + "video_id": "LRPBA", + "question_id": "LRPBA", + "video_file": "Charades/videos/LRPBA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinks from a glass in a closet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.7> to <13.3>." + } + ] + }, + { + "video_id": "LRPBA", + "question_id": "LRPBA", + "video_file": "Charades/videos/LRPBA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs in place' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.6>." + } + ] + }, + { + "video_id": "1T2W4", + "question_id": "1T2W4", + "video_file": "Charades/videos/1T2W4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws clothes into a washing machine' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.5> to <28.0>." + } + ] + }, + { + "video_id": "1T2W4", + "question_id": "1T2W4", + "video_file": "Charades/videos/1T2W4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throw clothes in the washer'." + }, + { + "from": "gpt", + "value": "From <22.5> to <28.0>." + } + ] + }, + { + "video_id": "YJQO3", + "question_id": "YJQO3", + "video_file": "Charades/videos/YJQO3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts their phone in a box'." + }, + { + "from": "gpt", + "value": "From <22.7> to <32.8>." + } + ] + }, + { + "video_id": "YJQO3", + "question_id": "YJQO3", + "video_file": "Charades/videos/YJQO3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a phone of their pocket'." + }, + { + "from": "gpt", + "value": "From <23.3> to <28.0>." + } + ] + }, + { + "video_id": "H2B7U", + "question_id": "H2B7U", + "video_file": "Charades/videos/H2B7U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is holding clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "H2B7U", + "question_id": "H2B7U", + "video_file": "Charades/videos/H2B7U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person shoves clothes into the washing machine' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "XHL3Z", + "question_id": "XHL3Z", + "video_file": "Charades/videos/XHL3Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes some medicine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.4> to <43.0>." + } + ] + }, + { + "video_id": "XHL3Z", + "question_id": "XHL3Z", + "video_file": "Charades/videos/XHL3Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take some medicine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.4> to <43.0>." + } + ] + }, + { + "video_id": "PPUOO", + "question_id": "PPUOO", + "video_file": "Charades/videos/PPUOO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person grabs a towel to put it away' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.2> to <16.7>." + } + ] + }, + { + "video_id": "PPUOO", + "question_id": "PPUOO", + "video_file": "Charades/videos/PPUOO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.8> to <13.8>." + } + ] + }, + { + "video_id": "PPUOO", + "question_id": "PPUOO", + "video_file": "Charades/videos/PPUOO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is opening a door to the bathroom'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "PPUOO", + "question_id": "PPUOO", + "video_file": "Charades/videos/PPUOO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts throwing a towel inside' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.2> to <19.5>." + } + ] + }, + { + "video_id": "PPUOO", + "question_id": "PPUOO", + "video_file": "Charades/videos/PPUOO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person grasps the doorknob to open the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "52D5Q", + "question_id": "52D5Q", + "video_file": "Charades/videos/52D5Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is laughing into a phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "52D5Q", + "question_id": "52D5Q", + "video_file": "Charades/videos/52D5Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is talking on the phone'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "8Z5SJ", + "question_id": "8Z5SJ", + "video_file": "Charades/videos/8Z5SJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person closes a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.2> to <18.8>." + } + ] + }, + { + "video_id": "8Z5SJ", + "question_id": "8Z5SJ", + "video_file": "Charades/videos/8Z5SJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes a window'." + }, + { + "from": "gpt", + "value": "From <4.2> to <18.8>." + } + ] + }, + { + "video_id": "8Z5SJ", + "question_id": "8Z5SJ", + "video_file": "Charades/videos/8Z5SJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person in the bathroom is closing a window'." + }, + { + "from": "gpt", + "value": "From <4.2> to <18.8>." + } + ] + }, + { + "video_id": "GHARQ", + "question_id": "GHARQ", + "video_file": "Charades/videos/GHARQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating from a plate balanced on a pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <22.7>." + } + ] + }, + { + "video_id": "GHARQ", + "question_id": "GHARQ", + "video_file": "Charades/videos/GHARQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a pillow standing next to a broom'." + }, + { + "from": "gpt", + "value": "From <19.0> to <25.6>." + } + ] + }, + { + "video_id": "GHARQ", + "question_id": "GHARQ", + "video_file": "Charades/videos/GHARQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is eating a sandwich on a plate'." + }, + { + "from": "gpt", + "value": "From <9.9> to <22.7>." + } + ] + }, + { + "video_id": "TZ27E", + "question_id": "TZ27E", + "video_file": "Charades/videos/TZ27E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a glass'." + }, + { + "from": "gpt", + "value": "From <9.0> to <14.0>." + } + ] + }, + { + "video_id": "TZ27E", + "question_id": "TZ27E", + "video_file": "Charades/videos/TZ27E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is putting a laptop on a bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.1> to <10.8>." + } + ] + }, + { + "video_id": "BALXZ", + "question_id": "BALXZ", + "video_file": "Charades/videos/BALXZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs down some stairs holding a phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "BALXZ", + "question_id": "BALXZ", + "video_file": "Charades/videos/BALXZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is running with their phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "BALXZ", + "question_id": "BALXZ", + "video_file": "Charades/videos/BALXZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <10.0>." + } + ] + }, + { + "video_id": "BALXZ", + "question_id": "BALXZ", + "video_file": "Charades/videos/BALXZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open the door'." + }, + { + "from": "gpt", + "value": "From <7.0> to <13.0>." + } + ] + }, + { + "video_id": "BALXZ", + "question_id": "BALXZ", + "video_file": "Charades/videos/BALXZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.0> to <13.0>." + } + ] + }, + { + "video_id": "56ASU", + "question_id": "56ASU", + "video_file": "Charades/videos/56ASU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks through the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <6.9>." + } + ] + }, + { + "video_id": "56ASU", + "question_id": "56ASU", + "video_file": "Charades/videos/56ASU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns on the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.6> to <9.2>." + } + ] + }, + { + "video_id": "56ASU", + "question_id": "56ASU", + "video_file": "Charades/videos/56ASU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person in a laundry room turns on a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.6> to <9.2>." + } + ] + }, + { + "video_id": "PDRVY", + "question_id": "PDRVY", + "video_file": "Charades/videos/PDRVY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.8> to <25.0>." + } + ] + }, + { + "video_id": "PDRVY", + "question_id": "PDRVY", + "video_file": "Charades/videos/PDRVY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes the bag into the room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <16.4>." + } + ] + }, + { + "video_id": "PDRVY", + "question_id": "PDRVY", + "video_file": "Charades/videos/PDRVY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person open the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <12.8>." + } + ] + }, + { + "video_id": "YRX5H", + "question_id": "YRX5H", + "video_file": "Charades/videos/YRX5H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is standing in the entryway eating some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.9> to <19.9>." + } + ] + }, + { + "video_id": "YRX5H", + "question_id": "YRX5H", + "video_file": "Charades/videos/YRX5H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <6.6>." + } + ] + }, + { + "video_id": "YRX5H", + "question_id": "YRX5H", + "video_file": "Charades/videos/YRX5H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats something from a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.9> to <19.9>." + } + ] + }, + { + "video_id": "YRX5H", + "question_id": "YRX5H", + "video_file": "Charades/videos/YRX5H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts down the food to grab the vacuum' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <36.5> to <39.0>." + } + ] + }, + { + "video_id": "YRX5H", + "question_id": "YRX5H", + "video_file": "Charades/videos/YRX5H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they close the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.8> to <21.7>." + } + ] + }, + { + "video_id": "YRX5H", + "question_id": "YRX5H", + "video_file": "Charades/videos/YRX5H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a door open' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <6.6>." + } + ] + }, + { + "video_id": "YRX5H", + "question_id": "YRX5H", + "video_file": "Charades/videos/YRX5H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is holding a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "FH9AB", + "question_id": "FH9AB", + "video_file": "Charades/videos/FH9AB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <29.4> to <36.0>." + } + ] + }, + { + "video_id": "O7N6K", + "question_id": "O7N6K", + "video_file": "Charades/videos/O7N6K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes off their shoes'." + }, + { + "from": "gpt", + "value": "From <2.6> to <7.7>." + } + ] + }, + { + "video_id": "O7N6K", + "question_id": "O7N6K", + "video_file": "Charades/videos/O7N6K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens the door to their bedroom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "O7N6K", + "question_id": "O7N6K", + "video_file": "Charades/videos/O7N6K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "O7N6K", + "question_id": "O7N6K", + "video_file": "Charades/videos/O7N6K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes their shoes off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.6> to <7.7>." + } + ] + }, + { + "video_id": "O7N6K", + "question_id": "O7N6K", + "video_file": "Charades/videos/O7N6K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.6>." + } + ] + }, + { + "video_id": "QAHJG", + "question_id": "QAHJG", + "video_file": "Charades/videos/QAHJG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens their laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <8.7>." + } + ] + }, + { + "video_id": "7N4TW", + "question_id": "7N4TW", + "video_file": "Charades/videos/7N4TW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person was in their bedroom undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <27.7>." + } + ] + }, + { + "video_id": "7N4TW", + "question_id": "7N4TW", + "video_file": "Charades/videos/7N4TW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person put some bedtime clothes on' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <58.5> to <68.6>." + } + ] + }, + { + "video_id": "7N4TW", + "question_id": "7N4TW", + "video_file": "Charades/videos/7N4TW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person gets undressed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.5> to <27.7>." + } + ] + }, + { + "video_id": "7N4TW", + "question_id": "7N4TW", + "video_file": "Charades/videos/7N4TW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they take their shoes off' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <37.2> to <45.1>." + } + ] + }, + { + "video_id": "7N4TW", + "question_id": "7N4TW", + "video_file": "Charades/videos/7N4TW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person get dressed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <38.0> to <61.6>." + } + ] + }, + { + "video_id": "JUCOK", + "question_id": "JUCOK", + "video_file": "Charades/videos/JUCOK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.4> to <28.9>." + } + ] + }, + { + "video_id": "Y05U8", + "question_id": "Y05U8", + "video_file": "Charades/videos/Y05U8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins dressing to go outside'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.0> to <34.0>." + } + ] + }, + { + "video_id": "Y05U8", + "question_id": "Y05U8", + "video_file": "Charades/videos/Y05U8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes out their phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.1> to <25.4>." + } + ] + }, + { + "video_id": "IB987", + "question_id": "IB987", + "video_file": "Charades/videos/IB987.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts throwing clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.7> to <26.1>." + } + ] + }, + { + "video_id": "NZX0U", + "question_id": "NZX0U", + "video_file": "Charades/videos/NZX0U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is cooking on the stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.9> to <13.4>." + } + ] + }, + { + "video_id": "NZX0U", + "question_id": "NZX0U", + "video_file": "Charades/videos/NZX0U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person cooking some food on the stove' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.9> to <13.4>." + } + ] + }, + { + "video_id": "5ZXA3", + "question_id": "5ZXA3", + "video_file": "Charades/videos/5ZXA3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes a laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.9> to <32.0>." + } + ] + }, + { + "video_id": "R1B8I", + "question_id": "R1B8I", + "video_file": "Charades/videos/R1B8I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holds a broom'." + }, + { + "from": "gpt", + "value": "From <15.8> to <32.2>." + } + ] + }, + { + "video_id": "CSYES", + "question_id": "CSYES", + "video_file": "Charades/videos/CSYES.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts them on the table'." + }, + { + "from": "gpt", + "value": "From <9.6> to <17.5>." + } + ] + }, + { + "video_id": "CSYES", + "question_id": "CSYES", + "video_file": "Charades/videos/CSYES.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes out their phone'." + }, + { + "from": "gpt", + "value": "From <10.7> to <17.0>." + } + ] + }, + { + "video_id": "5X5DT", + "question_id": "5X5DT", + "video_file": "Charades/videos/5X5DT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close the laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <23.8>." + } + ] + }, + { + "video_id": "5X5DT", + "question_id": "5X5DT", + "video_file": "Charades/videos/5X5DT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a laptop on their lap' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.8>." + } + ] + }, + { + "video_id": "5X5DT", + "question_id": "5X5DT", + "video_file": "Charades/videos/5X5DT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.6> to <23.8>." + } + ] + }, + { + "video_id": "MFQ5S", + "question_id": "MFQ5S", + "video_file": "Charades/videos/MFQ5S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.2> to <35.1>." + } + ] + }, + { + "video_id": "ZYVTC", + "question_id": "ZYVTC", + "video_file": "Charades/videos/ZYVTC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a book to read' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.1> to <30.9>." + } + ] + }, + { + "video_id": "ZYVTC", + "question_id": "ZYVTC", + "video_file": "Charades/videos/ZYVTC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.4> to <16.3>." + } + ] + }, + { + "video_id": "ZYVTC", + "question_id": "ZYVTC", + "video_file": "Charades/videos/ZYVTC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the food down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.5> to <16.9>." + } + ] + }, + { + "video_id": "Z6HEA", + "question_id": "Z6HEA", + "video_file": "Charades/videos/Z6HEA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats a sandwich'." + }, + { + "from": "gpt", + "value": "From <20.4> to <26.9>." + } + ] + }, + { + "video_id": "Z6HEA", + "question_id": "Z6HEA", + "video_file": "Charades/videos/Z6HEA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin to eat' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.4> to <26.9>." + } + ] + }, + { + "video_id": "Z6HEA", + "question_id": "Z6HEA", + "video_file": "Charades/videos/Z6HEA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person goes into the laundry room to wash some clothes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.6> to <11.4>." + } + ] + }, + { + "video_id": "DFPOJ", + "question_id": "DFPOJ", + "video_file": "Charades/videos/DFPOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.9> to <19.2>." + } + ] + }, + { + "video_id": "DFPOJ", + "question_id": "DFPOJ", + "video_file": "Charades/videos/DFPOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person was reading a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.4> to <31.0>." + } + ] + }, + { + "video_id": "DFPOJ", + "question_id": "DFPOJ", + "video_file": "Charades/videos/DFPOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is looking at a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.4> to <31.0>." + } + ] + }, + { + "video_id": "DFPOJ", + "question_id": "DFPOJ", + "video_file": "Charades/videos/DFPOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.9> to <19.2>." + } + ] + }, + { + "video_id": "AJLGU", + "question_id": "AJLGU", + "video_file": "Charades/videos/AJLGU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they open a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <18.6>." + } + ] + }, + { + "video_id": "AJLGU", + "question_id": "AJLGU", + "video_file": "Charades/videos/AJLGU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they open the cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <18.6>." + } + ] + }, + { + "video_id": "AJLGU", + "question_id": "AJLGU", + "video_file": "Charades/videos/AJLGU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes a dish from the cabinet'." + }, + { + "from": "gpt", + "value": "From <14.7> to <19.3>." + } + ] + }, + { + "video_id": "GQZOR", + "question_id": "GQZOR", + "video_file": "Charades/videos/GQZOR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drink from a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.2> to <25.7>." + } + ] + }, + { + "video_id": "GQZOR", + "question_id": "GQZOR", + "video_file": "Charades/videos/GQZOR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eat it along with a drink' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.0> to <18.0>." + } + ] + }, + { + "video_id": "GQZOR", + "question_id": "GQZOR", + "video_file": "Charades/videos/GQZOR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they eat some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.0> to <18.0>." + } + ] + }, + { + "video_id": "5SGU6", + "question_id": "5SGU6", + "video_file": "Charades/videos/5SGU6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person next opens a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.9> to <16.3>." + } + ] + }, + { + "video_id": "5SGU6", + "question_id": "5SGU6", + "video_file": "Charades/videos/5SGU6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is in a pantry opening a grocery bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <9.4>." + } + ] + }, + { + "video_id": "5SGU6", + "question_id": "5SGU6", + "video_file": "Charades/videos/5SGU6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts dishes on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.0> to <11.0>." + } + ] + }, + { + "video_id": "5SGU6", + "question_id": "5SGU6", + "video_file": "Charades/videos/5SGU6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put the dishes away first' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.9> to <18.8>." + } + ] + }, + { + "video_id": "4KO74", + "question_id": "4KO74", + "video_file": "Charades/videos/4KO74.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is smiling' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.4> to <23.3>." + } + ] + }, + { + "video_id": "MILRI", + "question_id": "MILRI", + "video_file": "Charades/videos/MILRI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is opening a closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.1> to <9.6>." + } + ] + }, + { + "video_id": "B9XZX", + "question_id": "B9XZX", + "video_file": "Charades/videos/B9XZX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is cooking something on their stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "41O3J", + "question_id": "41O3J", + "video_file": "Charades/videos/41O3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits on the bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.1> to <26.8>." + } + ] + }, + { + "video_id": "41O3J", + "question_id": "41O3J", + "video_file": "Charades/videos/41O3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person undresses' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.4> to <12.9>." + } + ] + }, + { + "video_id": "41O3J", + "question_id": "41O3J", + "video_file": "Charades/videos/41O3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the blanket on the bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <19.0>." + } + ] + }, + { + "video_id": "Y9930", + "question_id": "Y9930", + "video_file": "Charades/videos/Y9930.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person opens a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.1> to <19.1>." + } + ] + }, + { + "video_id": "A81AV", + "question_id": "A81AV", + "video_file": "Charades/videos/A81AV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits on the bed'." + }, + { + "from": "gpt", + "value": "From <1.8> to <7.4>." + } + ] + }, + { + "video_id": "A81AV", + "question_id": "A81AV", + "video_file": "Charades/videos/A81AV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they begin dressing by putting on a jacket' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.2> to <21.4>." + } + ] + }, + { + "video_id": "A81AV", + "question_id": "A81AV", + "video_file": "Charades/videos/A81AV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they start drinking from a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.4> to <31.0>." + } + ] + }, + { + "video_id": "IJOAK", + "question_id": "IJOAK", + "video_file": "Charades/videos/IJOAK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person holding a dish'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.9>." + } + ] + }, + { + "video_id": "ZC8L5", + "question_id": "ZC8L5", + "video_file": "Charades/videos/ZC8L5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they throw a some clothes on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.7> to <23.0>." + } + ] + }, + { + "video_id": "ZC8L5", + "question_id": "ZC8L5", + "video_file": "Charades/videos/ZC8L5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sitting on chair turns to watch tv' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.0> to <32.0>." + } + ] + }, + { + "video_id": "ZC8L5", + "question_id": "ZC8L5", + "video_file": "Charades/videos/ZC8L5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they throw off their blanket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.2> to <24.2>." + } + ] + }, + { + "video_id": "ZC8L5", + "question_id": "ZC8L5", + "video_file": "Charades/videos/ZC8L5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting in a chair watching television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.0> to <32.0>." + } + ] + }, + { + "video_id": "7HZHM", + "question_id": "7HZHM", + "video_file": "Charades/videos/7HZHM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is undressing in front of the mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.7> to <21.7>." + } + ] + }, + { + "video_id": "IN2RL", + "question_id": "IN2RL", + "video_file": "Charades/videos/IN2RL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws boxes on floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <32.9> to <38.6>." + } + ] + }, + { + "video_id": "IN2RL", + "question_id": "IN2RL", + "video_file": "Charades/videos/IN2RL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they begin throwing a box around the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <32.9> to <38.6>." + } + ] + }, + { + "video_id": "UVPSZ", + "question_id": "UVPSZ", + "video_file": "Charades/videos/UVPSZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person after closing the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.7> to <12.3>." + } + ] + }, + { + "video_id": "UVPSZ", + "question_id": "UVPSZ", + "video_file": "Charades/videos/UVPSZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <3.7> to <12.3>." + } + ] + }, + { + "video_id": "UVPSZ", + "question_id": "UVPSZ", + "video_file": "Charades/videos/UVPSZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person walks through a doorway'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "JFXG8", + "question_id": "JFXG8", + "video_file": "Charades/videos/JFXG8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening refrigerator to take a pot out' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "JFXG8", + "question_id": "JFXG8", + "video_file": "Charades/videos/JFXG8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "JFXG8", + "question_id": "JFXG8", + "video_file": "Charades/videos/JFXG8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person close the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.9> to <11.4>." + } + ] + }, + { + "video_id": "2DPW0", + "question_id": "2DPW0", + "video_file": "Charades/videos/2DPW0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinks water out of a large cup or jug'." + }, + { + "from": "gpt", + "value": "From <6.2> to <17.2>." + } + ] + }, + { + "video_id": "1NYKM", + "question_id": "1NYKM", + "video_file": "Charades/videos/1NYKM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put away the book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.3> to <22.2>." + } + ] + }, + { + "video_id": "1NYKM", + "question_id": "1NYKM", + "video_file": "Charades/videos/1NYKM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person lies in the bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.0> to <36.0>." + } + ] + }, + { + "video_id": "1NYKM", + "question_id": "1NYKM", + "video_file": "Charades/videos/1NYKM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person watches television' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.9> to <36.0>." + } + ] + }, + { + "video_id": "1NYKM", + "question_id": "1NYKM", + "video_file": "Charades/videos/1NYKM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person lied down in bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.0> to <36.0>." + } + ] + }, + { + "video_id": "897OI", + "question_id": "897OI", + "video_file": "Charades/videos/897OI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they pour something from the bottle'." + }, + { + "from": "gpt", + "value": "From <12.9> to <23.1>." + } + ] + }, + { + "video_id": "897OI", + "question_id": "897OI", + "video_file": "Charades/videos/897OI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person holding a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.1> to <16.7>." + } + ] + }, + { + "video_id": "97Y6T", + "question_id": "97Y6T", + "video_file": "Charades/videos/97Y6T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is talking on a phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "97Y6T", + "question_id": "97Y6T", + "video_file": "Charades/videos/97Y6T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sits talking on their phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "VPJLN", + "question_id": "VPJLN", + "video_file": "Charades/videos/VPJLN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person runs in through the entryway holding some clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.4> to <23.4>." + } + ] + }, + { + "video_id": "EG1MY", + "question_id": "EG1MY", + "video_file": "Charades/videos/EG1MY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they are holding a towel' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.7> to <14.6>." + } + ] + }, + { + "video_id": "EG1MY", + "question_id": "EG1MY", + "video_file": "Charades/videos/EG1MY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person began putting the towel on a towel rack' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.6> to <28.6>." + } + ] + }, + { + "video_id": "QX89T", + "question_id": "QX89T", + "video_file": "Charades/videos/QX89T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person fixes his hair'." + }, + { + "from": "gpt", + "value": "From <5.8> to <18.8>." + } + ] + }, + { + "video_id": "QX89T", + "question_id": "QX89T", + "video_file": "Charades/videos/QX89T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they fix their hair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.8> to <18.8>." + } + ] + }, + { + "video_id": "QX89T", + "question_id": "QX89T", + "video_file": "Charades/videos/QX89T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person fixes their hair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.8> to <18.8>." + } + ] + }, + { + "video_id": "QX89T", + "question_id": "QX89T", + "video_file": "Charades/videos/QX89T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes pictures of himself using his phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <33.0>." + } + ] + }, + { + "video_id": "QX89T", + "question_id": "QX89T", + "video_file": "Charades/videos/QX89T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a picture in the mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.1> to <33.0>." + } + ] + }, + { + "video_id": "CMC93", + "question_id": "CMC93", + "video_file": "Charades/videos/CMC93.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closed the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <13.4>." + } + ] + }, + { + "video_id": "CMC93", + "question_id": "CMC93", + "video_file": "Charades/videos/CMC93.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put on shoes,stood up from the chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.0> to <27.5>." + } + ] + }, + { + "video_id": "CMC93", + "question_id": "CMC93", + "video_file": "Charades/videos/CMC93.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is dressing in their bedroom'." + }, + { + "from": "gpt", + "value": "From <12.8> to <31.7>." + } + ] + }, + { + "video_id": "CMC93", + "question_id": "CMC93", + "video_file": "Charades/videos/CMC93.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sitting on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.5> to <20.2>." + } + ] + }, + { + "video_id": "CMC93", + "question_id": "CMC93", + "video_file": "Charades/videos/CMC93.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <46.9> to <53.0>." + } + ] + }, + { + "video_id": "CMC93", + "question_id": "CMC93", + "video_file": "Charades/videos/CMC93.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opened the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <46.9> to <53.0>." + } + ] + }, + { + "video_id": "CMC93", + "question_id": "CMC93", + "video_file": "Charades/videos/CMC93.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.6> to <13.4>." + } + ] + }, + { + "video_id": "CMC93", + "question_id": "CMC93", + "video_file": "Charades/videos/CMC93.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person stand up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <37.5> to <49.1>." + } + ] + }, + { + "video_id": "S0B9O", + "question_id": "S0B9O", + "video_file": "Charades/videos/S0B9O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person is smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <15.1>." + } + ] + }, + { + "video_id": "OQM2I", + "question_id": "OQM2I", + "video_file": "Charades/videos/OQM2I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closing the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <11.9>." + } + ] + }, + { + "video_id": "OQM2I", + "question_id": "OQM2I", + "video_file": "Charades/videos/OQM2I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person fixes their hair in a mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.2> to <6.3>." + } + ] + }, + { + "video_id": "PQYWB", + "question_id": "PQYWB", + "video_file": "Charades/videos/PQYWB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person he closes the door behind him' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.8> to <40.0>." + } + ] + }, + { + "video_id": "KVYOZ", + "question_id": "KVYOZ", + "video_file": "Charades/videos/KVYOZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.6>." + } + ] + }, + { + "video_id": "KVYOZ", + "question_id": "KVYOZ", + "video_file": "Charades/videos/KVYOZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sits on the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.5>." + } + ] + }, + { + "video_id": "EN9TD", + "question_id": "EN9TD", + "video_file": "Charades/videos/EN9TD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking a cup of coffee'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.3> to <10.6>." + } + ] + }, + { + "video_id": "V6M35", + "question_id": "V6M35", + "video_file": "Charades/videos/V6M35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinks from a cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.6> to <27.5>." + } + ] + }, + { + "video_id": "AQ3CP", + "question_id": "AQ3CP", + "video_file": "Charades/videos/AQ3CP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.3>." + } + ] + }, + { + "video_id": "AQ3CP", + "question_id": "AQ3CP", + "video_file": "Charades/videos/AQ3CP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sitting in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.3>." + } + ] + }, + { + "video_id": "EJL2F", + "question_id": "EJL2F", + "video_file": "Charades/videos/EJL2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "EJL2F", + "question_id": "EJL2F", + "video_file": "Charades/videos/EJL2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns the light on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <8.4>." + } + ] + }, + { + "video_id": "EJL2F", + "question_id": "EJL2F", + "video_file": "Charades/videos/EJL2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "EJL2F", + "question_id": "EJL2F", + "video_file": "Charades/videos/EJL2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns on the light in the recreation room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.2> to <8.4>." + } + ] + }, + { + "video_id": "EJL2F", + "question_id": "EJL2F", + "video_file": "Charades/videos/EJL2F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <8.4>." + } + ] + }, + { + "video_id": "RZ9SZ", + "question_id": "RZ9SZ", + "video_file": "Charades/videos/RZ9SZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is undressing in front of the mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.3> to <14.1>." + } + ] + }, + { + "video_id": "N4Q0O", + "question_id": "N4Q0O", + "video_file": "Charades/videos/N4Q0O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is awakening on a bed'." + }, + { + "from": "gpt", + "value": "From <1.0> to <11.9>." + } + ] + }, + { + "video_id": "N4Q0O", + "question_id": "N4Q0O", + "video_file": "Charades/videos/N4Q0O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person awakens from under a blanket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <11.9>." + } + ] + }, + { + "video_id": "XJS1X", + "question_id": "XJS1X", + "video_file": "Charades/videos/XJS1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person turns on a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <12.3>." + } + ] + }, + { + "video_id": "3Z08H", + "question_id": "3Z08H", + "video_file": "Charades/videos/3Z08H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <10.5>." + } + ] + }, + { + "video_id": "15TWQ", + "question_id": "15TWQ", + "video_file": "Charades/videos/15TWQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person turns on a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.9> to <5.7>." + } + ] + }, + { + "video_id": "15TWQ", + "question_id": "15TWQ", + "video_file": "Charades/videos/15TWQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person turns on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <5.7>." + } + ] + }, + { + "video_id": "J3UKN", + "question_id": "J3UKN", + "video_file": "Charades/videos/J3UKN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person walks through a doorway toward another room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.0> to <34.0>." + } + ] + }, + { + "video_id": "DZDD2", + "question_id": "DZDD2", + "video_file": "Charades/videos/DZDD2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.7>." + } + ] + }, + { + "video_id": "5GK8W", + "question_id": "5GK8W", + "video_file": "Charades/videos/5GK8W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens on the living room sofa'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "5GK8W", + "question_id": "5GK8W", + "video_file": "Charades/videos/5GK8W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person hurriedly putting on their shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.2> to <27.2>." + } + ] + }, + { + "video_id": "5GK8W", + "question_id": "5GK8W", + "video_file": "Charades/videos/5GK8W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts on their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <27.2>." + } + ] + }, + { + "video_id": "3STO2", + "question_id": "3STO2", + "video_file": "Charades/videos/3STO2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.4> to <20.4>." + } + ] + }, + { + "video_id": "0BH84", + "question_id": "0BH84", + "video_file": "Charades/videos/0BH84.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person talking on the phone'." + }, + { + "from": "gpt", + "value": "From <4.0> to <16.8>." + } + ] + }, + { + "video_id": "0BH84", + "question_id": "0BH84", + "video_file": "Charades/videos/0BH84.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they take off their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <26.0>." + } + ] + }, + { + "video_id": "0BH84", + "question_id": "0BH84", + "video_file": "Charades/videos/0BH84.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes their shoes off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <26.0>." + } + ] + }, + { + "video_id": "0PKMF", + "question_id": "0PKMF", + "video_file": "Charades/videos/0PKMF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person play on a phone'." + }, + { + "from": "gpt", + "value": "From <20.6> to <32.0>." + } + ] + }, + { + "video_id": "0PKMF", + "question_id": "0PKMF", + "video_file": "Charades/videos/0PKMF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person laughing at a television show' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <6.8>." + } + ] + }, + { + "video_id": "LY2GQ", + "question_id": "LY2GQ", + "video_file": "Charades/videos/LY2GQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the broom down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.8> to <26.3>." + } + ] + }, + { + "video_id": "W28C2", + "question_id": "W28C2", + "video_file": "Charades/videos/W28C2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person laughs as they cook on the stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.1> to <11.3>." + } + ] + }, + { + "video_id": "W28C2", + "question_id": "W28C2", + "video_file": "Charades/videos/W28C2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is cooking' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.1>." + } + ] + }, + { + "video_id": "W28C2", + "question_id": "W28C2", + "video_file": "Charades/videos/W28C2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes out a box of food'." + }, + { + "from": "gpt", + "value": "From <9.8> to <15.4>." + } + ] + }, + { + "video_id": "1ZBUS", + "question_id": "1ZBUS", + "video_file": "Charades/videos/1ZBUS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts a bag of groceries on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <13.6>." + } + ] + }, + { + "video_id": "1ZBUS", + "question_id": "1ZBUS", + "video_file": "Charades/videos/1ZBUS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting at a table'." + }, + { + "from": "gpt", + "value": "From <5.2> to <10.8>." + } + ] + }, + { + "video_id": "FS94F", + "question_id": "FS94F", + "video_file": "Charades/videos/FS94F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pours the contents from the bottle into the glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <12.4>." + } + ] + }, + { + "video_id": "FS94F", + "question_id": "FS94F", + "video_file": "Charades/videos/FS94F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pours it in a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.4> to <12.4>." + } + ] + }, + { + "video_id": "FS94F", + "question_id": "FS94F", + "video_file": "Charades/videos/FS94F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is pouring a drink into a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <12.4>." + } + ] + }, + { + "video_id": "QRZK5", + "question_id": "QRZK5", + "video_file": "Charades/videos/QRZK5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is lying on floor eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <18.7>." + } + ] + }, + { + "video_id": "QRZK5", + "question_id": "QRZK5", + "video_file": "Charades/videos/QRZK5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.1> to <18.7>." + } + ] + }, + { + "video_id": "QRZK5", + "question_id": "QRZK5", + "video_file": "Charades/videos/QRZK5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person watching television' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <13.3>." + } + ] + }, + { + "video_id": "QRZK5", + "question_id": "QRZK5", + "video_file": "Charades/videos/QRZK5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person watching television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <13.3>." + } + ] + }, + { + "video_id": "GQT29", + "question_id": "GQT29", + "video_file": "Charades/videos/GQT29.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take a book off the shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.2> to <23.8>." + } + ] + }, + { + "video_id": "GQT29", + "question_id": "GQT29", + "video_file": "Charades/videos/GQT29.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person read the book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.1> to <25.9>." + } + ] + }, + { + "video_id": "GQT29", + "question_id": "GQT29", + "video_file": "Charades/videos/GQT29.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is fixing a vacuum' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "APZUZ", + "question_id": "APZUZ", + "video_file": "Charades/videos/APZUZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person puts a sandwich on the shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.2>." + } + ] + }, + { + "video_id": "APZUZ", + "question_id": "APZUZ", + "video_file": "Charades/videos/APZUZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinks from a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <14.5>." + } + ] + }, + { + "video_id": "C8BKE", + "question_id": "C8BKE", + "video_file": "Charades/videos/C8BKE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put their groceries on the bed'." + }, + { + "from": "gpt", + "value": "From <11.0> to <16.5>." + } + ] + }, + { + "video_id": "C8BKE", + "question_id": "C8BKE", + "video_file": "Charades/videos/C8BKE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walks through doorway with groceries' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.1> to <9.9>." + } + ] + }, + { + "video_id": "C8BKE", + "question_id": "C8BKE", + "video_file": "Charades/videos/C8BKE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes off shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.6> to <12.3>." + } + ] + }, + { + "video_id": "C8BKE", + "question_id": "C8BKE", + "video_file": "Charades/videos/C8BKE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits down on bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <17.7>." + } + ] + }, + { + "video_id": "C8BKE", + "question_id": "C8BKE", + "video_file": "Charades/videos/C8BKE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes their shoes off' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.6> to <12.3>." + } + ] + }, + { + "video_id": "RXF2Q", + "question_id": "RXF2Q", + "video_file": "Charades/videos/RXF2Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens up their laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.4> to <33.5>." + } + ] + }, + { + "video_id": "RXF2Q", + "question_id": "RXF2Q", + "video_file": "Charades/videos/RXF2Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open the laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.4> to <33.5>." + } + ] + }, + { + "video_id": "EASCD", + "question_id": "EASCD", + "video_file": "Charades/videos/EASCD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person he pours a glass of soda' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.5> to <25.0>." + } + ] + }, + { + "video_id": "GWUM6", + "question_id": "GWUM6", + "video_file": "Charades/videos/GWUM6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person put away the dishes in the cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.0> to <10.6>." + } + ] + }, + { + "video_id": "GWUM6", + "question_id": "GWUM6", + "video_file": "Charades/videos/GWUM6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is seen putting dishes away' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.0> to <10.6>." + } + ] + }, + { + "video_id": "7C13O", + "question_id": "7C13O", + "video_file": "Charades/videos/7C13O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sit in a chair when they are done'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.4> to <50.0>." + } + ] + }, + { + "video_id": "7C13O", + "question_id": "7C13O", + "video_file": "Charades/videos/7C13O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits on chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.4> to <50.0>." + } + ] + }, + { + "video_id": "3SKPS", + "question_id": "3SKPS", + "video_file": "Charades/videos/3SKPS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person another works on a laptop'." + }, + { + "from": "gpt", + "value": "From <17.6> to <33.0>." + } + ] + }, + { + "video_id": "3SKPS", + "question_id": "3SKPS", + "video_file": "Charades/videos/3SKPS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a second person is working on their laptop'." + }, + { + "from": "gpt", + "value": "From <17.6> to <33.0>." + } + ] + }, + { + "video_id": "83V0A", + "question_id": "83V0A", + "video_file": "Charades/videos/83V0A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is cooking food on the stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "83V0A", + "question_id": "83V0A", + "video_file": "Charades/videos/83V0A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is at the kitchen stove cooking a meal' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "83V0A", + "question_id": "83V0A", + "video_file": "Charades/videos/83V0A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person look out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.4> to <11.7>." + } + ] + }, + { + "video_id": "83V0A", + "question_id": "83V0A", + "video_file": "Charades/videos/83V0A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person a kid cooking' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "83V0A", + "question_id": "83V0A", + "video_file": "Charades/videos/83V0A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts reading a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <21.0>." + } + ] + }, + { + "video_id": "83V0A", + "question_id": "83V0A", + "video_file": "Charades/videos/83V0A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looking out a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.4> to <11.7>." + } + ] + }, + { + "video_id": "83V0A", + "question_id": "83V0A", + "video_file": "Charades/videos/83V0A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person setting down to look at a cook book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "83V0A", + "question_id": "83V0A", + "video_file": "Charades/videos/83V0A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person turns around to look out of a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.4> to <11.7>." + } + ] + }, + { + "video_id": "83V0A", + "question_id": "83V0A", + "video_file": "Charades/videos/83V0A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person sits in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.4> to <16.6>." + } + ] + }, + { + "video_id": "1TAMK", + "question_id": "1TAMK", + "video_file": "Charades/videos/1TAMK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person runs into the entryway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <8.0>." + } + ] + }, + { + "video_id": "1TAMK", + "question_id": "1TAMK", + "video_file": "Charades/videos/1TAMK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is opening a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <9.1>." + } + ] + }, + { + "video_id": "T87QE", + "question_id": "T87QE", + "video_file": "Charades/videos/T87QE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins to undress their clothes with a smile' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.8> to <34.9>." + } + ] + }, + { + "video_id": "JTZZW", + "question_id": "JTZZW", + "video_file": "Charades/videos/JTZZW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person takes a laptop off a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "JTZZW", + "question_id": "JTZZW", + "video_file": "Charades/videos/JTZZW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begin watching a video off of the laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.8> to <32.0>." + } + ] + }, + { + "video_id": "2MAZY", + "question_id": "2MAZY", + "video_file": "Charades/videos/2MAZY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person begins to take groceries out of the box'." + }, + { + "from": "gpt", + "value": "From <1.0> to <6.0>." + } + ] + }, + { + "video_id": "2MAZY", + "question_id": "2MAZY", + "video_file": "Charades/videos/2MAZY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put them on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.5> to <8.1>." + } + ] + }, + { + "video_id": "2MAZY", + "question_id": "2MAZY", + "video_file": "Charades/videos/2MAZY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opening a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "2MAZY", + "question_id": "2MAZY", + "video_file": "Charades/videos/2MAZY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'sneezing as the person does so' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.2> to <7.1>." + } + ] + }, + { + "video_id": "2MAZY", + "question_id": "2MAZY", + "video_file": "Charades/videos/2MAZY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opens a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "Q07CU", + "question_id": "Q07CU", + "video_file": "Charades/videos/Q07CU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turn off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.3> to <35.0>." + } + ] + }, + { + "video_id": "Q07CU", + "question_id": "Q07CU", + "video_file": "Charades/videos/Q07CU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns the light off' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.3> to <35.0>." + } + ] + }, + { + "video_id": "DF6FY", + "question_id": "DF6FY", + "video_file": "Charades/videos/DF6FY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person snuggling a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.2> to <30.1>." + } + ] + }, + { + "video_id": "DF6FY", + "question_id": "DF6FY", + "video_file": "Charades/videos/DF6FY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking from a glass'." + }, + { + "from": "gpt", + "value": "From <10.9> to <18.9>." + } + ] + }, + { + "video_id": "DF6FY", + "question_id": "DF6FY", + "video_file": "Charades/videos/DF6FY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they begin to drink from a glass of water'." + }, + { + "from": "gpt", + "value": "From <10.9> to <18.9>." + } + ] + }, + { + "video_id": "HU409", + "question_id": "HU409", + "video_file": "Charades/videos/HU409.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the pillow away' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.1> to <31.0>." + } + ] + }, + { + "video_id": "HU409", + "question_id": "HU409", + "video_file": "Charades/videos/HU409.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they take the cover off the pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.6> to <22.5>." + } + ] + }, + { + "video_id": "I005F", + "question_id": "I005F", + "video_file": "Charades/videos/I005F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person the open the refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <13.1>." + } + ] + }, + { + "video_id": "I005F", + "question_id": "I005F", + "video_file": "Charades/videos/I005F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns down the heat' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <18.2>." + } + ] + }, + { + "video_id": "TZL6H", + "question_id": "TZL6H", + "video_file": "Charades/videos/TZL6H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person washing a mirror in the hallway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.0> to <32.0>." + } + ] + }, + { + "video_id": "D548M", + "question_id": "D548M", + "video_file": "Charades/videos/D548M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks through the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <23.7>." + } + ] + }, + { + "video_id": "D548M", + "question_id": "D548M", + "video_file": "Charades/videos/D548M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts the laptop down onto the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.1> to <31.5>." + } + ] + }, + { + "video_id": "QPX3S", + "question_id": "QPX3S", + "video_file": "Charades/videos/QPX3S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is looking out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "CC610", + "question_id": "CC610", + "video_file": "Charades/videos/CC610.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person pours something into a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <15.1>." + } + ] + }, + { + "video_id": "CC610", + "question_id": "CC610", + "video_file": "Charades/videos/CC610.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person pours soda in a glass at the sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <15.1>." + } + ] + }, + { + "video_id": "CC610", + "question_id": "CC610", + "video_file": "Charades/videos/CC610.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person sits down on the floor with a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.0> to <34.0>." + } + ] + }, + { + "video_id": "9SIZS", + "question_id": "9SIZS", + "video_file": "Charades/videos/9SIZS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws them on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <17.0>." + } + ] + }, + { + "video_id": "9SIZS", + "question_id": "9SIZS", + "video_file": "Charades/videos/9SIZS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is tidying up some clothes on their bed'." + }, + { + "from": "gpt", + "value": "From <10.3> to <17.0>." + } + ] + }, + { + "video_id": "MDXF4", + "question_id": "MDXF4", + "video_file": "Charades/videos/MDXF4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.1> to <12.2>." + } + ] + }, + { + "video_id": "MDXF4", + "question_id": "MDXF4", + "video_file": "Charades/videos/MDXF4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the closet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.1> to <12.2>." + } + ] + }, + { + "video_id": "MDXF4", + "question_id": "MDXF4", + "video_file": "Charades/videos/MDXF4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person was drinking a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.1> to <23.5>." + } + ] + }, + { + "video_id": "MDXF4", + "question_id": "MDXF4", + "video_file": "Charades/videos/MDXF4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person drinking a glass of water walks in' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.1> to <23.5>." + } + ] + }, + { + "video_id": "8MDSA", + "question_id": "8MDSA", + "video_file": "Charades/videos/8MDSA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is undressing in their closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.6> to <14.2>." + } + ] + }, + { + "video_id": "8MDSA", + "question_id": "8MDSA", + "video_file": "Charades/videos/8MDSA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.7> to <16.2>." + } + ] + }, + { + "video_id": "8MDSA", + "question_id": "8MDSA", + "video_file": "Charades/videos/8MDSA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turn the light off' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.4> to <33.0>." + } + ] + }, + { + "video_id": "8MDSA", + "question_id": "8MDSA", + "video_file": "Charades/videos/8MDSA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person turns off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.4> to <33.0>." + } + ] + }, + { + "video_id": "8MDSA", + "question_id": "8MDSA", + "video_file": "Charades/videos/8MDSA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.4> to <33.0>." + } + ] + }, + { + "video_id": "9MXDX", + "question_id": "9MXDX", + "video_file": "Charades/videos/9MXDX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.4> to <10.9>." + } + ] + }, + { + "video_id": "9MXDX", + "question_id": "9MXDX", + "video_file": "Charades/videos/9MXDX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sitting on couch eating'." + }, + { + "from": "gpt", + "value": "From <8.5> to <19.0>." + } + ] + }, + { + "video_id": "9MXDX", + "question_id": "9MXDX", + "video_file": "Charades/videos/9MXDX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating food off of two plates' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.5> to <19.0>." + } + ] + }, + { + "video_id": "9MXDX", + "question_id": "9MXDX", + "video_file": "Charades/videos/9MXDX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person is laughing at what he is watching' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.1> to <27.7>." + } + ] + }, + { + "video_id": "9MXDX", + "question_id": "9MXDX", + "video_file": "Charades/videos/9MXDX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking ans smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.4> to <27.8>." + } + ] + }, + { + "video_id": "9MXDX", + "question_id": "9MXDX", + "video_file": "Charades/videos/9MXDX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person laughs' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.1> to <27.7>." + } + ] + }, + { + "video_id": "9MXDX", + "question_id": "9MXDX", + "video_file": "Charades/videos/9MXDX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating food from a few dishes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.5> to <19.0>." + } + ] + }, + { + "video_id": "QNQOG", + "question_id": "QNQOG", + "video_file": "Charades/videos/QNQOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is undressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "QNQOG", + "question_id": "QNQOG", + "video_file": "Charades/videos/QNQOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they look out a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.0> to <19.0>." + } + ] + }, + { + "video_id": "QNQOG", + "question_id": "QNQOG", + "video_file": "Charades/videos/QNQOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throwing the clothes on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <10.8>." + } + ] + }, + { + "video_id": "QNQOG", + "question_id": "QNQOG", + "video_file": "Charades/videos/QNQOG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.0> to <27.0>." + } + ] + }, + { + "video_id": "45Z2A", + "question_id": "45Z2A", + "video_file": "Charades/videos/45Z2A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person awakens on the couch' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "DSG0F", + "question_id": "DSG0F", + "video_file": "Charades/videos/DSG0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks through the doorway of their home office' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "DSG0F", + "question_id": "DSG0F", + "video_file": "Charades/videos/DSG0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person drinks from a cup of coffee' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <13.1>." + } + ] + }, + { + "video_id": "E5O0F", + "question_id": "E5O0F", + "video_file": "Charades/videos/E5O0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person washes their hands'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <33.9>." + } + ] + }, + { + "video_id": "E5O0F", + "question_id": "E5O0F", + "video_file": "Charades/videos/E5O0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes their laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.7> to <16.7>." + } + ] + }, + { + "video_id": "E5O0F", + "question_id": "E5O0F", + "video_file": "Charades/videos/E5O0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person closes their laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.7> to <16.7>." + } + ] + }, + { + "video_id": "E5O0F", + "question_id": "E5O0F", + "video_file": "Charades/videos/E5O0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'washing the person's hands in a sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <33.9>." + } + ] + }, + { + "video_id": "E5O0F", + "question_id": "E5O0F", + "video_file": "Charades/videos/E5O0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person works on a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <30.9>." + } + ] + }, + { + "video_id": "E5O0F", + "question_id": "E5O0F", + "video_file": "Charades/videos/E5O0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person washes their hands' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.5> to <33.9>." + } + ] + }, + { + "video_id": "E5O0F", + "question_id": "E5O0F", + "video_file": "Charades/videos/E5O0F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.7> to <16.7>." + } + ] + }, + { + "video_id": "HOI6L", + "question_id": "HOI6L", + "video_file": "Charades/videos/HOI6L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opened the door to walk out of the room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.9> to <31.2>." + } + ] + }, + { + "video_id": "GWDF9", + "question_id": "GWDF9", + "video_file": "Charades/videos/GWDF9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person runs in room'." + }, + { + "from": "gpt", + "value": "From <2.7> to <13.0>." + } + ] + }, + { + "video_id": "GWDF9", + "question_id": "GWDF9", + "video_file": "Charades/videos/GWDF9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs into the living room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <13.0>." + } + ] + }, + { + "video_id": "GWDF9", + "question_id": "GWDF9", + "video_file": "Charades/videos/GWDF9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person comes running through the front door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <13.0>." + } + ] + }, + { + "video_id": "3VRTT", + "question_id": "3VRTT", + "video_file": "Charades/videos/3VRTT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person running away' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.9> to <12.0>." + } + ] + }, + { + "video_id": "WHVP2", + "question_id": "WHVP2", + "video_file": "Charades/videos/WHVP2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put it on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "VSLQT", + "question_id": "VSLQT", + "video_file": "Charades/videos/VSLQT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take a sandwich from a plate' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <11.5>." + } + ] + }, + { + "video_id": "VSLQT", + "question_id": "VSLQT", + "video_file": "Charades/videos/VSLQT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakens in bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "5P0ON", + "question_id": "5P0ON", + "video_file": "Charades/videos/5P0ON.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is standing in the laundry room eating' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "CN5G4", + "question_id": "CN5G4", + "video_file": "Charades/videos/CN5G4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sitting at the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <17.4>." + } + ] + }, + { + "video_id": "CYT5I", + "question_id": "CYT5I", + "video_file": "Charades/videos/CYT5I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they walk to the doorway of their bedroom'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.7> to <27.2>." + } + ] + }, + { + "video_id": "CYT5I", + "question_id": "CYT5I", + "video_file": "Charades/videos/CYT5I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put their pillow on top of it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <18.3>." + } + ] + }, + { + "video_id": "CYT5I", + "question_id": "CYT5I", + "video_file": "Charades/videos/CYT5I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closing the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.9> to <27.5>." + } + ] + }, + { + "video_id": "CYT5I", + "question_id": "CYT5I", + "video_file": "Charades/videos/CYT5I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts a pillow over it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <18.3>." + } + ] + }, + { + "video_id": "FI2DL", + "question_id": "FI2DL", + "video_file": "Charades/videos/FI2DL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person runs through playing a game'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <15.6>." + } + ] + }, + { + "video_id": "GIZ95", + "question_id": "GIZ95", + "video_file": "Charades/videos/GIZ95.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person look out window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.4> to <32.0>." + } + ] + }, + { + "video_id": "GIZ95", + "question_id": "GIZ95", + "video_file": "Charades/videos/GIZ95.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they look out the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.4> to <32.0>." + } + ] + }, + { + "video_id": "GIZ95", + "question_id": "GIZ95", + "video_file": "Charades/videos/GIZ95.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person looks out window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.4> to <32.0>." + } + ] + }, + { + "video_id": "GIZ95", + "question_id": "GIZ95", + "video_file": "Charades/videos/GIZ95.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.8> to <32.0>." + } + ] + }, + { + "video_id": "WYCRK", + "question_id": "WYCRK", + "video_file": "Charades/videos/WYCRK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person looks out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.9>." + } + ] + }, + { + "video_id": "WYCRK", + "question_id": "WYCRK", + "video_file": "Charades/videos/WYCRK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.7> to <17.8>." + } + ] + }, + { + "video_id": "WYCRK", + "question_id": "WYCRK", + "video_file": "Charades/videos/WYCRK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they drink out of the glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.7> to <17.8>." + } + ] + }, + { + "video_id": "WYCRK", + "question_id": "WYCRK", + "video_file": "Charades/videos/WYCRK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.6> to <12.6>." + } + ] + }, + { + "video_id": "56PK0", + "question_id": "56PK0", + "video_file": "Charades/videos/56PK0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.6> to <34.8>." + } + ] + }, + { + "video_id": "56PK0", + "question_id": "56PK0", + "video_file": "Charades/videos/56PK0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts the blanket into the washing machine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <35.2> to <42.7>." + } + ] + }, + { + "video_id": "56PK0", + "question_id": "56PK0", + "video_file": "Charades/videos/56PK0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the blanket into a dishwasher' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <35.2> to <42.7>." + } + ] + }, + { + "video_id": "56PK0", + "question_id": "56PK0", + "video_file": "Charades/videos/56PK0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they sit down in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.6> to <34.8>." + } + ] + }, + { + "video_id": "56PK0", + "question_id": "56PK0", + "video_file": "Charades/videos/56PK0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes a blanket out form the box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <14.7>." + } + ] + }, + { + "video_id": "56PK0", + "question_id": "56PK0", + "video_file": "Charades/videos/56PK0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person takes a blanket out of a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.1> to <14.7>." + } + ] + }, + { + "video_id": "56PK0", + "question_id": "56PK0", + "video_file": "Charades/videos/56PK0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a blanket out of it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.1> to <14.7>." + } + ] + }, + { + "video_id": "JKR9A", + "question_id": "JKR9A", + "video_file": "Charades/videos/JKR9A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns on the light'." + }, + { + "from": "gpt", + "value": "From <10.0> to <18.6>." + } + ] + }, + { + "video_id": "JKR9A", + "question_id": "JKR9A", + "video_file": "Charades/videos/JKR9A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turning on a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <18.6>." + } + ] + }, + { + "video_id": "JKR9A", + "question_id": "JKR9A", + "video_file": "Charades/videos/JKR9A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they turn on a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.0> to <18.6>." + } + ] + }, + { + "video_id": "KG2LF", + "question_id": "KG2LF", + "video_file": "Charades/videos/KG2LF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns off the lights'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.6> to <28.9>." + } + ] + }, + { + "video_id": "KG2LF", + "question_id": "KG2LF", + "video_file": "Charades/videos/KG2LF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.6> to <28.9>." + } + ] + }, + { + "video_id": "KG2LF", + "question_id": "KG2LF", + "video_file": "Charades/videos/KG2LF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they throw it on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.5> to <19.2>." + } + ] + }, + { + "video_id": "KG2LF", + "question_id": "KG2LF", + "video_file": "Charades/videos/KG2LF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person throws their clothes on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.1> to <18.5>." + } + ] + }, + { + "video_id": "CPZZ1", + "question_id": "CPZZ1", + "video_file": "Charades/videos/CPZZ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person running their fingers through their hair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.0> to <7.2>." + } + ] + }, + { + "video_id": "CPZZ1", + "question_id": "CPZZ1", + "video_file": "Charades/videos/CPZZ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs into their bathroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <7.2>." + } + ] + }, + { + "video_id": "CPZZ1", + "question_id": "CPZZ1", + "video_file": "Charades/videos/CPZZ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs into a bathroom'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <7.2>." + } + ] + }, + { + "video_id": "IXJKT", + "question_id": "IXJKT", + "video_file": "Charades/videos/IXJKT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs out of the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.7> to <26.8>." + } + ] + }, + { + "video_id": "IXJKT", + "question_id": "IXJKT", + "video_file": "Charades/videos/IXJKT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person runs out of the hallway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.7> to <26.8>." + } + ] + }, + { + "video_id": "UGJSB", + "question_id": "UGJSB", + "video_file": "Charades/videos/UGJSB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person switches off the light when exiting the room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.3> to <27.4>." + } + ] + }, + { + "video_id": "8XXZZ", + "question_id": "8XXZZ", + "video_file": "Charades/videos/8XXZZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts shoes on' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <17.1>." + } + ] + }, + { + "video_id": "8XXZZ", + "question_id": "8XXZZ", + "video_file": "Charades/videos/8XXZZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they begin putting on their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <17.1>." + } + ] + }, + { + "video_id": "8XXZZ", + "question_id": "8XXZZ", + "video_file": "Charades/videos/8XXZZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs around a room'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "8XXZZ", + "question_id": "8XXZZ", + "video_file": "Charades/videos/8XXZZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is running around in the garage' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "8XXZZ", + "question_id": "8XXZZ", + "video_file": "Charades/videos/8XXZZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person take out their phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.7> to <20.9>." + } + ] + }, + { + "video_id": "F8M2Y", + "question_id": "F8M2Y", + "video_file": "Charades/videos/F8M2Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is undressing in their basement'." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.5>." + } + ] + }, + { + "video_id": "ANA5N", + "question_id": "ANA5N", + "video_file": "Charades/videos/ANA5N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens door smiling folds towel puts ina closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.2> to <21.6>." + } + ] + }, + { + "video_id": "O95PG", + "question_id": "O95PG", + "video_file": "Charades/videos/O95PG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person partially undressing'." + }, + { + "from": "gpt", + "value": "From <17.3> to <32.0>." + } + ] + }, + { + "video_id": "4LDRK", + "question_id": "4LDRK", + "video_file": "Charades/videos/4LDRK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person is opening door with groceries in hand' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <16.9>." + } + ] + }, + { + "video_id": "4LDRK", + "question_id": "4LDRK", + "video_file": "Charades/videos/4LDRK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens the entry way door'." + }, + { + "from": "gpt", + "value": "From <2.4> to <16.9>." + } + ] + }, + { + "video_id": "4LDRK", + "question_id": "4LDRK", + "video_file": "Charades/videos/4LDRK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person is pouring coffee in cup'." + }, + { + "from": "gpt", + "value": "From <22.4> to <34.0>." + } + ] + }, + { + "video_id": "6UDGS", + "question_id": "6UDGS", + "video_file": "Charades/videos/6UDGS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.9> to <24.0>." + } + ] + }, + { + "video_id": "6UDGS", + "question_id": "6UDGS", + "video_file": "Charades/videos/6UDGS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts on a pair of shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <19.0>." + } + ] + }, + { + "video_id": "6UDGS", + "question_id": "6UDGS", + "video_file": "Charades/videos/6UDGS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts on their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <19.0>." + } + ] + }, + { + "video_id": "564DX", + "question_id": "564DX", + "video_file": "Charades/videos/564DX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes a bottle of water from the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.4> to <8.4>." + } + ] + }, + { + "video_id": "4KOBZ", + "question_id": "4KOBZ", + "video_file": "Charades/videos/4KOBZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting down eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "4KOBZ", + "question_id": "4KOBZ", + "video_file": "Charades/videos/4KOBZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "4KOBZ", + "question_id": "4KOBZ", + "video_file": "Charades/videos/4KOBZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.7> to <24.0>." + } + ] + }, + { + "video_id": "2M9YL", + "question_id": "2M9YL", + "video_file": "Charades/videos/2M9YL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opened the door'." + }, + { + "from": "gpt", + "value": "From <22.9> to <31.0>." + } + ] + }, + { + "video_id": "2M9YL", + "question_id": "2M9YL", + "video_file": "Charades/videos/2M9YL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person was smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.0> to <26.9>." + } + ] + }, + { + "video_id": "2M9YL", + "question_id": "2M9YL", + "video_file": "Charades/videos/2M9YL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opening the bedroom door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.9> to <31.0>." + } + ] + }, + { + "video_id": "OM2NF", + "question_id": "OM2NF", + "video_file": "Charades/videos/OM2NF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts his shoes on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.7> to <34.8>." + } + ] + }, + { + "video_id": "OM2NF", + "question_id": "OM2NF", + "video_file": "Charades/videos/OM2NF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking coffee from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.2> to <8.8>." + } + ] + }, + { + "video_id": "OM2NF", + "question_id": "OM2NF", + "video_file": "Charades/videos/OM2NF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person he puts on his shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.7> to <34.8>." + } + ] + }, + { + "video_id": "OM2NF", + "question_id": "OM2NF", + "video_file": "Charades/videos/OM2NF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walks to the stove to cook breakfast' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <34.1> to <45.0>." + } + ] + }, + { + "video_id": "OM2NF", + "question_id": "OM2NF", + "video_file": "Charades/videos/OM2NF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks to cook at the oven' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <34.1> to <45.0>." + } + ] + }, + { + "video_id": "OM2NF", + "question_id": "OM2NF", + "video_file": "Charades/videos/OM2NF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts on shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.7> to <34.8>." + } + ] + }, + { + "video_id": "AH4UW", + "question_id": "AH4UW", + "video_file": "Charades/videos/AH4UW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.4> to <8.5>." + } + ] + }, + { + "video_id": "AH4UW", + "question_id": "AH4UW", + "video_file": "Charades/videos/AH4UW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "AH4UW", + "question_id": "AH4UW", + "video_file": "Charades/videos/AH4UW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person closes the door behind them'." + }, + { + "from": "gpt", + "value": "From <1.4> to <8.5>." + } + ] + }, + { + "video_id": "AH4UW", + "question_id": "AH4UW", + "video_file": "Charades/videos/AH4UW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put the thing on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "SOTQ1", + "question_id": "SOTQ1", + "video_file": "Charades/videos/SOTQ1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person laughs' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.9> to <31.3>." + } + ] + }, + { + "video_id": "G752J", + "question_id": "G752J", + "video_file": "Charades/videos/G752J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person playing with the light on a nightstand' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <25.7>." + } + ] + }, + { + "video_id": "G752J", + "question_id": "G752J", + "video_file": "Charades/videos/G752J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they are playing with the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <25.7>." + } + ] + }, + { + "video_id": "G752J", + "question_id": "G752J", + "video_file": "Charades/videos/G752J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person is sitting on the bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.1> to <19.5>." + } + ] + }, + { + "video_id": "G752J", + "question_id": "G752J", + "video_file": "Charades/videos/G752J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person is flicking the lights on' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.5> to <25.7>." + } + ] + }, + { + "video_id": "G752J", + "question_id": "G752J", + "video_file": "Charades/videos/G752J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person walks through a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.9>." + } + ] + }, + { + "video_id": "F8E8C", + "question_id": "F8E8C", + "video_file": "Charades/videos/F8E8C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a drink of a glass of water'." + }, + { + "from": "gpt", + "value": "From <12.6> to <26.1>." + } + ] + }, + { + "video_id": "F8E8C", + "question_id": "F8E8C", + "video_file": "Charades/videos/F8E8C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take a drink from a cup on their counter' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.6> to <26.1>." + } + ] + }, + { + "video_id": "O8PO3", + "question_id": "O8PO3", + "video_file": "Charades/videos/O8PO3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person runs away' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <18.1>." + } + ] + }, + { + "video_id": "JL1W2", + "question_id": "JL1W2", + "video_file": "Charades/videos/JL1W2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is putting groceries on their table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.7> to <11.3>." + } + ] + }, + { + "video_id": "JOYAJ", + "question_id": "JOYAJ", + "video_file": "Charades/videos/JOYAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking on the stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.8>." + } + ] + }, + { + "video_id": "JOYAJ", + "question_id": "JOYAJ", + "video_file": "Charades/videos/JOYAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is cooking something on a stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.8>." + } + ] + }, + { + "video_id": "2K5GR", + "question_id": "2K5GR", + "video_file": "Charades/videos/2K5GR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take a drink from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <25.7>." + } + ] + }, + { + "video_id": "E33IO", + "question_id": "E33IO", + "video_file": "Charades/videos/E33IO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'starts laughing.the person turn' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.5> to <13.8>." + } + ] + }, + { + "video_id": "E33IO", + "question_id": "E33IO", + "video_file": "Charades/videos/E33IO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they laugh as the leave through the doorway'." + }, + { + "from": "gpt", + "value": "From <8.5> to <13.8>." + } + ] + }, + { + "video_id": "REH2U", + "question_id": "REH2U", + "video_file": "Charades/videos/REH2U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is lying on a bed snuggling a pillow' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.3>." + } + ] + }, + { + "video_id": "0RP6B", + "question_id": "0RP6B", + "video_file": "Charades/videos/0RP6B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person reads the book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.1> to <36.6>." + } + ] + }, + { + "video_id": "KDK62", + "question_id": "KDK62", + "video_file": "Charades/videos/KDK62.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes out a towel'." + }, + { + "from": "gpt", + "value": "From <1.9> to <9.3>." + } + ] + }, + { + "video_id": "8ISS6", + "question_id": "8ISS6", + "video_file": "Charades/videos/8ISS6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person starts holding a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.4> to <33.0>." + } + ] + }, + { + "video_id": "1DGC9", + "question_id": "1DGC9", + "video_file": "Charades/videos/1DGC9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person awakens in a closet snuggling their pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "1DGC9", + "question_id": "1DGC9", + "video_file": "Charades/videos/1DGC9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <14.1>." + } + ] + }, + { + "video_id": "1DGC9", + "question_id": "1DGC9", + "video_file": "Charades/videos/1DGC9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person close the door behind them' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.6> to <25.0>." + } + ] + }, + { + "video_id": "1DGC9", + "question_id": "1DGC9", + "video_file": "Charades/videos/1DGC9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person awakens in the bathroom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "IZ2XX", + "question_id": "IZ2XX", + "video_file": "Charades/videos/IZ2XX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put the books down on a table'." + }, + { + "from": "gpt", + "value": "From <8.4> to <16.4>." + } + ] + }, + { + "video_id": "MDZVU", + "question_id": "MDZVU", + "video_file": "Charades/videos/MDZVU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sneezing at their desk' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.0> to <10.0>." + } + ] + }, + { + "video_id": "IYU90", + "question_id": "IYU90", + "video_file": "Charades/videos/IYU90.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is running' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <7.4>." + } + ] + }, + { + "video_id": "IYU90", + "question_id": "IYU90", + "video_file": "Charades/videos/IYU90.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person runs by'." + }, + { + "from": "gpt", + "value": "From <0.1> to <7.4>." + } + ] + }, + { + "video_id": "IYU90", + "question_id": "IYU90", + "video_file": "Charades/videos/IYU90.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runnings in' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.1> to <7.4>." + } + ] + }, + { + "video_id": "T9DF0", + "question_id": "T9DF0", + "video_file": "Charades/videos/T9DF0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person smiling at the homework' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.8> to <26.3>." + } + ] + }, + { + "video_id": "GLOMM", + "question_id": "GLOMM", + "video_file": "Charades/videos/GLOMM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "7S3KW", + "question_id": "7S3KW", + "video_file": "Charades/videos/7S3KW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting on the floor opening a box'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "KTXE0", + "question_id": "KTXE0", + "video_file": "Charades/videos/KTXE0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats some cereal' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.0> to <31.0>." + } + ] + }, + { + "video_id": "KTXE0", + "question_id": "KTXE0", + "video_file": "Charades/videos/KTXE0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating a snack'." + }, + { + "from": "gpt", + "value": "From <25.0> to <31.0>." + } + ] + }, + { + "video_id": "NZG0R", + "question_id": "NZG0R", + "video_file": "Charades/videos/NZG0R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is smiling'." + }, + { + "from": "gpt", + "value": "From <11.1> to <16.7>." + } + ] + }, + { + "video_id": "NZG0R", + "question_id": "NZG0R", + "video_file": "Charades/videos/NZG0R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closing the door to a refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <12.2>." + } + ] + }, + { + "video_id": "NZG0R", + "question_id": "NZG0R", + "video_file": "Charades/videos/NZG0R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes a refrigerator door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.0> to <12.2>." + } + ] + }, + { + "video_id": "WMBOO", + "question_id": "WMBOO", + "video_file": "Charades/videos/WMBOO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person sitting on the floor opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <7.3>." + } + ] + }, + { + "video_id": "WMBOO", + "question_id": "WMBOO", + "video_file": "Charades/videos/WMBOO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens the door to the pantry' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <7.5>." + } + ] + }, + { + "video_id": "II0EJ", + "question_id": "II0EJ", + "video_file": "Charades/videos/II0EJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person sitting on stairs holds a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.1>." + } + ] + }, + { + "video_id": "II0EJ", + "question_id": "II0EJ", + "video_file": "Charades/videos/II0EJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting the glass down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.6> to <14.9>." + } + ] + }, + { + "video_id": "II0EJ", + "question_id": "II0EJ", + "video_file": "Charades/videos/II0EJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.1> to <18.4>." + } + ] + }, + { + "video_id": "PE4BQ", + "question_id": "PE4BQ", + "video_file": "Charades/videos/PE4BQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is putting a sandwich on the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "L8120", + "question_id": "L8120", + "video_file": "Charades/videos/L8120.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is smiling' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.5> to <29.0>." + } + ] + }, + { + "video_id": "L8120", + "question_id": "L8120", + "video_file": "Charades/videos/L8120.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person came out into the kitchen holding a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <11.3>." + } + ] + }, + { + "video_id": "K7E8L", + "question_id": "K7E8L", + "video_file": "Charades/videos/K7E8L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is tidying the hall with a broom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.5>." + } + ] + }, + { + "video_id": "X16WJ", + "question_id": "X16WJ", + "video_file": "Charades/videos/X16WJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sits down in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.8>." + } + ] + }, + { + "video_id": "X16WJ", + "question_id": "X16WJ", + "video_file": "Charades/videos/X16WJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the cup on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "X16WJ", + "question_id": "X16WJ", + "video_file": "Charades/videos/X16WJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person sitting down in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.8>." + } + ] + }, + { + "video_id": "GNNAE", + "question_id": "GNNAE", + "video_file": "Charades/videos/GNNAE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking from a cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.7> to <18.5>." + } + ] + }, + { + "video_id": "2VCOB", + "question_id": "2VCOB", + "video_file": "Charades/videos/2VCOB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the fridge door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.0> to <26.1>." + } + ] + }, + { + "video_id": "2VCOB", + "question_id": "2VCOB", + "video_file": "Charades/videos/2VCOB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts closing the refrigerator nearby' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <26.1>." + } + ] + }, + { + "video_id": "2VCOB", + "question_id": "2VCOB", + "video_file": "Charades/videos/2VCOB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closed it's door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.0> to <26.1>." + } + ] + }, + { + "video_id": "2VCOB", + "question_id": "2VCOB", + "video_file": "Charades/videos/2VCOB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is opening a sandwich bag in the pantry'." + }, + { + "from": "gpt", + "value": "From <4.7> to <13.7>." + } + ] + }, + { + "video_id": "V6LP3", + "question_id": "V6LP3", + "video_file": "Charades/videos/V6LP3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.7> to <28.9>." + } + ] + }, + { + "video_id": "EN0CD", + "question_id": "EN0CD", + "video_file": "Charades/videos/EN0CD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they start undressing by taking their clothes off' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <32.1>." + } + ] + }, + { + "video_id": "EN0CD", + "question_id": "EN0CD", + "video_file": "Charades/videos/EN0CD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person in the hallway is running around'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.3>." + } + ] + }, + { + "video_id": "EN0CD", + "question_id": "EN0CD", + "video_file": "Charades/videos/EN0CD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs into the hallway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.3>." + } + ] + }, + { + "video_id": "673VM", + "question_id": "673VM", + "video_file": "Charades/videos/673VM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a camera from a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <155.1> to <163.5>." + } + ] + }, + { + "video_id": "673VM", + "question_id": "673VM", + "video_file": "Charades/videos/673VM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put it back on the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <142.6> to <159.1>." + } + ] + }, + { + "video_id": "673VM", + "question_id": "673VM", + "video_file": "Charades/videos/673VM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts it back on the shelf'." + }, + { + "from": "gpt", + "value": "From <142.6> to <159.1>." + } + ] + }, + { + "video_id": "673VM", + "question_id": "673VM", + "video_file": "Charades/videos/673VM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a camera off a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <155.1> to <163.5>." + } + ] + }, + { + "video_id": "673VM", + "question_id": "673VM", + "video_file": "Charades/videos/673VM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person walks through the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <91.7> to <102.8>." + } + ] + }, + { + "video_id": "C4DFA", + "question_id": "C4DFA", + "video_file": "Charades/videos/C4DFA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes an item from the bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "QO0UU", + "question_id": "QO0UU", + "video_file": "Charades/videos/QO0UU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <5.8>." + } + ] + }, + { + "video_id": "K4XO7", + "question_id": "K4XO7", + "video_file": "Charades/videos/K4XO7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person awakens on the sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.9> to <30.9>." + } + ] + }, + { + "video_id": "X3LEM", + "question_id": "X3LEM", + "video_file": "Charades/videos/X3LEM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits down on a sofa,' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.7> to <32.0>." + } + ] + }, + { + "video_id": "X3LEM", + "question_id": "X3LEM", + "video_file": "Charades/videos/X3LEM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person b begins to sit on a couch'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.7> to <32.0>." + } + ] + }, + { + "video_id": "X3LEM", + "question_id": "X3LEM", + "video_file": "Charades/videos/X3LEM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turning on a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.5> to <19.4>." + } + ] + }, + { + "video_id": "X3LEM", + "question_id": "X3LEM", + "video_file": "Charades/videos/X3LEM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person turns off the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.3> to <21.5>." + } + ] + }, + { + "video_id": "X3LEM", + "question_id": "X3LEM", + "video_file": "Charades/videos/X3LEM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person takes a phone from a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.5> to <29.0>." + } + ] + }, + { + "video_id": "YYHX3", + "question_id": "YYHX3", + "video_file": "Charades/videos/YYHX3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put a towel on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <11.6>." + } + ] + }, + { + "video_id": "YYHX3", + "question_id": "YYHX3", + "video_file": "Charades/videos/YYHX3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting it on the table as well' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <11.6>." + } + ] + }, + { + "video_id": "YYHX3", + "question_id": "YYHX3", + "video_file": "Charades/videos/YYHX3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is putting a towel on a table'." + }, + { + "from": "gpt", + "value": "From <5.2> to <11.6>." + } + ] + }, + { + "video_id": "NBE4C", + "question_id": "NBE4C", + "video_file": "Charades/videos/NBE4C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <27.6> to <32.0>." + } + ] + }, + { + "video_id": "NBE4C", + "question_id": "NBE4C", + "video_file": "Charades/videos/NBE4C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person take medicine' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.5> to <32.0>." + } + ] + }, + { + "video_id": "NBE4C", + "question_id": "NBE4C", + "video_file": "Charades/videos/NBE4C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.6> to <32.0>." + } + ] + }, + { + "video_id": "NBE4C", + "question_id": "NBE4C", + "video_file": "Charades/videos/NBE4C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts the phone in a pocket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.5> to <22.2>." + } + ] + }, + { + "video_id": "61RMD", + "question_id": "61RMD", + "video_file": "Charades/videos/61RMD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the camera on a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.9>." + } + ] + }, + { + "video_id": "RB9VI", + "question_id": "RB9VI", + "video_file": "Charades/videos/RB9VI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person in an entryway holding a broom sweeping' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.3>." + } + ] + }, + { + "video_id": "RB9VI", + "question_id": "RB9VI", + "video_file": "Charades/videos/RB9VI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they pick up a bag take out a bottle' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.6> to <24.0>." + } + ] + }, + { + "video_id": "RB9VI", + "question_id": "RB9VI", + "video_file": "Charades/videos/RB9VI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take a bottle out of a backpack' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.6> to <24.0>." + } + ] + }, + { + "video_id": "RB9VI", + "question_id": "RB9VI", + "video_file": "Charades/videos/RB9VI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws the broom down'." + }, + { + "from": "gpt", + "value": "From <7.9> to <14.9>." + } + ] + }, + { + "video_id": "J2J3Q", + "question_id": "J2J3Q", + "video_file": "Charades/videos/J2J3Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person starts sneezing on a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <22.6>." + } + ] + }, + { + "video_id": "9JHH9", + "question_id": "9JHH9", + "video_file": "Charades/videos/9JHH9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person close the door behind them' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.8> to <30.0>." + } + ] + }, + { + "video_id": "TSI2S", + "question_id": "TSI2S", + "video_file": "Charades/videos/TSI2S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts a broom under the bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.8> to <22.2>." + } + ] + }, + { + "video_id": "TSI2S", + "question_id": "TSI2S", + "video_file": "Charades/videos/TSI2S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs into a bedroom holding a broom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.5> to <22.7>." + } + ] + }, + { + "video_id": "TSI2S", + "question_id": "TSI2S", + "video_file": "Charades/videos/TSI2S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs through the room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.0> to <17.0>." + } + ] + }, + { + "video_id": "M5XFI", + "question_id": "M5XFI", + "video_file": "Charades/videos/M5XFI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person begins laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.5> to <24.5>." + } + ] + }, + { + "video_id": "7JZTY", + "question_id": "7JZTY", + "video_file": "Charades/videos/7JZTY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a drink from a soda bottle' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.9> to <31.5>." + } + ] + }, + { + "video_id": "7JZTY", + "question_id": "7JZTY", + "video_file": "Charades/videos/7JZTY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinks from a bottle of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.9> to <31.5>." + } + ] + }, + { + "video_id": "7JZTY", + "question_id": "7JZTY", + "video_file": "Charades/videos/7JZTY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating from a bag of food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <14.0>." + } + ] + }, + { + "video_id": "7JZTY", + "question_id": "7JZTY", + "video_file": "Charades/videos/7JZTY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats the food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <14.0>." + } + ] + }, + { + "video_id": "7JZTY", + "question_id": "7JZTY", + "video_file": "Charades/videos/7JZTY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating some' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <14.0>." + } + ] + }, + { + "video_id": "LBG7L", + "question_id": "LBG7L", + "video_file": "Charades/videos/LBG7L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts sits in a chair'." + }, + { + "from": "gpt", + "value": "From <19.0> to <32.0>." + } + ] + }, + { + "video_id": "LBG7L", + "question_id": "LBG7L", + "video_file": "Charades/videos/LBG7L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes a laptop out from the bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.1> to <10.7>." + } + ] + }, + { + "video_id": "LBG7L", + "question_id": "LBG7L", + "video_file": "Charades/videos/LBG7L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes a laptop computer out of a bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.1> to <10.7>." + } + ] + }, + { + "video_id": "LBG7L", + "question_id": "LBG7L", + "video_file": "Charades/videos/LBG7L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person sits in the chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <32.0>." + } + ] + }, + { + "video_id": "LBG7L", + "question_id": "LBG7L", + "video_file": "Charades/videos/LBG7L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins working on the laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.5> to <32.0>." + } + ] + }, + { + "video_id": "TE34B", + "question_id": "TE34B", + "video_file": "Charades/videos/TE34B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person works on a laptop'." + }, + { + "from": "gpt", + "value": "From <1.3> to <15.5>." + } + ] + }, + { + "video_id": "TE34B", + "question_id": "TE34B", + "video_file": "Charades/videos/TE34B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person walk out of the room holding the phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.6> to <26.5>." + } + ] + }, + { + "video_id": "TE34B", + "question_id": "TE34B", + "video_file": "Charades/videos/TE34B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they reenter the room holding a bag of groceries' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <55.4> to <68.5>." + } + ] + }, + { + "video_id": "9T11N", + "question_id": "9T11N", + "video_file": "Charades/videos/9T11N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looking at a desk picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.2>." + } + ] + }, + { + "video_id": "9T11N", + "question_id": "9T11N", + "video_file": "Charades/videos/9T11N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person picks up a picture to look at' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.2>." + } + ] + }, + { + "video_id": "9T11N", + "question_id": "9T11N", + "video_file": "Charades/videos/9T11N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looking at the picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.2>." + } + ] + }, + { + "video_id": "45FWP", + "question_id": "45FWP", + "video_file": "Charades/videos/45FWP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'this person is eating some dough'." + }, + { + "from": "gpt", + "value": "From <8.3> to <25.8>." + } + ] + }, + { + "video_id": "45FWP", + "question_id": "45FWP", + "video_file": "Charades/videos/45FWP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin to eat' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.3> to <25.8>." + } + ] + }, + { + "video_id": "45FWP", + "question_id": "45FWP", + "video_file": "Charades/videos/45FWP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats by the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <25.8>." + } + ] + }, + { + "video_id": "K9GJF", + "question_id": "K9GJF", + "video_file": "Charades/videos/K9GJF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person closes a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.2> to <17.9>." + } + ] + }, + { + "video_id": "D7M6N", + "question_id": "D7M6N", + "video_file": "Charades/videos/D7M6N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is opening a bag of groceries' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "D7M6N", + "question_id": "D7M6N", + "video_file": "Charades/videos/D7M6N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sneezing on a picture in a kitchen'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.9> to <32.0>." + } + ] + }, + { + "video_id": "RBAO4", + "question_id": "RBAO4", + "video_file": "Charades/videos/RBAO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person beginning to eat from a bag of potato chips' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.8> to <32.0>." + } + ] + }, + { + "video_id": "RBAO4", + "question_id": "RBAO4", + "video_file": "Charades/videos/RBAO4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opens the front door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.7> to <15.7>." + } + ] + }, + { + "video_id": "LJ0JS", + "question_id": "LJ0JS", + "video_file": "Charades/videos/LJ0JS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person periodically takes a drink from a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.1>." + } + ] + }, + { + "video_id": "89MAO", + "question_id": "89MAO", + "video_file": "Charades/videos/89MAO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a picture of it with a camera' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.9> to <25.0>." + } + ] + }, + { + "video_id": "0OSJY", + "question_id": "0OSJY", + "video_file": "Charades/videos/0OSJY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person sits on their bed holding a broom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.6>." + } + ] + }, + { + "video_id": "0OSJY", + "question_id": "0OSJY", + "video_file": "Charades/videos/0OSJY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs to touch the doorknob' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <15.7>." + } + ] + }, + { + "video_id": "0OSJY", + "question_id": "0OSJY", + "video_file": "Charades/videos/0OSJY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting on a bed holding a broom'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.6>." + } + ] + }, + { + "video_id": "ILQ2R", + "question_id": "ILQ2R", + "video_file": "Charades/videos/ILQ2R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they turn off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.3> to <23.3>." + } + ] + }, + { + "video_id": "OL2JP", + "question_id": "OL2JP", + "video_file": "Charades/videos/OL2JP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person fixes a broken door knob at the bathroom door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.1>." + } + ] + }, + { + "video_id": "OL2JP", + "question_id": "OL2JP", + "video_file": "Charades/videos/OL2JP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is fixing a doorknob'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.1>." + } + ] + }, + { + "video_id": "AQ5M6", + "question_id": "AQ5M6", + "video_file": "Charades/videos/AQ5M6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person getting dressed'." + }, + { + "from": "gpt", + "value": "From <15.8> to <30.0>." + } + ] + }, + { + "video_id": "AQ5M6", + "question_id": "AQ5M6", + "video_file": "Charades/videos/AQ5M6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns the light on' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.8> to <9.5>." + } + ] + }, + { + "video_id": "AQ5M6", + "question_id": "AQ5M6", + "video_file": "Charades/videos/AQ5M6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turned on the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.8> to <9.5>." + } + ] + }, + { + "video_id": "AQ5M6", + "question_id": "AQ5M6", + "video_file": "Charades/videos/AQ5M6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person turns a light on'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <9.5>." + } + ] + }, + { + "video_id": "24XHS", + "question_id": "24XHS", + "video_file": "Charades/videos/24XHS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <24.1>." + } + ] + }, + { + "video_id": "24XHS", + "question_id": "24XHS", + "video_file": "Charades/videos/24XHS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person laughing in the kitchen'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.0> to <24.1>." + } + ] + }, + { + "video_id": "24XHS", + "question_id": "24XHS", + "video_file": "Charades/videos/24XHS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opening the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.3> to <26.3>." + } + ] + }, + { + "video_id": "T5JY8", + "question_id": "T5JY8", + "video_file": "Charades/videos/T5JY8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes a drink from a cup of coffee' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.4> to <23.5>." + } + ] + }, + { + "video_id": "0VMUZ", + "question_id": "0VMUZ", + "video_file": "Charades/videos/0VMUZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding a picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.5> to <16.7>." + } + ] + }, + { + "video_id": "0VMUZ", + "question_id": "0VMUZ", + "video_file": "Charades/videos/0VMUZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'in the laundry room the person is washing their clothes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.7>." + } + ] + }, + { + "video_id": "1L220", + "question_id": "1L220", + "video_file": "Charades/videos/1L220.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they begin to eat as they use their phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <73.4> to <129.0>." + } + ] + }, + { + "video_id": "2CDOS", + "question_id": "2CDOS", + "video_file": "Charades/videos/2CDOS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws them on the floor'." + }, + { + "from": "gpt", + "value": "From <10.6> to <20.7>." + } + ] + }, + { + "video_id": "RH44U", + "question_id": "RH44U", + "video_file": "Charades/videos/RH44U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a book off the shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.1> to <12.8>." + } + ] + }, + { + "video_id": "RH44U", + "question_id": "RH44U", + "video_file": "Charades/videos/RH44U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person holding a cup of coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.8>." + } + ] + }, + { + "video_id": "DZ0SB", + "question_id": "DZ0SB", + "video_file": "Charades/videos/DZ0SB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a book'." + }, + { + "from": "gpt", + "value": "From <21.8> to <28.1>." + } + ] + }, + { + "video_id": "DZ0SB", + "question_id": "DZ0SB", + "video_file": "Charades/videos/DZ0SB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person puts down a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.5> to <29.7>." + } + ] + }, + { + "video_id": "3SDJE", + "question_id": "3SDJE", + "video_file": "Charades/videos/3SDJE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts a cook book back where it belongs' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.8> to <20.2>." + } + ] + }, + { + "video_id": "3SDJE", + "question_id": "3SDJE", + "video_file": "Charades/videos/3SDJE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person closes the book'." + }, + { + "from": "gpt", + "value": "From <7.0> to <14.1>." + } + ] + }, + { + "video_id": "3SDJE", + "question_id": "3SDJE", + "video_file": "Charades/videos/3SDJE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the book inside the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.8> to <20.2>." + } + ] + }, + { + "video_id": "GMMVC", + "question_id": "GMMVC", + "video_file": "Charades/videos/GMMVC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person running to the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.3> to <19.0>." + } + ] + }, + { + "video_id": "Y25DQ", + "question_id": "Y25DQ", + "video_file": "Charades/videos/Y25DQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.4> to <17.2>." + } + ] + }, + { + "video_id": "Y25DQ", + "question_id": "Y25DQ", + "video_file": "Charades/videos/Y25DQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.4> to <17.2>." + } + ] + }, + { + "video_id": "W4W97", + "question_id": "W4W97", + "video_file": "Charades/videos/W4W97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take a drink from a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.9> to <21.6>." + } + ] + }, + { + "video_id": "W4W97", + "question_id": "W4W97", + "video_file": "Charades/videos/W4W97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drink from the glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <21.6>." + } + ] + }, + { + "video_id": "FQ6OB", + "question_id": "FQ6OB", + "video_file": "Charades/videos/FQ6OB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.3> to <25.3>." + } + ] + }, + { + "video_id": "CRTUY", + "question_id": "CRTUY", + "video_file": "Charades/videos/CRTUY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throw something on the floor'." + }, + { + "from": "gpt", + "value": "From <17.7> to <22.1>." + } + ] + }, + { + "video_id": "CRTUY", + "question_id": "CRTUY", + "video_file": "Charades/videos/CRTUY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door to it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.1> to <28.0>." + } + ] + }, + { + "video_id": "CRTUY", + "question_id": "CRTUY", + "video_file": "Charades/videos/CRTUY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <10.9>." + } + ] + }, + { + "video_id": "93ESZ", + "question_id": "93ESZ", + "video_file": "Charades/videos/93ESZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person dresses up by the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.6> to <12.1>." + } + ] + }, + { + "video_id": "EWCV4", + "question_id": "EWCV4", + "video_file": "Charades/videos/EWCV4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes his shoes off' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.4> to <17.1>." + } + ] + }, + { + "video_id": "EWCV4", + "question_id": "EWCV4", + "video_file": "Charades/videos/EWCV4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a man is laying on a floor in the doorway'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "NKTWF", + "question_id": "NKTWF", + "video_file": "Charades/videos/NKTWF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks into the living room holding a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.9>." + } + ] + }, + { + "video_id": "NKTWF", + "question_id": "NKTWF", + "video_file": "Charades/videos/NKTWF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the bag on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.9>." + } + ] + }, + { + "video_id": "NKTWF", + "question_id": "NKTWF", + "video_file": "Charades/videos/NKTWF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'this person walks in holding a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.9>." + } + ] + }, + { + "video_id": "PQUF6", + "question_id": "PQUF6", + "video_file": "Charades/videos/PQUF6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person washes their hands' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <14.0>." + } + ] + }, + { + "video_id": "PQUF6", + "question_id": "PQUF6", + "video_file": "Charades/videos/PQUF6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person try to open a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <19.1>." + } + ] + }, + { + "video_id": "PQUF6", + "question_id": "PQUF6", + "video_file": "Charades/videos/PQUF6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person tries opening a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.5> to <19.1>." + } + ] + }, + { + "video_id": "R08G9", + "question_id": "R08G9", + "video_file": "Charades/videos/R08G9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.1> to <27.3>." + } + ] + }, + { + "video_id": "QDRWK", + "question_id": "QDRWK", + "video_file": "Charades/videos/QDRWK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "QDRWK", + "question_id": "QDRWK", + "video_file": "Charades/videos/QDRWK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts a book on a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.8> to <30.0>." + } + ] + }, + { + "video_id": "QDRWK", + "question_id": "QDRWK", + "video_file": "Charades/videos/QDRWK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts on shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.7> to <30.0>." + } + ] + }, + { + "video_id": "7EYIL", + "question_id": "7EYIL", + "video_file": "Charades/videos/7EYIL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person is opening a refrigerator' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <19.1>." + } + ] + }, + { + "video_id": "7EYIL", + "question_id": "7EYIL", + "video_file": "Charades/videos/7EYIL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes out a bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.3> to <22.9>." + } + ] + }, + { + "video_id": "7EYIL", + "question_id": "7EYIL", + "video_file": "Charades/videos/7EYIL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they open the refrigerator'." + }, + { + "from": "gpt", + "value": "From <12.8> to <19.1>." + } + ] + }, + { + "video_id": "KWONB", + "question_id": "KWONB", + "video_file": "Charades/videos/KWONB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is throwing clothes into a closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.0> to <8.4>." + } + ] + }, + { + "video_id": "KWONB", + "question_id": "KWONB", + "video_file": "Charades/videos/KWONB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens the closet door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "6CMWO", + "question_id": "6CMWO", + "video_file": "Charades/videos/6CMWO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is undressing'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "6CMWO", + "question_id": "6CMWO", + "video_file": "Charades/videos/6CMWO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.9> to <27.8>." + } + ] + }, + { + "video_id": "ZWE6N", + "question_id": "ZWE6N", + "video_file": "Charades/videos/ZWE6N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person takes a cup of coffee from the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.0> to <8.8>." + } + ] + }, + { + "video_id": "ZWE6N", + "question_id": "ZWE6N", + "video_file": "Charades/videos/ZWE6N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person takes a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <29.0>." + } + ] + }, + { + "video_id": "6NWDR", + "question_id": "6NWDR", + "video_file": "Charades/videos/6NWDR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes a bottle of medicine from the desk'." + }, + { + "from": "gpt", + "value": "From <25.1> to <32.7>." + } + ] + }, + { + "video_id": "6NWDR", + "question_id": "6NWDR", + "video_file": "Charades/videos/6NWDR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sneezing'." + }, + { + "from": "gpt", + "value": "From <1.2> to <14.9>." + } + ] + }, + { + "video_id": "0EGNU", + "question_id": "0EGNU", + "video_file": "Charades/videos/0EGNU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the pillow on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.7> to <9.0>." + } + ] + }, + { + "video_id": "UBC9N", + "question_id": "UBC9N", + "video_file": "Charades/videos/UBC9N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins pouring it into a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.4> to <24.4>." + } + ] + }, + { + "video_id": "UBC9N", + "question_id": "UBC9N", + "video_file": "Charades/videos/UBC9N.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pours medicine into a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.4> to <24.4>." + } + ] + }, + { + "video_id": "8MR9I", + "question_id": "8MR9I", + "video_file": "Charades/videos/8MR9I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.8> to <6.5>." + } + ] + }, + { + "video_id": "8MR9I", + "question_id": "8MR9I", + "video_file": "Charades/videos/8MR9I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person talking on the phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <6.8>." + } + ] + }, + { + "video_id": "UK8KW", + "question_id": "UK8KW", + "video_file": "Charades/videos/UK8KW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eat a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.3> to <31.0>." + } + ] + }, + { + "video_id": "UK8KW", + "question_id": "UK8KW", + "video_file": "Charades/videos/UK8KW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person standing in a doorway eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.3> to <31.0>." + } + ] + }, + { + "video_id": "UK8KW", + "question_id": "UK8KW", + "video_file": "Charades/videos/UK8KW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person stands in the door holding a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.2> to <31.0>." + } + ] + }, + { + "video_id": "07AWM", + "question_id": "07AWM", + "video_file": "Charades/videos/07AWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person snuggling a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.1> to <20.2>." + } + ] + }, + { + "video_id": "07AWM", + "question_id": "07AWM", + "video_file": "Charades/videos/07AWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins to laugh'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <25.0>." + } + ] + }, + { + "video_id": "5S1WQ", + "question_id": "5S1WQ", + "video_file": "Charades/videos/5S1WQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person throws a broom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.6>." + } + ] + }, + { + "video_id": "OVMCM", + "question_id": "OVMCM", + "video_file": "Charades/videos/OVMCM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person takes a towel from the shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.9> to <19.8>." + } + ] + }, + { + "video_id": "OVMCM", + "question_id": "OVMCM", + "video_file": "Charades/videos/OVMCM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakens sitting in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.2>." + } + ] + }, + { + "video_id": "OVMCM", + "question_id": "OVMCM", + "video_file": "Charades/videos/OVMCM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens in a chair in their recreation room'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "LWINQ", + "question_id": "LWINQ", + "video_file": "Charades/videos/LWINQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person begins undressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "LWINQ", + "question_id": "LWINQ", + "video_file": "Charades/videos/LWINQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sit on the sofa with blanket over the head' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.8> to <17.8>." + } + ] + }, + { + "video_id": "LWINQ", + "question_id": "LWINQ", + "video_file": "Charades/videos/LWINQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits on a couch' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.8> to <17.8>." + } + ] + }, + { + "video_id": "LWINQ", + "question_id": "LWINQ", + "video_file": "Charades/videos/LWINQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person sits on the sofa'." + }, + { + "from": "gpt", + "value": "From <9.8> to <17.8>." + } + ] + }, + { + "video_id": "3C8OG", + "question_id": "3C8OG", + "video_file": "Charades/videos/3C8OG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take a bottle' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.7> to <5.6>." + } + ] + }, + { + "video_id": "3C8OG", + "question_id": "3C8OG", + "video_file": "Charades/videos/3C8OG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person the opens a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.6> to <29.2>." + } + ] + }, + { + "video_id": "3C8OG", + "question_id": "3C8OG", + "video_file": "Charades/videos/3C8OG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take a book from their bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.9> to <23.5>." + } + ] + }, + { + "video_id": "3C8OG", + "question_id": "3C8OG", + "video_file": "Charades/videos/3C8OG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is working on a laptop'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "NKOJ8", + "question_id": "NKOJ8", + "video_file": "Charades/videos/NKOJ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens the door to pick up a device' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.6> to <13.9>." + } + ] + }, + { + "video_id": "NKOJ8", + "question_id": "NKOJ8", + "video_file": "Charades/videos/NKOJ8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person walks through the doorway into an office' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.0> to <17.9>." + } + ] + }, + { + "video_id": "DOGYU", + "question_id": "DOGYU", + "video_file": "Charades/videos/DOGYU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes out a shoe box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <8.7>." + } + ] + }, + { + "video_id": "DOGYU", + "question_id": "DOGYU", + "video_file": "Charades/videos/DOGYU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'there is a person sitting in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.8> to <16.0>." + } + ] + }, + { + "video_id": "DOGYU", + "question_id": "DOGYU", + "video_file": "Charades/videos/DOGYU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person in the chair is smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.9> to <12.7>." + } + ] + }, + { + "video_id": "DOGYU", + "question_id": "DOGYU", + "video_file": "Charades/videos/DOGYU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person is sitting in a chair'." + }, + { + "from": "gpt", + "value": "From <3.8> to <16.0>." + } + ] + }, + { + "video_id": "DOGYU", + "question_id": "DOGYU", + "video_file": "Charades/videos/DOGYU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.6>." + } + ] + }, + { + "video_id": "DOGYU", + "question_id": "DOGYU", + "video_file": "Charades/videos/DOGYU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person recloses the box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.3> to <32.0>." + } + ] + }, + { + "video_id": "A96EQ", + "question_id": "A96EQ", + "video_file": "Charades/videos/A96EQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person pushes the door open' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "A96EQ", + "question_id": "A96EQ", + "video_file": "Charades/videos/A96EQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks through the doorway into the room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <7.5>." + } + ] + }, + { + "video_id": "3XDQC", + "question_id": "3XDQC", + "video_file": "Charades/videos/3XDQC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person with hair in a towel is smiling' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "USNP1", + "question_id": "USNP1", + "video_file": "Charades/videos/USNP1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take a phone out of their pocket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.3> to <11.4>." + } + ] + }, + { + "video_id": "VSAA7", + "question_id": "VSAA7", + "video_file": "Charades/videos/VSAA7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin snuggling under a blanket in the living room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <30.6> to <37.0>." + } + ] + }, + { + "video_id": "IXISD", + "question_id": "IXISD", + "video_file": "Charades/videos/IXISD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person is putting their shoes on the floor'." + }, + { + "from": "gpt", + "value": "From <10.7> to <15.6>." + } + ] + }, + { + "video_id": "IXISD", + "question_id": "IXISD", + "video_file": "Charades/videos/IXISD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.9> to <13.1>." + } + ] + }, + { + "video_id": "IXISD", + "question_id": "IXISD", + "video_file": "Charades/videos/IXISD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put their feet up on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <18.7>." + } + ] + }, + { + "video_id": "IXISD", + "question_id": "IXISD", + "video_file": "Charades/videos/IXISD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take off their shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.2> to <13.5>." + } + ] + }, + { + "video_id": "IXISD", + "question_id": "IXISD", + "video_file": "Charades/videos/IXISD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they laugh' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <13.1>." + } + ] + }, + { + "video_id": "D1D55", + "question_id": "D1D55", + "video_file": "Charades/videos/D1D55.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sitting in the office chair for a moment' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.1> to <36.7>." + } + ] + }, + { + "video_id": "D1D55", + "question_id": "D1D55", + "video_file": "Charades/videos/D1D55.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.1> to <36.7>." + } + ] + }, + { + "video_id": "7SUTI", + "question_id": "7SUTI", + "video_file": "Charades/videos/7SUTI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person looks out the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.9> to <29.5>." + } + ] + }, + { + "video_id": "SWZWL", + "question_id": "SWZWL", + "video_file": "Charades/videos/SWZWL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'takes a picture of the second person' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.7> to <33.0>." + } + ] + }, + { + "video_id": "8K380", + "question_id": "8K380", + "video_file": "Charades/videos/8K380.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person the other was sneezing into a pillow'." + }, + { + "from": "gpt", + "value": "From <16.8> to <32.0>." + } + ] + }, + { + "video_id": "00T4B", + "question_id": "00T4B", + "video_file": "Charades/videos/00T4B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they ended by putting on their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.9> to <20.3>." + } + ] + }, + { + "video_id": "00T4B", + "question_id": "00T4B", + "video_file": "Charades/videos/00T4B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts on shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.9> to <20.3>." + } + ] + }, + { + "video_id": "ZEID0", + "question_id": "ZEID0", + "video_file": "Charades/videos/ZEID0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person is sneezing on homework'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.0>." + } + ] + }, + { + "video_id": "RZDEA", + "question_id": "RZDEA", + "video_file": "Charades/videos/RZDEA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is smiling with a glass in their hand' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.7> to <14.6>." + } + ] + }, + { + "video_id": "RZDEA", + "question_id": "RZDEA", + "video_file": "Charades/videos/RZDEA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is holding a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.9>." + } + ] + }, + { + "video_id": "EIK9W", + "question_id": "EIK9W", + "video_file": "Charades/videos/EIK9W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <6.8>." + } + ] + }, + { + "video_id": "I049A", + "question_id": "I049A", + "video_file": "Charades/videos/I049A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put on some shoes'." + }, + { + "from": "gpt", + "value": "From <23.7> to <31.0>." + } + ] + }, + { + "video_id": "D7KU2", + "question_id": "D7KU2", + "video_file": "Charades/videos/D7KU2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is putting a broom in a corner' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.7> to <14.1>." + } + ] + }, + { + "video_id": "5XLBY", + "question_id": "5XLBY", + "video_file": "Charades/videos/5XLBY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts book on table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <27.7> to <32.0>." + } + ] + }, + { + "video_id": "E6IL3", + "question_id": "E6IL3", + "video_file": "Charades/videos/E6IL3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting on the floor reading a magazine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "E6IL3", + "question_id": "E6IL3", + "video_file": "Charades/videos/E6IL3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is smiling'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "E6IL3", + "question_id": "E6IL3", + "video_file": "Charades/videos/E6IL3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person reading a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.1>." + } + ] + }, + { + "video_id": "YBVVZ", + "question_id": "YBVVZ", + "video_file": "Charades/videos/YBVVZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person starts to tidy up a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "YBVVZ", + "question_id": "YBVVZ", + "video_file": "Charades/videos/YBVVZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put food on a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.2> to <23.1>." + } + ] + }, + { + "video_id": "YBVVZ", + "question_id": "YBVVZ", + "video_file": "Charades/videos/YBVVZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is tidying a table'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.4>." + } + ] + }, + { + "video_id": "YBVVZ", + "question_id": "YBVVZ", + "video_file": "Charades/videos/YBVVZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting food on a shelf'." + }, + { + "from": "gpt", + "value": "From <15.2> to <23.1>." + } + ] + }, + { + "video_id": "WPFKR", + "question_id": "WPFKR", + "video_file": "Charades/videos/WPFKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.5> to <27.4>." + } + ] + }, + { + "video_id": "WPFKR", + "question_id": "WPFKR", + "video_file": "Charades/videos/WPFKR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes off their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <7.2>." + } + ] + }, + { + "video_id": "SJ6X6", + "question_id": "SJ6X6", + "video_file": "Charades/videos/SJ6X6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person runs up the driveway into the garage' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <10.3>." + } + ] + }, + { + "video_id": "SJ6X6", + "question_id": "SJ6X6", + "video_file": "Charades/videos/SJ6X6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws the book into red pillow over' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.9> to <12.8>." + } + ] + }, + { + "video_id": "YH7K4", + "question_id": "YH7K4", + "video_file": "Charades/videos/YH7K4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.9> to <30.0>." + } + ] + }, + { + "video_id": "YH7K4", + "question_id": "YH7K4", + "video_file": "Charades/videos/YH7K4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking a cup of coffee' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <16.4>." + } + ] + }, + { + "video_id": "ZRITA", + "question_id": "ZRITA", + "video_file": "Charades/videos/ZRITA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "ZRITA", + "question_id": "ZRITA", + "video_file": "Charades/videos/ZRITA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person one on sofa eating'." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.6>." + } + ] + }, + { + "video_id": "D52S5", + "question_id": "D52S5", + "video_file": "Charades/videos/D52S5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person he sits down in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.4> to <23.6>." + } + ] + }, + { + "video_id": "D52S5", + "question_id": "D52S5", + "video_file": "Charades/videos/D52S5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a seat at the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.2> to <12.9>." + } + ] + }, + { + "video_id": "D52S5", + "question_id": "D52S5", + "video_file": "Charades/videos/D52S5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person reaches behind the person to open the cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.2> to <33.0>." + } + ] + }, + { + "video_id": "D52S5", + "question_id": "D52S5", + "video_file": "Charades/videos/D52S5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person grabs a drink from a bottle'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.4> to <33.0>." + } + ] + }, + { + "video_id": "D52S5", + "question_id": "D52S5", + "video_file": "Charades/videos/D52S5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats some more food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.2> to <12.9>." + } + ] + }, + { + "video_id": "D52S5", + "question_id": "D52S5", + "video_file": "Charades/videos/D52S5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats it'." + }, + { + "from": "gpt", + "value": "From <6.2> to <12.9>." + } + ] + }, + { + "video_id": "IWF0U", + "question_id": "IWF0U", + "video_file": "Charades/videos/IWF0U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sits on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.6> to <21.0>." + } + ] + }, + { + "video_id": "IWF0U", + "question_id": "IWF0U", + "video_file": "Charades/videos/IWF0U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is pouring a drink in a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.4> to <19.3>." + } + ] + }, + { + "video_id": "IWF0U", + "question_id": "IWF0U", + "video_file": "Charades/videos/IWF0U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours a drink into a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.4> to <19.3>." + } + ] + }, + { + "video_id": "K5IAK", + "question_id": "K5IAK", + "video_file": "Charades/videos/K5IAK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person enters the bedroom holding a vacuum' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <17.1>." + } + ] + }, + { + "video_id": "1J3I0", + "question_id": "1J3I0", + "video_file": "Charades/videos/1J3I0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting the laptop on the couch' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <7.1>." + } + ] + }, + { + "video_id": "1J3I0", + "question_id": "1J3I0", + "video_file": "Charades/videos/1J3I0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person closes a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.8>." + } + ] + }, + { + "video_id": "GR5N6", + "question_id": "GR5N6", + "video_file": "Charades/videos/GR5N6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they stand up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.1> to <18.0>." + } + ] + }, + { + "video_id": "GR5N6", + "question_id": "GR5N6", + "video_file": "Charades/videos/GR5N6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throw the book across the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.2> to <17.9>." + } + ] + }, + { + "video_id": "GR5N6", + "question_id": "GR5N6", + "video_file": "Charades/videos/GR5N6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throw the book through a doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.2> to <17.9>." + } + ] + }, + { + "video_id": "KW3BO", + "question_id": "KW3BO", + "video_file": "Charades/videos/KW3BO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person taking a drink of water from a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.6> to <13.1>." + } + ] + }, + { + "video_id": "KW3BO", + "question_id": "KW3BO", + "video_file": "Charades/videos/KW3BO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person eats some food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "KW3BO", + "question_id": "KW3BO", + "video_file": "Charades/videos/KW3BO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person is eating box of candy' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "KW3BO", + "question_id": "KW3BO", + "video_file": "Charades/videos/KW3BO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts laughing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <23.9>." + } + ] + }, + { + "video_id": "FHXKQ", + "question_id": "FHXKQ", + "video_file": "Charades/videos/FHXKQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is putting some groceries on a table'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "A2771", + "question_id": "A2771", + "video_file": "Charades/videos/A2771.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person takes a laptop from a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.7> to <32.0>." + } + ] + }, + { + "video_id": "YAPN7", + "question_id": "YAPN7", + "video_file": "Charades/videos/YAPN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits on the couch' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.6> to <17.5>." + } + ] + }, + { + "video_id": "YAPN7", + "question_id": "YAPN7", + "video_file": "Charades/videos/YAPN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a bite from a piece of food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.4> to <38.4>." + } + ] + }, + { + "video_id": "YAPN7", + "question_id": "YAPN7", + "video_file": "Charades/videos/YAPN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person undresses in the recreation room'." + }, + { + "from": "gpt", + "value": "From <2.3> to <6.6>." + } + ] + }, + { + "video_id": "YAPN7", + "question_id": "YAPN7", + "video_file": "Charades/videos/YAPN7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating some food from a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.8> to <8.6>." + } + ] + }, + { + "video_id": "CIBGY", + "question_id": "CIBGY", + "video_file": "Charades/videos/CIBGY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person washes their hands in a sink'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.4>." + } + ] + }, + { + "video_id": "CIBGY", + "question_id": "CIBGY", + "video_file": "Charades/videos/CIBGY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person washing hands at the sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.4>." + } + ] + }, + { + "video_id": "CIBGY", + "question_id": "CIBGY", + "video_file": "Charades/videos/CIBGY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open a door'." + }, + { + "from": "gpt", + "value": "From <24.2> to <29.8>." + } + ] + }, + { + "video_id": "CIBGY", + "question_id": "CIBGY", + "video_file": "Charades/videos/CIBGY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is washing their hands'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.4>." + } + ] + }, + { + "video_id": "85OEC", + "question_id": "85OEC", + "video_file": "Charades/videos/85OEC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sitting on the sofa' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <28.0> to <35.5>." + } + ] + }, + { + "video_id": "I0HIX", + "question_id": "I0HIX", + "video_file": "Charades/videos/I0HIX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is working on a laptop on a desk'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "I0HIX", + "question_id": "I0HIX", + "video_file": "Charades/videos/I0HIX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person sits at their desk working on their laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "2G6A8", + "question_id": "2G6A8", + "video_file": "Charades/videos/2G6A8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person runs from the room though the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <23.9>." + } + ] + }, + { + "video_id": "2G6A8", + "question_id": "2G6A8", + "video_file": "Charades/videos/2G6A8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs out of the room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <23.9>." + } + ] + }, + { + "video_id": "2G6A8", + "question_id": "2G6A8", + "video_file": "Charades/videos/2G6A8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turning off the light behind them' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.0> to <20.9>." + } + ] + }, + { + "video_id": "2G6A8", + "question_id": "2G6A8", + "video_file": "Charades/videos/2G6A8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person suddenly turns off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.0> to <20.9>." + } + ] + }, + { + "video_id": "FFBK3", + "question_id": "FFBK3", + "video_file": "Charades/videos/FFBK3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they put the book in a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.3> to <29.1>." + } + ] + }, + { + "video_id": "SXI64", + "question_id": "SXI64", + "video_file": "Charades/videos/SXI64.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person undressing in front of their mirror glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.6> to <13.5>." + } + ] + }, + { + "video_id": "13YII", + "question_id": "13YII", + "video_file": "Charades/videos/13YII.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is putting things in a kitchen cupboard' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "JVISI", + "question_id": "JVISI", + "video_file": "Charades/videos/JVISI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes off his shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.7> to <35.6>." + } + ] + }, + { + "video_id": "JVISI", + "question_id": "JVISI", + "video_file": "Charades/videos/JVISI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person he takes a drink from the cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.8> to <39.0>." + } + ] + }, + { + "video_id": "B8BSY", + "question_id": "B8BSY", + "video_file": "Charades/videos/B8BSY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sneezing at the same time' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.7> to <18.4>." + } + ] + }, + { + "video_id": "ZAIHG", + "question_id": "ZAIHG", + "video_file": "Charades/videos/ZAIHG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts it on a table near a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.4> to <18.9>." + } + ] + }, + { + "video_id": "KU2T0", + "question_id": "KU2T0", + "video_file": "Charades/videos/KU2T0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person is sitting in a chair tossing a pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.2>." + } + ] + }, + { + "video_id": "KX7FC", + "question_id": "KX7FC", + "video_file": "Charades/videos/KX7FC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins closing the cabinet over the sink' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.0> to <38.0>." + } + ] + }, + { + "video_id": "KX7FC", + "question_id": "KX7FC", + "video_file": "Charades/videos/KX7FC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closed the cabinets beneath the sink'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.0> to <38.0>." + } + ] + }, + { + "video_id": "EFI64", + "question_id": "EFI64", + "video_file": "Charades/videos/EFI64.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person watching laptop kept something on the desk' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.0> to <26.9>." + } + ] + }, + { + "video_id": "EFI64", + "question_id": "EFI64", + "video_file": "Charades/videos/EFI64.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person watching something in the laptop drinking some coffee' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.0> to <26.9>." + } + ] + }, + { + "video_id": "P3WQV", + "question_id": "P3WQV", + "video_file": "Charades/videos/P3WQV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they sit on bed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.5> to <44.2>." + } + ] + }, + { + "video_id": "P3WQV", + "question_id": "P3WQV", + "video_file": "Charades/videos/P3WQV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sitting in front of a mirror in a bedroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.5> to <44.2>." + } + ] + }, + { + "video_id": "K7X5G", + "question_id": "K7X5G", + "video_file": "Charades/videos/K7X5G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person started to close the laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.4> to <32.2>." + } + ] + }, + { + "video_id": "K7X5G", + "question_id": "K7X5G", + "video_file": "Charades/videos/K7X5G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes their laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.4> to <32.2>." + } + ] + }, + { + "video_id": "YE991", + "question_id": "YE991", + "video_file": "Charades/videos/YE991.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking a glass of water at the stove'." + }, + { + "from": "gpt", + "value": "From <18.4> to <28.0>." + } + ] + }, + { + "video_id": "0FVYR", + "question_id": "0FVYR", + "video_file": "Charades/videos/0FVYR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.6> to <17.4>." + } + ] + }, + { + "video_id": "0FVYR", + "question_id": "0FVYR", + "video_file": "Charades/videos/0FVYR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinks from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.2> to <27.5>." + } + ] + }, + { + "video_id": "0FVYR", + "question_id": "0FVYR", + "video_file": "Charades/videos/0FVYR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opening a closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.2> to <16.0>." + } + ] + }, + { + "video_id": "MNGT1", + "question_id": "MNGT1", + "video_file": "Charades/videos/MNGT1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they close the laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.7> to <16.8>." + } + ] + }, + { + "video_id": "MNGT1", + "question_id": "MNGT1", + "video_file": "Charades/videos/MNGT1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put it in a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.0> to <35.0>." + } + ] + }, + { + "video_id": "MNGT1", + "question_id": "MNGT1", + "video_file": "Charades/videos/MNGT1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person closed the laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.7> to <16.8>." + } + ] + }, + { + "video_id": "MNGT1", + "question_id": "MNGT1", + "video_file": "Charades/videos/MNGT1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person sits on the stairs working on their laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <14.3>." + } + ] + }, + { + "video_id": "MNGT1", + "question_id": "MNGT1", + "video_file": "Charades/videos/MNGT1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the bag over their shoulder'." + }, + { + "from": "gpt", + "value": "From <28.0> to <35.0>." + } + ] + }, + { + "video_id": "PZ0MM", + "question_id": "PZ0MM", + "video_file": "Charades/videos/PZ0MM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they pour some more water into their cup'." + }, + { + "from": "gpt", + "value": "From <5.7> to <14.5>." + } + ] + }, + { + "video_id": "PZ0MM", + "question_id": "PZ0MM", + "video_file": "Charades/videos/PZ0MM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person drinks from the glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "PZ0MM", + "question_id": "PZ0MM", + "video_file": "Charades/videos/PZ0MM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walks through a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.9> to <19.5>." + } + ] + }, + { + "video_id": "PZ0MM", + "question_id": "PZ0MM", + "video_file": "Charades/videos/PZ0MM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they walk through the doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.9> to <19.5>." + } + ] + }, + { + "video_id": "3GXB7", + "question_id": "3GXB7", + "video_file": "Charades/videos/3GXB7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sitting down at the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.3> to <31.1>." + } + ] + }, + { + "video_id": "4GDTQ", + "question_id": "4GDTQ", + "video_file": "Charades/videos/4GDTQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <16.9>." + } + ] + }, + { + "video_id": "10INO", + "question_id": "10INO", + "video_file": "Charades/videos/10INO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person play on their phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <62.1> to <83.1>." + } + ] + }, + { + "video_id": "XNO6M", + "question_id": "XNO6M", + "video_file": "Charades/videos/XNO6M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person turns off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <21.7>." + } + ] + }, + { + "video_id": "XNO6M", + "question_id": "XNO6M", + "video_file": "Charades/videos/XNO6M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turn off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.8> to <21.7>." + } + ] + }, + { + "video_id": "39T0D", + "question_id": "39T0D", + "video_file": "Charades/videos/39T0D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person started sneezing because of all the dust'." + }, + { + "from": "gpt", + "value": "From <17.3> to <33.5>." + } + ] + }, + { + "video_id": "555RO", + "question_id": "555RO", + "video_file": "Charades/videos/555RO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the bathroom door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <21.2>." + } + ] + }, + { + "video_id": "555RO", + "question_id": "555RO", + "video_file": "Charades/videos/555RO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person working on fixing the bathroom door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.7> to <31.0>." + } + ] + }, + { + "video_id": "C6LDA", + "question_id": "C6LDA", + "video_file": "Charades/videos/C6LDA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sitting in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <32.0>." + } + ] + }, + { + "video_id": "C6LDA", + "question_id": "C6LDA", + "video_file": "Charades/videos/C6LDA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a man fixed his hair in a mirror'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.1>." + } + ] + }, + { + "video_id": "E1U37", + "question_id": "E1U37", + "video_file": "Charades/videos/E1U37.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throwing it on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.1> to <10.0>." + } + ] + }, + { + "video_id": "9SP9P", + "question_id": "9SP9P", + "video_file": "Charades/videos/9SP9P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinking from a coffee cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.7> to <13.0>." + } + ] + }, + { + "video_id": "9SP9P", + "question_id": "9SP9P", + "video_file": "Charades/videos/9SP9P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts drinking from a cup'." + }, + { + "from": "gpt", + "value": "From <5.7> to <13.0>." + } + ] + }, + { + "video_id": "6XGUE", + "question_id": "6XGUE", + "video_file": "Charades/videos/6XGUE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person eats a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.4> to <10.8>." + } + ] + }, + { + "video_id": "6XGUE", + "question_id": "6XGUE", + "video_file": "Charades/videos/6XGUE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.0> to <22.4>." + } + ] + }, + { + "video_id": "GYAR9", + "question_id": "GYAR9", + "video_file": "Charades/videos/GYAR9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person throws the blanket on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.3> to <30.5>." + } + ] + }, + { + "video_id": "GANC6", + "question_id": "GANC6", + "video_file": "Charades/videos/GANC6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person start sneezing from the dust' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.6> to <19.6>." + } + ] + }, + { + "video_id": "GANC6", + "question_id": "GANC6", + "video_file": "Charades/videos/GANC6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts it on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <6.5>." + } + ] + }, + { + "video_id": "P23HN", + "question_id": "P23HN", + "video_file": "Charades/videos/P23HN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open up a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.3> to <10.9>." + } + ] + }, + { + "video_id": "PBGYV", + "question_id": "PBGYV", + "video_file": "Charades/videos/PBGYV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "PBGYV", + "question_id": "PBGYV", + "video_file": "Charades/videos/PBGYV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens the door to their bathroom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "PBGYV", + "question_id": "PBGYV", + "video_file": "Charades/videos/PBGYV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a picture'." + }, + { + "from": "gpt", + "value": "From <11.8> to <25.8>." + } + ] + }, + { + "video_id": "330ZE", + "question_id": "330ZE", + "video_file": "Charades/videos/330ZE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they take some bagged food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "330ZE", + "question_id": "330ZE", + "video_file": "Charades/videos/330ZE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting groceries on the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.5> to <12.9>." + } + ] + }, + { + "video_id": "OLSMC", + "question_id": "OLSMC", + "video_file": "Charades/videos/OLSMC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sneezing into the mirror'." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.1>." + } + ] + }, + { + "video_id": "K5OQO", + "question_id": "K5OQO", + "video_file": "Charades/videos/K5OQO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person closes a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.2> to <11.7>." + } + ] + }, + { + "video_id": "K5OQO", + "question_id": "K5OQO", + "video_file": "Charades/videos/K5OQO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person puts a sandwich on the bathroom shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "K5OQO", + "question_id": "K5OQO", + "video_file": "Charades/videos/K5OQO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the doorknob to the bathroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.2> to <11.7>." + } + ] + }, + { + "video_id": "L8N31", + "question_id": "L8N31", + "video_file": "Charades/videos/L8N31.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turn off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.2> to <26.9>." + } + ] + }, + { + "video_id": "L8N31", + "question_id": "L8N31", + "video_file": "Charades/videos/L8N31.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.2> to <26.9>." + } + ] + }, + { + "video_id": "A0T30", + "question_id": "A0T30", + "video_file": "Charades/videos/A0T30.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take a small towel' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.6> to <15.2>." + } + ] + }, + { + "video_id": "A0T30", + "question_id": "A0T30", + "video_file": "Charades/videos/A0T30.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they take a towel from a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.6> to <15.2>." + } + ] + }, + { + "video_id": "QB8O7", + "question_id": "QB8O7", + "video_file": "Charades/videos/QB8O7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person smiling at a picture on a laptop'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "F3ULU", + "question_id": "F3ULU", + "video_file": "Charades/videos/F3ULU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person is putting it on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.4> to <21.9>." + } + ] + }, + { + "video_id": "TS75H", + "question_id": "TS75H", + "video_file": "Charades/videos/TS75H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person grabs a bite to eat' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.4> to <20.0>." + } + ] + }, + { + "video_id": "TS75H", + "question_id": "TS75H", + "video_file": "Charades/videos/TS75H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is standing in a garage sneezing'." + }, + { + "from": "gpt", + "value": "From <0.9> to <6.1>." + } + ] + }, + { + "video_id": "TS75H", + "question_id": "TS75H", + "video_file": "Charades/videos/TS75H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they eat a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.4> to <20.0>." + } + ] + }, + { + "video_id": "TS75H", + "question_id": "TS75H", + "video_file": "Charades/videos/TS75H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person holding medicine sneezes twice' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.6>." + } + ] + }, + { + "video_id": "TS75H", + "question_id": "TS75H", + "video_file": "Charades/videos/TS75H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holding a bottle of medicine sneezing they sit down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.6>." + } + ] + }, + { + "video_id": "MEYZ5", + "question_id": "MEYZ5", + "video_file": "Charades/videos/MEYZ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they stand up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <19.4>." + } + ] + }, + { + "video_id": "MEYZ5", + "question_id": "MEYZ5", + "video_file": "Charades/videos/MEYZ5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open a window on a cabinet behind their head' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.4> to <25.4>." + } + ] + }, + { + "video_id": "NU09Z", + "question_id": "NU09Z", + "video_file": "Charades/videos/NU09Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is tidying up a shelf in the kitchen' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.3> to <17.8>." + } + ] + }, + { + "video_id": "AXIW1", + "question_id": "AXIW1", + "video_file": "Charades/videos/AXIW1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take a drink from a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <32.9> to <38.0>." + } + ] + }, + { + "video_id": "AXIW1", + "question_id": "AXIW1", + "video_file": "Charades/videos/AXIW1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes off the person's shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.7> to <34.6>." + } + ] + }, + { + "video_id": "GLH5S", + "question_id": "GLH5S", + "video_file": "Charades/videos/GLH5S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person sits in a chair'." + }, + { + "from": "gpt", + "value": "From <0.4> to <5.5>." + } + ] + }, + { + "video_id": "GLH5S", + "question_id": "GLH5S", + "video_file": "Charades/videos/GLH5S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they stand up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <9.8>." + } + ] + }, + { + "video_id": "GLH5S", + "question_id": "GLH5S", + "video_file": "Charades/videos/GLH5S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put the thing back on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <14.6>." + } + ] + }, + { + "video_id": "RBCU4", + "question_id": "RBCU4", + "video_file": "Charades/videos/RBCU4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.8> to <24.4>." + } + ] + }, + { + "video_id": "K0C0A", + "question_id": "K0C0A", + "video_file": "Charades/videos/K0C0A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person sits in a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.5>." + } + ] + }, + { + "video_id": "K0C0A", + "question_id": "K0C0A", + "video_file": "Charades/videos/K0C0A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begin to put groceries away'." + }, + { + "from": "gpt", + "value": "From <23.6> to <38.0>." + } + ] + }, + { + "video_id": "K0C0A", + "question_id": "K0C0A", + "video_file": "Charades/videos/K0C0A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opens a bag of groceries' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.0> to <21.3>." + } + ] + }, + { + "video_id": "L58KN", + "question_id": "L58KN", + "video_file": "Charades/videos/L58KN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put the towel down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.2> to <13.6>." + } + ] + }, + { + "video_id": "L58KN", + "question_id": "L58KN", + "video_file": "Charades/videos/L58KN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is in their bedroom holding a towel'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.4>." + } + ] + }, + { + "video_id": "9D37U", + "question_id": "9D37U", + "video_file": "Charades/videos/9D37U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begin playing a short puzzle game on their phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.5> to <32.0>." + } + ] + }, + { + "video_id": "4GWNV", + "question_id": "4GWNV", + "video_file": "Charades/videos/4GWNV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a camera'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <12.8>." + } + ] + }, + { + "video_id": "4GWNV", + "question_id": "4GWNV", + "video_file": "Charades/videos/4GWNV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person closes a door'." + }, + { + "from": "gpt", + "value": "From <10.6> to <17.8>." + } + ] + }, + { + "video_id": "4GWNV", + "question_id": "4GWNV", + "video_file": "Charades/videos/4GWNV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pulling the door closed behind them'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.6> to <17.8>." + } + ] + }, + { + "video_id": "4GWNV", + "question_id": "4GWNV", + "video_file": "Charades/videos/4GWNV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes a medicine cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.6> to <10.0>." + } + ] + }, + { + "video_id": "0RI7C", + "question_id": "0RI7C", + "video_file": "Charades/videos/0RI7C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person hey close the book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.5> to <7.5>." + } + ] + }, + { + "video_id": "0RI7C", + "question_id": "0RI7C", + "video_file": "Charades/videos/0RI7C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting in the bathroom reading a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "CMD0M", + "question_id": "CMD0M", + "video_file": "Charades/videos/CMD0M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pours themselves a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <30.7>." + } + ] + }, + { + "video_id": "CMD0M", + "question_id": "CMD0M", + "video_file": "Charades/videos/CMD0M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person pours some liquid into a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <30.7>." + } + ] + }, + { + "video_id": "YM5CV", + "question_id": "YM5CV", + "video_file": "Charades/videos/YM5CV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put the camera on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.9> to <24.8>." + } + ] + }, + { + "video_id": "U8X15", + "question_id": "U8X15", + "video_file": "Charades/videos/U8X15.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person stands in a room holding a camera' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.2> to <15.7>." + } + ] + }, + { + "video_id": "U8X15", + "question_id": "U8X15", + "video_file": "Charades/videos/U8X15.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person prepare to take a picture'." + }, + { + "from": "gpt", + "value": "From <10.3> to <20.2>." + } + ] + }, + { + "video_id": "KPHP5", + "question_id": "KPHP5", + "video_file": "Charades/videos/KPHP5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turning on a light'." + }, + { + "from": "gpt", + "value": "From <2.3> to <10.4>." + } + ] + }, + { + "video_id": "KPHP5", + "question_id": "KPHP5", + "video_file": "Charades/videos/KPHP5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begins sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.9> to <32.0>." + } + ] + }, + { + "video_id": "Z6HSZ", + "question_id": "Z6HSZ", + "video_file": "Charades/videos/Z6HSZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they close a nearby open door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.4> to <26.5>." + } + ] + }, + { + "video_id": "Z6HSZ", + "question_id": "Z6HSZ", + "video_file": "Charades/videos/Z6HSZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich who walks to the bathroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <21.3>." + } + ] + }, + { + "video_id": "Z6HSZ", + "question_id": "Z6HSZ", + "video_file": "Charades/videos/Z6HSZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <21.3>." + } + ] + }, + { + "video_id": "B7T3B", + "question_id": "B7T3B", + "video_file": "Charades/videos/B7T3B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person takes a picture from a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <22.0>." + } + ] + }, + { + "video_id": "B7T3B", + "question_id": "B7T3B", + "video_file": "Charades/videos/B7T3B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take a photograph from a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.2> to <22.0>." + } + ] + }, + { + "video_id": "B7T3B", + "question_id": "B7T3B", + "video_file": "Charades/videos/B7T3B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looked out a window'." + }, + { + "from": "gpt", + "value": "From <7.9> to <14.5>." + } + ] + }, + { + "video_id": "B7T3B", + "question_id": "B7T3B", + "video_file": "Charades/videos/B7T3B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks through a doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.4> to <28.4>." + } + ] + }, + { + "video_id": "I5Q2Z", + "question_id": "I5Q2Z", + "video_file": "Charades/videos/I5Q2Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a woman is holding two cups of tea'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.5> to <32.0>." + } + ] + }, + { + "video_id": "I5Q2Z", + "question_id": "I5Q2Z", + "video_file": "Charades/videos/I5Q2Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is pouring coffee between two cups' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "I5Q2Z", + "question_id": "I5Q2Z", + "video_file": "Charades/videos/I5Q2Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person prepares to eat his food'." + }, + { + "from": "gpt", + "value": "From <12.6> to <16.8>." + } + ] + }, + { + "video_id": "9L71W", + "question_id": "9L71W", + "video_file": "Charades/videos/9L71W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person looks at the picture for a moment' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <30.2> to <33.0>." + } + ] + }, + { + "video_id": "CMI7U", + "question_id": "CMI7U", + "video_file": "Charades/videos/CMI7U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting on a chair holding books' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.2>." + } + ] + }, + { + "video_id": "CMI7U", + "question_id": "CMI7U", + "video_file": "Charades/videos/CMI7U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the books on a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <21.3>." + } + ] + }, + { + "video_id": "CMI7U", + "question_id": "CMI7U", + "video_file": "Charades/videos/CMI7U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person the putting books on a shelf in the hallway'." + }, + { + "from": "gpt", + "value": "From <14.6> to <21.3>." + } + ] + }, + { + "video_id": "RCA3E", + "question_id": "RCA3E", + "video_file": "Charades/videos/RCA3E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they go to wash their hands in a sink'." + }, + { + "from": "gpt", + "value": "From <23.4> to <31.0>." + } + ] + }, + { + "video_id": "RCA3E", + "question_id": "RCA3E", + "video_file": "Charades/videos/RCA3E.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they are washing their hands in the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.4> to <31.0>." + } + ] + }, + { + "video_id": "JW2HC", + "question_id": "JW2HC", + "video_file": "Charades/videos/JW2HC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs into a bedroom holding a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "JW2HC", + "question_id": "JW2HC", + "video_file": "Charades/videos/JW2HC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs into a room holding a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "JW2HC", + "question_id": "JW2HC", + "video_file": "Charades/videos/JW2HC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they eat the sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <6.0>." + } + ] + }, + { + "video_id": "JW2HC", + "question_id": "JW2HC", + "video_file": "Charades/videos/JW2HC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person entered the room by eating sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.0> to <6.0>." + } + ] + }, + { + "video_id": "JW2HC", + "question_id": "JW2HC", + "video_file": "Charades/videos/JW2HC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.0> to <6.0>." + } + ] + }, + { + "video_id": "GF6TR", + "question_id": "GF6TR", + "video_file": "Charades/videos/GF6TR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts something else on the shelf'." + }, + { + "from": "gpt", + "value": "From <4.1> to <9.4>." + } + ] + }, + { + "video_id": "GF6TR", + "question_id": "GF6TR", + "video_file": "Charades/videos/GF6TR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person runs into the closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "GF6TR", + "question_id": "GF6TR", + "video_file": "Charades/videos/GF6TR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws the pillow onto a shelf'." + }, + { + "from": "gpt", + "value": "From <4.2> to <9.2>." + } + ] + }, + { + "video_id": "GF6TR", + "question_id": "GF6TR", + "video_file": "Charades/videos/GF6TR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "6VF2L", + "question_id": "6VF2L", + "video_file": "Charades/videos/6VF2L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks through a doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.7> to <15.8>." + } + ] + }, + { + "video_id": "UU2LH", + "question_id": "UU2LH", + "video_file": "Charades/videos/UU2LH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sit on the bed to put her shoes on'." + }, + { + "from": "gpt", + "value": "From <8.6> to <18.2>." + } + ] + }, + { + "video_id": "LTYK5", + "question_id": "LTYK5", + "video_file": "Charades/videos/LTYK5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is smiling in the laundry room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.3>." + } + ] + }, + { + "video_id": "LTYK5", + "question_id": "LTYK5", + "video_file": "Charades/videos/LTYK5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the dryer door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <17.9>." + } + ] + }, + { + "video_id": "LTYK5", + "question_id": "LTYK5", + "video_file": "Charades/videos/LTYK5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting clothes in the washer' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <16.2>." + } + ] + }, + { + "video_id": "LTYK5", + "question_id": "LTYK5", + "video_file": "Charades/videos/LTYK5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting the clothes inside of the dryer' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <16.2>." + } + ] + }, + { + "video_id": "DJ7DG", + "question_id": "DJ7DG", + "video_file": "Charades/videos/DJ7DG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they stand up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.7> to <18.0>." + } + ] + }, + { + "video_id": "DJ7DG", + "question_id": "DJ7DG", + "video_file": "Charades/videos/DJ7DG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sits on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.6>." + } + ] + }, + { + "video_id": "DJ7DG", + "question_id": "DJ7DG", + "video_file": "Charades/videos/DJ7DG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sits on the floor in a doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.6>." + } + ] + }, + { + "video_id": "KHWG2", + "question_id": "KHWG2", + "video_file": "Charades/videos/KHWG2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinking from a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.1> to <9.3>." + } + ] + }, + { + "video_id": "XIUQJ", + "question_id": "XIUQJ", + "video_file": "Charades/videos/XIUQJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.0>." + } + ] + }, + { + "video_id": "XIUQJ", + "question_id": "XIUQJ", + "video_file": "Charades/videos/XIUQJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they take their shoes off' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <26.6>." + } + ] + }, + { + "video_id": "XIUQJ", + "question_id": "XIUQJ", + "video_file": "Charades/videos/XIUQJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they take off their shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <26.6>." + } + ] + }, + { + "video_id": "QFVMA", + "question_id": "QFVMA", + "video_file": "Charades/videos/QFVMA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looks out a window'." + }, + { + "from": "gpt", + "value": "From <23.3> to <34.0>." + } + ] + }, + { + "video_id": "QFVMA", + "question_id": "QFVMA", + "video_file": "Charades/videos/QFVMA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they look out the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.3> to <34.0>." + } + ] + }, + { + "video_id": "HRR0X", + "question_id": "HRR0X", + "video_file": "Charades/videos/HRR0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.3>." + } + ] + }, + { + "video_id": "HRR0X", + "question_id": "HRR0X", + "video_file": "Charades/videos/HRR0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take a bite of some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.0> to <25.1>." + } + ] + }, + { + "video_id": "HRR0X", + "question_id": "HRR0X", + "video_file": "Charades/videos/HRR0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person hits a light switch' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.1> to <20.8>." + } + ] + }, + { + "video_id": "HRR0X", + "question_id": "HRR0X", + "video_file": "Charades/videos/HRR0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eat an apple' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.5> to <27.0>." + } + ] + }, + { + "video_id": "HRR0X", + "question_id": "HRR0X", + "video_file": "Charades/videos/HRR0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they turn on a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <20.8>." + } + ] + }, + { + "video_id": "HRR0X", + "question_id": "HRR0X", + "video_file": "Charades/videos/HRR0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.4> to <16.5>." + } + ] + }, + { + "video_id": "HRR0X", + "question_id": "HRR0X", + "video_file": "Charades/videos/HRR0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens in their dining room holding a book'." + }, + { + "from": "gpt", + "value": "From <9.4> to <16.5>." + } + ] + }, + { + "video_id": "OWLRQ", + "question_id": "OWLRQ", + "video_file": "Charades/videos/OWLRQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they start sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.8> to <31.3>." + } + ] + }, + { + "video_id": "OWLRQ", + "question_id": "OWLRQ", + "video_file": "Charades/videos/OWLRQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take a picture with the camera' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.6> to <23.0>." + } + ] + }, + { + "video_id": "ZP5TG", + "question_id": "ZP5TG", + "video_file": "Charades/videos/ZP5TG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person is laughing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.3> to <10.7>." + } + ] + }, + { + "video_id": "ZP5TG", + "question_id": "ZP5TG", + "video_file": "Charades/videos/ZP5TG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person is eating'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "ZP5TG", + "question_id": "ZP5TG", + "video_file": "Charades/videos/ZP5TG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person 1 of them is eating food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "ZP5TG", + "question_id": "ZP5TG", + "video_file": "Charades/videos/ZP5TG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is eating food on the bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.8>." + } + ] + }, + { + "video_id": "9IXS4", + "question_id": "9IXS4", + "video_file": "Charades/videos/9IXS4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns off a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <20.4>." + } + ] + }, + { + "video_id": "9IXS4", + "question_id": "9IXS4", + "video_file": "Charades/videos/9IXS4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they turn off the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.6> to <20.4>." + } + ] + }, + { + "video_id": "X4WWH", + "question_id": "X4WWH", + "video_file": "Charades/videos/X4WWH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.2> to <15.8>." + } + ] + }, + { + "video_id": "X4WWH", + "question_id": "X4WWH", + "video_file": "Charades/videos/X4WWH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person opens the closet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.8> to <13.2>." + } + ] + }, + { + "video_id": "9KYMU", + "question_id": "9KYMU", + "video_file": "Charades/videos/9KYMU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opening a box with shoes in it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.9> to <23.9>." + } + ] + }, + { + "video_id": "9KYMU", + "question_id": "9KYMU", + "video_file": "Charades/videos/9KYMU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs back in forth in the garage' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "9KYMU", + "question_id": "9KYMU", + "video_file": "Charades/videos/9KYMU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is running by a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "TCGAY", + "question_id": "TCGAY", + "video_file": "Charades/videos/TCGAY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put a glass onto a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.7> to <31.9>." + } + ] + }, + { + "video_id": "TCGAY", + "question_id": "TCGAY", + "video_file": "Charades/videos/TCGAY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person start tidying themselves with a towel' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.3> to <32.0>." + } + ] + }, + { + "video_id": "MY9UC", + "question_id": "MY9UC", + "video_file": "Charades/videos/MY9UC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is in the kitchen washing dishes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "MY9UC", + "question_id": "MY9UC", + "video_file": "Charades/videos/MY9UC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a different person is sitting on the kitchen table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.6> to <31.0>." + } + ] + }, + { + "video_id": "MY9UC", + "question_id": "MY9UC", + "video_file": "Charades/videos/MY9UC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is washing dishes using a sponge' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "MY9UC", + "question_id": "MY9UC", + "video_file": "Charades/videos/MY9UC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is standing in a kitchen washing dishes'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "ISTQI", + "question_id": "ISTQI", + "video_file": "Charades/videos/ISTQI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person takes a book off a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <11.5>." + } + ] + }, + { + "video_id": "63GDX", + "question_id": "63GDX", + "video_file": "Charades/videos/63GDX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person puts the box down'." + }, + { + "from": "gpt", + "value": "From <25.3> to <29.6>." + } + ] + }, + { + "video_id": "63GDX", + "question_id": "63GDX", + "video_file": "Charades/videos/63GDX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <18.3>." + } + ] + }, + { + "video_id": "F3O37", + "question_id": "F3O37", + "video_file": "Charades/videos/F3O37.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is throwing dirty shoes into the laundry' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.8> to <10.5>." + } + ] + }, + { + "video_id": "KQLIJ", + "question_id": "KQLIJ", + "video_file": "Charades/videos/KQLIJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws it on the floor'." + }, + { + "from": "gpt", + "value": "From <11.0> to <16.6>." + } + ] + }, + { + "video_id": "KQLIJ", + "question_id": "KQLIJ", + "video_file": "Charades/videos/KQLIJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is playing with a camera' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.6>." + } + ] + }, + { + "video_id": "KQLIJ", + "question_id": "KQLIJ", + "video_file": "Charades/videos/KQLIJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is holding a camera'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "Y9WZW", + "question_id": "Y9WZW", + "video_file": "Charades/videos/Y9WZW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person she drinks from a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.7> to <46.6>." + } + ] + }, + { + "video_id": "NU4XA", + "question_id": "NU4XA", + "video_file": "Charades/videos/NU4XA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sit down in another chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.9> to <22.9>." + } + ] + }, + { + "video_id": "NU4XA", + "question_id": "NU4XA", + "video_file": "Charades/videos/NU4XA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person watch television' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.6> to <32.0>." + } + ] + }, + { + "video_id": "NU4XA", + "question_id": "NU4XA", + "video_file": "Charades/videos/NU4XA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.8> to <13.5>." + } + ] + }, + { + "video_id": "NU4XA", + "question_id": "NU4XA", + "video_file": "Charades/videos/NU4XA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sitting in a chair who gets up' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.9> to <22.9>." + } + ] + }, + { + "video_id": "GLLNH", + "question_id": "GLLNH", + "video_file": "Charades/videos/GLLNH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a picture of something through the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.3> to <31.1>." + } + ] + }, + { + "video_id": "GLLNH", + "question_id": "GLLNH", + "video_file": "Charades/videos/GLLNH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes out a phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <20.8>." + } + ] + }, + { + "video_id": "GLLNH", + "question_id": "GLLNH", + "video_file": "Charades/videos/GLLNH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a picture in the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.3> to <31.1>." + } + ] + }, + { + "video_id": "GLLNH", + "question_id": "GLLNH", + "video_file": "Charades/videos/GLLNH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a picture'." + }, + { + "from": "gpt", + "value": "From <24.3> to <31.1>." + } + ] + }, + { + "video_id": "GLLNH", + "question_id": "GLLNH", + "video_file": "Charades/videos/GLLNH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes out a camera' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.3> to <20.8>." + } + ] + }, + { + "video_id": "GLLNH", + "question_id": "GLLNH", + "video_file": "Charades/videos/GLLNH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes out their phone'." + }, + { + "from": "gpt", + "value": "From <14.3> to <20.8>." + } + ] + }, + { + "video_id": "TW6NZ", + "question_id": "TW6NZ", + "video_file": "Charades/videos/TW6NZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting on a sofa'." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.7>." + } + ] + }, + { + "video_id": "TW6NZ", + "question_id": "TW6NZ", + "video_file": "Charades/videos/TW6NZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting on a couch' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.7>." + } + ] + }, + { + "video_id": "TW6NZ", + "question_id": "TW6NZ", + "video_file": "Charades/videos/TW6NZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they are smiling' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.4> to <27.4>." + } + ] + }, + { + "video_id": "TW6NZ", + "question_id": "TW6NZ", + "video_file": "Charades/videos/TW6NZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person sits down on the sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.7>." + } + ] + }, + { + "video_id": "TW6NZ", + "question_id": "TW6NZ", + "video_file": "Charades/videos/TW6NZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person begins undressing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <12.4>." + } + ] + }, + { + "video_id": "SQY6A", + "question_id": "SQY6A", + "video_file": "Charades/videos/SQY6A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person picks up a box closes itand puts it back' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <27.5>." + } + ] + }, + { + "video_id": "SQY6A", + "question_id": "SQY6A", + "video_file": "Charades/videos/SQY6A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person picks up shoes puts them back down'." + }, + { + "from": "gpt", + "value": "From <10.1> to <15.4>." + } + ] + }, + { + "video_id": "GNPSK", + "question_id": "GNPSK", + "video_file": "Charades/videos/GNPSK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sitting in a chair eating' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "GNPSK", + "question_id": "GNPSK", + "video_file": "Charades/videos/GNPSK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sitting at a desk eating an apple' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "GNPSK", + "question_id": "GNPSK", + "video_file": "Charades/videos/GNPSK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating at the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "B8W55", + "question_id": "B8W55", + "video_file": "Charades/videos/B8W55.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person the open a cabinet'." + }, + { + "from": "gpt", + "value": "From <18.8> to <27.7>." + } + ] + }, + { + "video_id": "SL9P2", + "question_id": "SL9P2", + "video_file": "Charades/videos/SL9P2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking some medicine from a rather large cup'." + }, + { + "from": "gpt", + "value": "From <26.0> to <32.0>." + } + ] + }, + { + "video_id": "SL9P2", + "question_id": "SL9P2", + "video_file": "Charades/videos/SL9P2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is playing on their laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.9>." + } + ] + }, + { + "video_id": "SL9P2", + "question_id": "SL9P2", + "video_file": "Charades/videos/SL9P2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person working on laptop sitting down taking medicine'." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.9>." + } + ] + }, + { + "video_id": "J8781", + "question_id": "J8781", + "video_file": "Charades/videos/J8781.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.5> to <14.1>." + } + ] + }, + { + "video_id": "J8781", + "question_id": "J8781", + "video_file": "Charades/videos/J8781.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting on their shoes'." + }, + { + "from": "gpt", + "value": "From <3.4> to <10.1>." + } + ] + }, + { + "video_id": "J8781", + "question_id": "J8781", + "video_file": "Charades/videos/J8781.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sitting on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.5> to <14.1>." + } + ] + }, + { + "video_id": "J8781", + "question_id": "J8781", + "video_file": "Charades/videos/J8781.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <21.0>." + } + ] + }, + { + "video_id": "J8781", + "question_id": "J8781", + "video_file": "Charades/videos/J8781.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is getting dressed in the bedroom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "J8781", + "question_id": "J8781", + "video_file": "Charades/videos/J8781.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the door behind them'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <21.0>." + } + ] + }, + { + "video_id": "J8781", + "question_id": "J8781", + "video_file": "Charades/videos/J8781.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sitting in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.5> to <14.1>." + } + ] + }, + { + "video_id": "A0ZW3", + "question_id": "A0ZW3", + "video_file": "Charades/videos/A0ZW3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is awakening on a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <15.8>." + } + ] + }, + { + "video_id": "D4LRR", + "question_id": "D4LRR", + "video_file": "Charades/videos/D4LRR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they open a laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.6> to <19.1>." + } + ] + }, + { + "video_id": "D4LRR", + "question_id": "D4LRR", + "video_file": "Charades/videos/D4LRR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a picture of themselves'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.1>." + } + ] + }, + { + "video_id": "X4CZH", + "question_id": "X4CZH", + "video_file": "Charades/videos/X4CZH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person off a light switch'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <10.3>." + } + ] + }, + { + "video_id": "X4CZH", + "question_id": "X4CZH", + "video_file": "Charades/videos/X4CZH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turned on the light'." + }, + { + "from": "gpt", + "value": "From <3.6> to <10.3>." + } + ] + }, + { + "video_id": "X4CZH", + "question_id": "X4CZH", + "video_file": "Charades/videos/X4CZH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is looking out a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "X4CZH", + "question_id": "X4CZH", + "video_file": "Charades/videos/X4CZH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'opening it the person fixed the light bulb!' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.6> to <10.3>." + } + ] + }, + { + "video_id": "X4CZH", + "question_id": "X4CZH", + "video_file": "Charades/videos/X4CZH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take the light bulb from over head'." + }, + { + "from": "gpt", + "value": "From <3.6> to <10.3>." + } + ] + }, + { + "video_id": "LT13Z", + "question_id": "LT13Z", + "video_file": "Charades/videos/LT13Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person fixes his hair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.3> to <14.0>." + } + ] + }, + { + "video_id": "LT13Z", + "question_id": "LT13Z", + "video_file": "Charades/videos/LT13Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing by the mirror' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.1> to <21.2>." + } + ] + }, + { + "video_id": "8MY8P", + "question_id": "8MY8P", + "video_file": "Charades/videos/8MY8P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person begins cooking it in a microwave'." + }, + { + "from": "gpt", + "value": "From <7.8> to <17.2>." + } + ] + }, + { + "video_id": "8MY8P", + "question_id": "8MY8P", + "video_file": "Charades/videos/8MY8P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person takes some food off the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <7.1>." + } + ] + }, + { + "video_id": "GIZCL", + "question_id": "GIZCL", + "video_file": "Charades/videos/GIZCL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person holds a box of groceries'." + }, + { + "from": "gpt", + "value": "From <3.9> to <16.6>." + } + ] + }, + { + "video_id": "0WU7S", + "question_id": "0WU7S", + "video_file": "Charades/videos/0WU7S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person takes a bag from a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.2> to <11.5>." + } + ] + }, + { + "video_id": "0WU7S", + "question_id": "0WU7S", + "video_file": "Charades/videos/0WU7S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they open the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.1> to <8.2>." + } + ] + }, + { + "video_id": "0WU7S", + "question_id": "0WU7S", + "video_file": "Charades/videos/0WU7S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they start eating the chips' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <34.0>." + } + ] + }, + { + "video_id": "D740L", + "question_id": "D740L", + "video_file": "Charades/videos/D740L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <21.4>." + } + ] + }, + { + "video_id": "D740L", + "question_id": "D740L", + "video_file": "Charades/videos/D740L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes clothes out'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <10.1>." + } + ] + }, + { + "video_id": "D740L", + "question_id": "D740L", + "video_file": "Charades/videos/D740L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws empty bag on floor'." + }, + { + "from": "gpt", + "value": "From <26.4> to <31.5>." + } + ] + }, + { + "video_id": "LP1F3", + "question_id": "LP1F3", + "video_file": "Charades/videos/LP1F3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "LP1F3", + "question_id": "LP1F3", + "video_file": "Charades/videos/LP1F3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person sitting at their desk begins sneezing uncontrollably'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.4>." + } + ] + }, + { + "video_id": "QL92H", + "question_id": "QL92H", + "video_file": "Charades/videos/QL92H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is running in holding a bag' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <8.1>." + } + ] + }, + { + "video_id": "QL92H", + "question_id": "QL92H", + "video_file": "Charades/videos/QL92H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sitting on a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.9> to <15.3>." + } + ] + }, + { + "video_id": "JXNEN", + "question_id": "JXNEN", + "video_file": "Charades/videos/JXNEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they sit down on a sofa'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.5> to <27.0>." + } + ] + }, + { + "video_id": "JXNEN", + "question_id": "JXNEN", + "video_file": "Charades/videos/JXNEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <17.7>." + } + ] + }, + { + "video_id": "JXNEN", + "question_id": "JXNEN", + "video_file": "Charades/videos/JXNEN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person laughs at them'." + }, + { + "from": "gpt", + "value": "From <15.5> to <27.0>." + } + ] + }, + { + "video_id": "HCM5Y", + "question_id": "HCM5Y", + "video_file": "Charades/videos/HCM5Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person seated in front of a mirror' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "HCM5Y", + "question_id": "HCM5Y", + "video_file": "Charades/videos/HCM5Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting on the floor eating some food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "HCM5Y", + "question_id": "HCM5Y", + "video_file": "Charades/videos/HCM5Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person eats' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "HCM5Y", + "question_id": "HCM5Y", + "video_file": "Charades/videos/HCM5Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looks like eating something' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "STAFD", + "question_id": "STAFD", + "video_file": "Charades/videos/STAFD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.3> to <12.0>." + } + ] + }, + { + "video_id": "N50C6", + "question_id": "N50C6", + "video_file": "Charades/videos/N50C6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is holding a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.2> to <9.6>." + } + ] + }, + { + "video_id": "7K2CS", + "question_id": "7K2CS", + "video_file": "Charades/videos/7K2CS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts eating' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <31.0>." + } + ] + }, + { + "video_id": "7K2CS", + "question_id": "7K2CS", + "video_file": "Charades/videos/7K2CS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins eating it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <31.0>." + } + ] + }, + { + "video_id": "7K2CS", + "question_id": "7K2CS", + "video_file": "Charades/videos/7K2CS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats from s container' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.9> to <31.0>." + } + ] + }, + { + "video_id": "7K2CS", + "question_id": "7K2CS", + "video_file": "Charades/videos/7K2CS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins undressing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.8> to <19.2>." + } + ] + }, + { + "video_id": "2LMQF", + "question_id": "2LMQF", + "video_file": "Charades/videos/2LMQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they pour water into a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.8> to <22.9>." + } + ] + }, + { + "video_id": "2LMQF", + "question_id": "2LMQF", + "video_file": "Charades/videos/2LMQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "2LMQF", + "question_id": "2LMQF", + "video_file": "Charades/videos/2LMQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting at the table eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "2LMQF", + "question_id": "2LMQF", + "video_file": "Charades/videos/2LMQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "2LMQF", + "question_id": "2LMQF", + "video_file": "Charades/videos/2LMQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they pour water into another glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.8> to <22.9>." + } + ] + }, + { + "video_id": "D8YOL", + "question_id": "D8YOL", + "video_file": "Charades/videos/D8YOL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is seen working on their laptop'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.8>." + } + ] + }, + { + "video_id": "D8YOL", + "question_id": "D8YOL", + "video_file": "Charades/videos/D8YOL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits on a chair working on their laptop'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.8>." + } + ] + }, + { + "video_id": "REOES", + "question_id": "REOES", + "video_file": "Charades/videos/REOES.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the book down' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.6> to <17.8>." + } + ] + }, + { + "video_id": "REOES", + "question_id": "REOES", + "video_file": "Charades/videos/REOES.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they take out a camera'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.4> to <21.3>." + } + ] + }, + { + "video_id": "REOES", + "question_id": "REOES", + "video_file": "Charades/videos/REOES.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a picture of themselves' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.6> to <26.0>." + } + ] + }, + { + "video_id": "TFQNW", + "question_id": "TFQNW", + "video_file": "Charades/videos/TFQNW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person goes to the window to run in place' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.7> to <21.0>." + } + ] + }, + { + "video_id": "AUYOZ", + "question_id": "AUYOZ", + "video_file": "Charades/videos/AUYOZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a phone from their pocket'." + }, + { + "from": "gpt", + "value": "From <15.6> to <20.3>." + } + ] + }, + { + "video_id": "KSDFR", + "question_id": "KSDFR", + "video_file": "Charades/videos/KSDFR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'this person pours something in a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.5> to <26.9>." + } + ] + }, + { + "video_id": "OK2TU", + "question_id": "OK2TU", + "video_file": "Charades/videos/OK2TU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <37.7> to <46.0>." + } + ] + }, + { + "video_id": "OK2TU", + "question_id": "OK2TU", + "video_file": "Charades/videos/OK2TU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.5> to <42.0>." + } + ] + }, + { + "video_id": "CC87Z", + "question_id": "CC87Z", + "video_file": "Charades/videos/CC87Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closed the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.5> to <18.1>." + } + ] + }, + { + "video_id": "CC87Z", + "question_id": "CC87Z", + "video_file": "Charades/videos/CC87Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opened the door to the room' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.2> to <12.6>." + } + ] + }, + { + "video_id": "AVJFE", + "question_id": "AVJFE", + "video_file": "Charades/videos/AVJFE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person is sitting at table eating'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.8> to <19.8>." + } + ] + }, + { + "video_id": "ZVCMM", + "question_id": "ZVCMM", + "video_file": "Charades/videos/ZVCMM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is standing in the kitchen eating something' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <10.9>." + } + ] + }, + { + "video_id": "ZVCMM", + "question_id": "ZVCMM", + "video_file": "Charades/videos/ZVCMM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <20.9>." + } + ] + }, + { + "video_id": "ZVCMM", + "question_id": "ZVCMM", + "video_file": "Charades/videos/ZVCMM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.7> to <10.9>." + } + ] + }, + { + "video_id": "ZVCMM", + "question_id": "ZVCMM", + "video_file": "Charades/videos/ZVCMM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting dishes in a sink in a kitchen' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.6> to <27.0>." + } + ] + }, + { + "video_id": "ZVCMM", + "question_id": "ZVCMM", + "video_file": "Charades/videos/ZVCMM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.6> to <14.3>." + } + ] + }, + { + "video_id": "J84TZ", + "question_id": "J84TZ", + "video_file": "Charades/videos/J84TZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is awakening on a couch'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.5>." + } + ] + }, + { + "video_id": "J84TZ", + "question_id": "J84TZ", + "video_file": "Charades/videos/J84TZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walk to the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.2> to <21.0>." + } + ] + }, + { + "video_id": "J84TZ", + "question_id": "J84TZ", + "video_file": "Charades/videos/J84TZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walks through a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <21.0>." + } + ] + }, + { + "video_id": "243Y9", + "question_id": "243Y9", + "video_file": "Charades/videos/243Y9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats something out of a plastic bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.5> to <28.5>." + } + ] + }, + { + "video_id": "243Y9", + "question_id": "243Y9", + "video_file": "Charades/videos/243Y9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits down in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <27.5>." + } + ] + }, + { + "video_id": "243Y9", + "question_id": "243Y9", + "video_file": "Charades/videos/243Y9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts eating' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.5> to <28.5>." + } + ] + }, + { + "video_id": "243Y9", + "question_id": "243Y9", + "video_file": "Charades/videos/243Y9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person sits on a chair'." + }, + { + "from": "gpt", + "value": "From <17.7> to <27.5>." + } + ] + }, + { + "video_id": "243Y9", + "question_id": "243Y9", + "video_file": "Charades/videos/243Y9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person is sitting in a chair eating breakfast' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.7> to <27.5>." + } + ] + }, + { + "video_id": "PWO7A", + "question_id": "PWO7A", + "video_file": "Charades/videos/PWO7A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door behind them'." + }, + { + "from": "gpt", + "value": "From <11.9> to <19.9>." + } + ] + }, + { + "video_id": "9GOUI", + "question_id": "9GOUI", + "video_file": "Charades/videos/9GOUI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person running up the steps' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.0> to <19.9>." + } + ] + }, + { + "video_id": "9GOUI", + "question_id": "9GOUI", + "video_file": "Charades/videos/9GOUI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person smiles as they eat a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <6.2>." + } + ] + }, + { + "video_id": "9GOUI", + "question_id": "9GOUI", + "video_file": "Charades/videos/9GOUI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person run up the stairs'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <19.9>." + } + ] + }, + { + "video_id": "9GOUI", + "question_id": "9GOUI", + "video_file": "Charades/videos/9GOUI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drink a cup of coffee on the stairs' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <9.5>." + } + ] + }, + { + "video_id": "9GOUI", + "question_id": "9GOUI", + "video_file": "Charades/videos/9GOUI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person runs up the stairs' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <19.9>." + } + ] + }, + { + "video_id": "9GOUI", + "question_id": "9GOUI", + "video_file": "Charades/videos/9GOUI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting on the stairs eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <6.2>." + } + ] + }, + { + "video_id": "YA7M0", + "question_id": "YA7M0", + "video_file": "Charades/videos/YA7M0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a sip of coffee from a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.1> to <38.7>." + } + ] + }, + { + "video_id": "YA7M0", + "question_id": "YA7M0", + "video_file": "Charades/videos/YA7M0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a sip of water from a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.1> to <38.7>." + } + ] + }, + { + "video_id": "GQR08", + "question_id": "GQR08", + "video_file": "Charades/videos/GQR08.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person awakens holding a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.6>." + } + ] + }, + { + "video_id": "JZB97", + "question_id": "JZB97", + "video_file": "Charades/videos/JZB97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put the towel down'." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.3>." + } + ] + }, + { + "video_id": "JZB97", + "question_id": "JZB97", + "video_file": "Charades/videos/JZB97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes some clothes'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "JZB97", + "question_id": "JZB97", + "video_file": "Charades/videos/JZB97.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person laughs at a video that's playing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.1> to <31.9>." + } + ] + }, + { + "video_id": "XUTJZ", + "question_id": "XUTJZ", + "video_file": "Charades/videos/XUTJZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looking out a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <24.0>." + } + ] + }, + { + "video_id": "XUTJZ", + "question_id": "XUTJZ", + "video_file": "Charades/videos/XUTJZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person looking into a room out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <24.0>." + } + ] + }, + { + "video_id": "GXDGV", + "question_id": "GXDGV", + "video_file": "Charades/videos/GXDGV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts a camera around their neck' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <8.6>." + } + ] + }, + { + "video_id": "VNQES", + "question_id": "VNQES", + "video_file": "Charades/videos/VNQES.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put the box down on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.9> to <31.0>." + } + ] + }, + { + "video_id": "VNQES", + "question_id": "VNQES", + "video_file": "Charades/videos/VNQES.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put the box on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.9> to <31.0>." + } + ] + }, + { + "video_id": "VNQES", + "question_id": "VNQES", + "video_file": "Charades/videos/VNQES.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is laughing at something in a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.0> to <13.4>." + } + ] + }, + { + "video_id": "M7KH7", + "question_id": "M7KH7", + "video_file": "Charades/videos/M7KH7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting it back on the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.9> to <28.7>." + } + ] + }, + { + "video_id": "88LYX", + "question_id": "88LYX", + "video_file": "Charades/videos/88LYX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is cooking on their stove' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "88LYX", + "question_id": "88LYX", + "video_file": "Charades/videos/88LYX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person go back to cooking'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "88LYX", + "question_id": "88LYX", + "video_file": "Charades/videos/88LYX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is cooking on the stove'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "88LYX", + "question_id": "88LYX", + "video_file": "Charades/videos/88LYX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person wash a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.5> to <16.4>." + } + ] + }, + { + "video_id": "88LYX", + "question_id": "88LYX", + "video_file": "Charades/videos/88LYX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'there is a person cooking on a stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.7>." + } + ] + }, + { + "video_id": "O7KDK", + "question_id": "O7KDK", + "video_file": "Charades/videos/O7KDK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is running into the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "O7KDK", + "question_id": "O7KDK", + "video_file": "Charades/videos/O7KDK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs in to a bedroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "ZZ9RN", + "question_id": "ZZ9RN", + "video_file": "Charades/videos/ZZ9RN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.9> to <19.6>." + } + ] + }, + { + "video_id": "WBX8F", + "question_id": "WBX8F", + "video_file": "Charades/videos/WBX8F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they stand back up turn on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.2> to <30.0>." + } + ] + }, + { + "video_id": "WBX8F", + "question_id": "WBX8F", + "video_file": "Charades/videos/WBX8F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person leaving the light on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.2> to <30.0>." + } + ] + }, + { + "video_id": "WBX8F", + "question_id": "WBX8F", + "video_file": "Charades/videos/WBX8F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person turns the light on'." + }, + { + "from": "gpt", + "value": "From <25.2> to <30.0>." + } + ] + }, + { + "video_id": "MTA02", + "question_id": "MTA02", + "video_file": "Charades/videos/MTA02.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person is closing the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.0> to <25.1>." + } + ] + }, + { + "video_id": "6QPNW", + "question_id": "6QPNW", + "video_file": "Charades/videos/6QPNW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is smiling at a pair of shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.5> to <21.4>." + } + ] + }, + { + "video_id": "6QPNW", + "question_id": "6QPNW", + "video_file": "Charades/videos/6QPNW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person stands near smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.5> to <21.4>." + } + ] + }, + { + "video_id": "CVX9F", + "question_id": "CVX9F", + "video_file": "Charades/videos/CVX9F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a picture of the stove with a camera'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "CVX9F", + "question_id": "CVX9F", + "video_file": "Charades/videos/CVX9F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they take a picture of the stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.7>." + } + ] + }, + { + "video_id": "8IBNY", + "question_id": "8IBNY", + "video_file": "Charades/videos/8IBNY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes some medicine' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.1> to <42.7>." + } + ] + }, + { + "video_id": "8IBNY", + "question_id": "8IBNY", + "video_file": "Charades/videos/8IBNY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person pours something into a glass'." + }, + { + "from": "gpt", + "value": "From <9.6> to <17.2>." + } + ] + }, + { + "video_id": "8IBNY", + "question_id": "8IBNY", + "video_file": "Charades/videos/8IBNY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person pours a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.6> to <17.2>." + } + ] + }, + { + "video_id": "YQ6Z6", + "question_id": "YQ6Z6", + "video_file": "Charades/videos/YQ6Z6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes a pillow from the dryer' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.5> to <16.3>." + } + ] + }, + { + "video_id": "YQ6Z6", + "question_id": "YQ6Z6", + "video_file": "Charades/videos/YQ6Z6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.8> to <22.9>." + } + ] + }, + { + "video_id": "YQ6Z6", + "question_id": "YQ6Z6", + "video_file": "Charades/videos/YQ6Z6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <10.3>." + } + ] + }, + { + "video_id": "YQ6Z6", + "question_id": "YQ6Z6", + "video_file": "Charades/videos/YQ6Z6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes pillow out of dryer' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.5> to <16.3>." + } + ] + }, + { + "video_id": "YQ6Z6", + "question_id": "YQ6Z6", + "video_file": "Charades/videos/YQ6Z6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a pillow out of a dryer' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <16.3>." + } + ] + }, + { + "video_id": "QJIAV", + "question_id": "QJIAV", + "video_file": "Charades/videos/QJIAV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is eating food at the kitchen table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "QJIAV", + "question_id": "QJIAV", + "video_file": "Charades/videos/QJIAV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is standing at a counter eating' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.3>." + } + ] + }, + { + "video_id": "K90LN", + "question_id": "K90LN", + "video_file": "Charades/videos/K90LN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opening rice from a bag into a pot' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.9> to <9.3>." + } + ] + }, + { + "video_id": "KK8N9", + "question_id": "KK8N9", + "video_file": "Charades/videos/KK8N9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a phone off a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.7> to <11.0>." + } + ] + }, + { + "video_id": "KK8N9", + "question_id": "KK8N9", + "video_file": "Charades/videos/KK8N9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person so they start laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.8> to <13.0>." + } + ] + }, + { + "video_id": "IZSPN", + "question_id": "IZSPN", + "video_file": "Charades/videos/IZSPN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.2> to <8.3>." + } + ] + }, + { + "video_id": "IZSPN", + "question_id": "IZSPN", + "video_file": "Charades/videos/IZSPN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is opening a cabinet in their home office' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.2> to <8.3>." + } + ] + }, + { + "video_id": "OBTP8", + "question_id": "OBTP8", + "video_file": "Charades/videos/OBTP8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <11.8>." + } + ] + }, + { + "video_id": "OBTP8", + "question_id": "OBTP8", + "video_file": "Charades/videos/OBTP8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <11.8>." + } + ] + }, + { + "video_id": "PV856", + "question_id": "PV856", + "video_file": "Charades/videos/PV856.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.7> to <33.3>." + } + ] + }, + { + "video_id": "PV856", + "question_id": "PV856", + "video_file": "Charades/videos/PV856.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a door in their garage'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.7> to <33.3>." + } + ] + }, + { + "video_id": "PV856", + "question_id": "PV856", + "video_file": "Charades/videos/PV856.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.6> to <7.3>." + } + ] + }, + { + "video_id": "RE1M6", + "question_id": "RE1M6", + "video_file": "Charades/videos/RE1M6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turn off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <26.1>." + } + ] + }, + { + "video_id": "RE1M6", + "question_id": "RE1M6", + "video_file": "Charades/videos/RE1M6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turn off a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.6> to <26.1>." + } + ] + }, + { + "video_id": "DH9JU", + "question_id": "DH9JU", + "video_file": "Charades/videos/DH9JU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a broom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.2> to <17.1>." + } + ] + }, + { + "video_id": "DH9JU", + "question_id": "DH9JU", + "video_file": "Charades/videos/DH9JU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins to wash dishes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.8>." + } + ] + }, + { + "video_id": "RFMRB", + "question_id": "RFMRB", + "video_file": "Charades/videos/RFMRB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a man holding a blanket'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.5>." + } + ] + }, + { + "video_id": "RFMRB", + "question_id": "RFMRB", + "video_file": "Charades/videos/RFMRB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is putting a blanket on a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.1> to <17.1>." + } + ] + }, + { + "video_id": "B7HDN", + "question_id": "B7HDN", + "video_file": "Charades/videos/B7HDN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person starts laughing as he is eating' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.0> to <22.0>." + } + ] + }, + { + "video_id": "B7HDN", + "question_id": "B7HDN", + "video_file": "Charades/videos/B7HDN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person eats a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.0> to <27.0>." + } + ] + }, + { + "video_id": "B7HDN", + "question_id": "B7HDN", + "video_file": "Charades/videos/B7HDN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <27.0>." + } + ] + }, + { + "video_id": "B7HDN", + "question_id": "B7HDN", + "video_file": "Charades/videos/B7HDN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person resumes eating' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <27.0>." + } + ] + }, + { + "video_id": "B7HDN", + "question_id": "B7HDN", + "video_file": "Charades/videos/B7HDN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they pause to laugh at something on the screen'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <22.0>." + } + ] + }, + { + "video_id": "42PMR", + "question_id": "42PMR", + "video_file": "Charades/videos/42PMR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.5> to <19.9>." + } + ] + }, + { + "video_id": "WIOOY", + "question_id": "WIOOY", + "video_file": "Charades/videos/WIOOY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts bag of groceries on table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.9> to <9.1>." + } + ] + }, + { + "video_id": "WIOOY", + "question_id": "WIOOY", + "video_file": "Charades/videos/WIOOY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person is cooking something' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.4> to <27.0>." + } + ] + }, + { + "video_id": "WIOOY", + "question_id": "WIOOY", + "video_file": "Charades/videos/WIOOY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts cooking'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <27.0>." + } + ] + }, + { + "video_id": "ONT61", + "question_id": "ONT61", + "video_file": "Charades/videos/ONT61.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts a towel onto the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <18.0>." + } + ] + }, + { + "video_id": "ONT61", + "question_id": "ONT61", + "video_file": "Charades/videos/ONT61.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looks at the picture on the wall'." + }, + { + "from": "gpt", + "value": "From <16.9> to <23.1>." + } + ] + }, + { + "video_id": "BCONW", + "question_id": "BCONW", + "video_file": "Charades/videos/BCONW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pours water into a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.3> to <21.4>." + } + ] + }, + { + "video_id": "BCONW", + "question_id": "BCONW", + "video_file": "Charades/videos/BCONW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person holding a bag pours water into a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.3> to <21.4>." + } + ] + }, + { + "video_id": "BCONW", + "question_id": "BCONW", + "video_file": "Charades/videos/BCONW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the glass in the refrigerator'." + }, + { + "from": "gpt", + "value": "From <21.2> to <30.1>." + } + ] + }, + { + "video_id": "V9DQI", + "question_id": "V9DQI", + "video_file": "Charades/videos/V9DQI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating an apple' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.1>." + } + ] + }, + { + "video_id": "V9DQI", + "question_id": "V9DQI", + "video_file": "Charades/videos/V9DQI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person seated in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.1>." + } + ] + }, + { + "video_id": "V9DQI", + "question_id": "V9DQI", + "video_file": "Charades/videos/V9DQI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person eats' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.1>." + } + ] + }, + { + "video_id": "V9DQI", + "question_id": "V9DQI", + "video_file": "Charades/videos/V9DQI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person watches television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.9> to <30.9>." + } + ] + }, + { + "video_id": "3UWW0", + "question_id": "3UWW0", + "video_file": "Charades/videos/3UWW0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person is eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "3UWW0", + "question_id": "3UWW0", + "video_file": "Charades/videos/3UWW0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking from a glass mug' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <23.3>." + } + ] + }, + { + "video_id": "3UWW0", + "question_id": "3UWW0", + "video_file": "Charades/videos/3UWW0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "WTU0I", + "question_id": "WTU0I", + "video_file": "Charades/videos/WTU0I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "WTU0I", + "question_id": "WTU0I", + "video_file": "Charades/videos/WTU0I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sitting in a chair'." + }, + { + "from": "gpt", + "value": "From <37.2> to <47.4>." + } + ] + }, + { + "video_id": "UDF8X", + "question_id": "UDF8X", + "video_file": "Charades/videos/UDF8X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person the open the bathroom door'." + }, + { + "from": "gpt", + "value": "From <9.6> to <17.0>." + } + ] + }, + { + "video_id": "8J723", + "question_id": "8J723", + "video_file": "Charades/videos/8J723.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looking out the window'." + }, + { + "from": "gpt", + "value": "From <29.6> to <36.3>." + } + ] + }, + { + "video_id": "8J723", + "question_id": "8J723", + "video_file": "Charades/videos/8J723.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.9> to <35.5>." + } + ] + }, + { + "video_id": "7UZZI", + "question_id": "7UZZI", + "video_file": "Charades/videos/7UZZI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person start sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.1> to <13.6>." + } + ] + }, + { + "video_id": "U6L1X", + "question_id": "U6L1X", + "video_file": "Charades/videos/U6L1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turned the light back off' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.0>." + } + ] + }, + { + "video_id": "U6L1X", + "question_id": "U6L1X", + "video_file": "Charades/videos/U6L1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person turned on the light in the pantry' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.0>." + } + ] + }, + { + "video_id": "U6L1X", + "question_id": "U6L1X", + "video_file": "Charades/videos/U6L1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.0>." + } + ] + }, + { + "video_id": "U6L1X", + "question_id": "U6L1X", + "video_file": "Charades/videos/U6L1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the broom back in the pantry' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.8> to <24.6>." + } + ] + }, + { + "video_id": "U6L1X", + "question_id": "U6L1X", + "video_file": "Charades/videos/U6L1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.4> to <27.3>." + } + ] + }, + { + "video_id": "U6L1X", + "question_id": "U6L1X", + "video_file": "Charades/videos/U6L1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts the broom back into the closet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.8> to <24.6>." + } + ] + }, + { + "video_id": "U6L1X", + "question_id": "U6L1X", + "video_file": "Charades/videos/U6L1X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put the broom back away' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.8> to <24.6>." + } + ] + }, + { + "video_id": "BE43L", + "question_id": "BE43L", + "video_file": "Charades/videos/BE43L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they sit at the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.4> to <32.0>." + } + ] + }, + { + "video_id": "BE43L", + "question_id": "BE43L", + "video_file": "Charades/videos/BE43L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits down at table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.4> to <32.0>." + } + ] + }, + { + "video_id": "78XRT", + "question_id": "78XRT", + "video_file": "Charades/videos/78XRT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person throws a pillow'." + }, + { + "from": "gpt", + "value": "From <7.9> to <17.0>." + } + ] + }, + { + "video_id": "78XRT", + "question_id": "78XRT", + "video_file": "Charades/videos/78XRT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person at the desk begins sneezing'." + }, + { + "from": "gpt", + "value": "From <6.4> to <12.1>." + } + ] + }, + { + "video_id": "GFVSV", + "question_id": "GFVSV", + "video_file": "Charades/videos/GFVSV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the cup down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <25.7>." + } + ] + }, + { + "video_id": "GFVSV", + "question_id": "GFVSV", + "video_file": "Charades/videos/GFVSV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open the book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.3> to <34.0>." + } + ] + }, + { + "video_id": "GFVSV", + "question_id": "GFVSV", + "video_file": "Charades/videos/GFVSV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens a book to start reading' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.3> to <34.0>." + } + ] + }, + { + "video_id": "27JQL", + "question_id": "27JQL", + "video_file": "Charades/videos/27JQL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eat a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <24.1>." + } + ] + }, + { + "video_id": "27JQL", + "question_id": "27JQL", + "video_file": "Charades/videos/27JQL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eats a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <24.1>." + } + ] + }, + { + "video_id": "J8KPE", + "question_id": "J8KPE", + "video_file": "Charades/videos/J8KPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes out a bag of food' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.9> to <18.9>." + } + ] + }, + { + "video_id": "J8KPE", + "question_id": "J8KPE", + "video_file": "Charades/videos/J8KPE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes packet out of cupboard,closes cabinet'." + }, + { + "from": "gpt", + "value": "From <12.1> to <17.9>." + } + ] + }, + { + "video_id": "3GA59", + "question_id": "3GA59", + "video_file": "Charades/videos/3GA59.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'this person runs out of the room with their homework' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.7> to <23.3>." + } + ] + }, + { + "video_id": "KGU21", + "question_id": "KGU21", + "video_file": "Charades/videos/KGU21.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is standing by the window laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <23.2>." + } + ] + }, + { + "video_id": "NH9JS", + "question_id": "NH9JS", + "video_file": "Charades/videos/NH9JS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding the door open'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.4> to <23.8>." + } + ] + }, + { + "video_id": "NH9JS", + "question_id": "NH9JS", + "video_file": "Charades/videos/NH9JS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person opened the door of cup board' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.4> to <23.8>." + } + ] + }, + { + "video_id": "NH9JS", + "question_id": "NH9JS", + "video_file": "Charades/videos/NH9JS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the bag into the cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.2> to <29.5>." + } + ] + }, + { + "video_id": "194WK", + "question_id": "194WK", + "video_file": "Charades/videos/194WK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is reading a book'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.8>." + } + ] + }, + { + "video_id": "194WK", + "question_id": "194WK", + "video_file": "Charades/videos/194WK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person laughs as they read something in a book'." + }, + { + "from": "gpt", + "value": "From <2.5> to <10.8>." + } + ] + }, + { + "video_id": "194WK", + "question_id": "194WK", + "video_file": "Charades/videos/194WK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begins sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.3> to <18.7>." + } + ] + }, + { + "video_id": "MWC7V", + "question_id": "MWC7V", + "video_file": "Charades/videos/MWC7V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person playing games on phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "3V7MT", + "question_id": "3V7MT", + "video_file": "Charades/videos/3V7MT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person flip off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <31.0>." + } + ] + }, + { + "video_id": "3V7MT", + "question_id": "3V7MT", + "video_file": "Charades/videos/3V7MT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <31.0>." + } + ] + }, + { + "video_id": "3V7MT", + "question_id": "3V7MT", + "video_file": "Charades/videos/3V7MT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looks out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "4JSK6", + "question_id": "4JSK6", + "video_file": "Charades/videos/4JSK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person watching television with book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.2> to <44.0>." + } + ] + }, + { + "video_id": "E64R4", + "question_id": "E64R4", + "video_file": "Charades/videos/E64R4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes door behind them' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <32.3> to <37.0>." + } + ] + }, + { + "video_id": "E64R4", + "question_id": "E64R4", + "video_file": "Charades/videos/E64R4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <30.8> to <36.4>." + } + ] + }, + { + "video_id": "E64R4", + "question_id": "E64R4", + "video_file": "Charades/videos/E64R4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is dressing in their garage'." + }, + { + "from": "gpt", + "value": "From <1.6> to <13.6>." + } + ] + }, + { + "video_id": "5Q0CX", + "question_id": "5Q0CX", + "video_file": "Charades/videos/5Q0CX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person awakens' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <12.4>." + } + ] + }, + { + "video_id": "5Q0CX", + "question_id": "5Q0CX", + "video_file": "Charades/videos/5Q0CX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person look out a window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.4> to <30.7>." + } + ] + }, + { + "video_id": "5Q0CX", + "question_id": "5Q0CX", + "video_file": "Charades/videos/5Q0CX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person look out the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.4> to <30.7>." + } + ] + }, + { + "video_id": "5Q0CX", + "question_id": "5Q0CX", + "video_file": "Charades/videos/5Q0CX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looks out a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.4> to <30.7>." + } + ] + }, + { + "video_id": "5Q0CX", + "question_id": "5Q0CX", + "video_file": "Charades/videos/5Q0CX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.2> to <18.7>." + } + ] + }, + { + "video_id": "5Q0CX", + "question_id": "5Q0CX", + "video_file": "Charades/videos/5Q0CX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turn on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.2> to <18.7>." + } + ] + }, + { + "video_id": "XAW3U", + "question_id": "XAW3U", + "video_file": "Charades/videos/XAW3U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.2>." + } + ] + }, + { + "video_id": "XAW3U", + "question_id": "XAW3U", + "video_file": "Charades/videos/XAW3U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is standing by a curtained window eating something' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.2>." + } + ] + }, + { + "video_id": "XAW3U", + "question_id": "XAW3U", + "video_file": "Charades/videos/XAW3U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person eating' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.2>." + } + ] + }, + { + "video_id": "IGT82", + "question_id": "IGT82", + "video_file": "Charades/videos/IGT82.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <7.8>." + } + ] + }, + { + "video_id": "IGT82", + "question_id": "IGT82", + "video_file": "Charades/videos/IGT82.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person walks through the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.9>." + } + ] + }, + { + "video_id": "IGT82", + "question_id": "IGT82", + "video_file": "Charades/videos/IGT82.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person running out' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.7> to <29.8>." + } + ] + }, + { + "video_id": "3DNW1", + "question_id": "3DNW1", + "video_file": "Charades/videos/3DNW1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.4> to <14.8>." + } + ] + }, + { + "video_id": "3DNW1", + "question_id": "3DNW1", + "video_file": "Charades/videos/3DNW1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person running out of the room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.1> to <20.0>." + } + ] + }, + { + "video_id": "3DNW1", + "question_id": "3DNW1", + "video_file": "Charades/videos/3DNW1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person run out of the room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.1> to <20.0>." + } + ] + }, + { + "video_id": "NYXJ6", + "question_id": "NYXJ6", + "video_file": "Charades/videos/NYXJ6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sitting on the chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.7> to <37.0>." + } + ] + }, + { + "video_id": "NYXJ6", + "question_id": "NYXJ6", + "video_file": "Charades/videos/NYXJ6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting at the table doing homework' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.6> to <37.0>." + } + ] + }, + { + "video_id": "NYXJ6", + "question_id": "NYXJ6", + "video_file": "Charades/videos/NYXJ6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts their bag on the dining room table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.4> to <25.8>." + } + ] + }, + { + "video_id": "CLORP", + "question_id": "CLORP", + "video_file": "Charades/videos/CLORP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person sits down on the sofa'." + }, + { + "from": "gpt", + "value": "From <28.2> to <40.0>." + } + ] + }, + { + "video_id": "CLORP", + "question_id": "CLORP", + "video_file": "Charades/videos/CLORP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looks out of the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.6> to <24.4>." + } + ] + }, + { + "video_id": "CLORP", + "question_id": "CLORP", + "video_file": "Charades/videos/CLORP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person reads a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.6> to <40.0>." + } + ] + }, + { + "video_id": "CLORP", + "question_id": "CLORP", + "video_file": "Charades/videos/CLORP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person awakens from a nap on the couch' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.5> to <7.4>." + } + ] + }, + { + "video_id": "CLORP", + "question_id": "CLORP", + "video_file": "Charades/videos/CLORP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person starts eating pizza from the table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.2> to <39.5>." + } + ] + }, + { + "video_id": "CLORP", + "question_id": "CLORP", + "video_file": "Charades/videos/CLORP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looks out the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <24.4>." + } + ] + }, + { + "video_id": "CLORP", + "question_id": "CLORP", + "video_file": "Charades/videos/CLORP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.8> to <37.8>." + } + ] + }, + { + "video_id": "CLORP", + "question_id": "CLORP", + "video_file": "Charades/videos/CLORP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.2> to <39.5>." + } + ] + }, + { + "video_id": "UCPP8", + "question_id": "UCPP8", + "video_file": "Charades/videos/UCPP8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is putting away groceries in the pantry'." + }, + { + "from": "gpt", + "value": "From <22.1> to <38.3>." + } + ] + }, + { + "video_id": "UCPP8", + "question_id": "UCPP8", + "video_file": "Charades/videos/UCPP8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes off shoes'." + }, + { + "from": "gpt", + "value": "From <34.6> to <48.9>." + } + ] + }, + { + "video_id": "UCPP8", + "question_id": "UCPP8", + "video_file": "Charades/videos/UCPP8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <34.6> to <48.9>." + } + ] + }, + { + "video_id": "UCPP8", + "question_id": "UCPP8", + "video_file": "Charades/videos/UCPP8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person take their shoes off' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <34.6> to <48.9>." + } + ] + }, + { + "video_id": "F76GD", + "question_id": "F76GD", + "video_file": "Charades/videos/F76GD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they stand up' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.7> to <14.3>." + } + ] + }, + { + "video_id": "F76GD", + "question_id": "F76GD", + "video_file": "Charades/videos/F76GD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person wash their hands' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.8> to <35.2>." + } + ] + }, + { + "video_id": "F76GD", + "question_id": "F76GD", + "video_file": "Charades/videos/F76GD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person follow up by washing their hands in the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.8> to <35.2>." + } + ] + }, + { + "video_id": "F76GD", + "question_id": "F76GD", + "video_file": "Charades/videos/F76GD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinks from a glass'." + }, + { + "from": "gpt", + "value": "From <11.8> to <19.9>." + } + ] + }, + { + "video_id": "F76GD", + "question_id": "F76GD", + "video_file": "Charades/videos/F76GD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drink a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.8> to <19.9>." + } + ] + }, + { + "video_id": "F76GD", + "question_id": "F76GD", + "video_file": "Charades/videos/F76GD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person washes their hands' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.8> to <35.2>." + } + ] + }, + { + "video_id": "F76GD", + "question_id": "F76GD", + "video_file": "Charades/videos/F76GD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drink water from a glass on the counter'." + }, + { + "from": "gpt", + "value": "From <11.8> to <19.9>." + } + ] + }, + { + "video_id": "OSENM", + "question_id": "OSENM", + "video_file": "Charades/videos/OSENM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person holding a camera'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.0> to <34.0>." + } + ] + }, + { + "video_id": "OSENM", + "question_id": "OSENM", + "video_file": "Charades/videos/OSENM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person attempting to take a picture of themselves' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.7> to <34.0>." + } + ] + }, + { + "video_id": "NIRNP", + "question_id": "NIRNP", + "video_file": "Charades/videos/NIRNP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs over grabs' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <8.5>." + } + ] + }, + { + "video_id": "5T9CK", + "question_id": "5T9CK", + "video_file": "Charades/videos/5T9CK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinks from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <26.8>." + } + ] + }, + { + "video_id": "5T9CK", + "question_id": "5T9CK", + "video_file": "Charades/videos/5T9CK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person dresses in a winter coat' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.8> to <36.0>." + } + ] + }, + { + "video_id": "DJ6ZW", + "question_id": "DJ6ZW", + "video_file": "Charades/videos/DJ6ZW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinks some water from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.2> to <22.4>." + } + ] + }, + { + "video_id": "DJ6ZW", + "question_id": "DJ6ZW", + "video_file": "Charades/videos/DJ6ZW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is washing their hands in a sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.2> to <7.6>." + } + ] + }, + { + "video_id": "DJ6ZW", + "question_id": "DJ6ZW", + "video_file": "Charades/videos/DJ6ZW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they drink from a glass of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <22.4>." + } + ] + }, + { + "video_id": "PP5KJ", + "question_id": "PP5KJ", + "video_file": "Charades/videos/PP5KJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a vacuum out of a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <16.0>." + } + ] + }, + { + "video_id": "H1GWM", + "question_id": "H1GWM", + "video_file": "Charades/videos/H1GWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person walks to the doorway'." + }, + { + "from": "gpt", + "value": "From <13.7> to <19.6>." + } + ] + }, + { + "video_id": "H1GWM", + "question_id": "H1GWM", + "video_file": "Charades/videos/H1GWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walk through multiple doorways' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.7> to <19.6>." + } + ] + }, + { + "video_id": "H1GWM", + "question_id": "H1GWM", + "video_file": "Charades/videos/H1GWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person washes their hand in a sink' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.9>." + } + ] + }, + { + "video_id": "JG7UX", + "question_id": "JG7UX", + "video_file": "Charades/videos/JG7UX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing in front of the wardrobe' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.1>." + } + ] + }, + { + "video_id": "JG7UX", + "question_id": "JG7UX", + "video_file": "Charades/videos/JG7UX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opens the refrigerator'." + }, + { + "from": "gpt", + "value": "From <22.0> to <29.0>." + } + ] + }, + { + "video_id": "JG7UX", + "question_id": "JG7UX", + "video_file": "Charades/videos/JG7UX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a cabinet to place their hat inside' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "8RSZ4", + "question_id": "8RSZ4", + "video_file": "Charades/videos/8RSZ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person looks out the window'." + }, + { + "from": "gpt", + "value": "From <1.9> to <12.7>." + } + ] + }, + { + "video_id": "8RSZ4", + "question_id": "8RSZ4", + "video_file": "Charades/videos/8RSZ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they look out the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <12.7>." + } + ] + }, + { + "video_id": "SJGZ2", + "question_id": "SJGZ2", + "video_file": "Charades/videos/SJGZ2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens up a cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.5> to <9.1>." + } + ] + }, + { + "video_id": "SJGZ2", + "question_id": "SJGZ2", + "video_file": "Charades/videos/SJGZ2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person opens the doors of the cupboard' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.5> to <9.1>." + } + ] + }, + { + "video_id": "HJOJ0", + "question_id": "HJOJ0", + "video_file": "Charades/videos/HJOJ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs over to the sofa'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <7.0>." + } + ] + }, + { + "video_id": "HJOJ0", + "question_id": "HJOJ0", + "video_file": "Charades/videos/HJOJ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is running into the living room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <7.0>." + } + ] + }, + { + "video_id": "1THHL", + "question_id": "1THHL", + "video_file": "Charades/videos/1THHL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts smiling' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.0> to <27.8>." + } + ] + }, + { + "video_id": "1THHL", + "question_id": "1THHL", + "video_file": "Charades/videos/1THHL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person is opening medicine cabinets' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.1>." + } + ] + }, + { + "video_id": "T8XGN", + "question_id": "T8XGN", + "video_file": "Charades/videos/T8XGN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.8> to <17.9>." + } + ] + }, + { + "video_id": "T8XGN", + "question_id": "T8XGN", + "video_file": "Charades/videos/T8XGN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a door'." + }, + { + "from": "gpt", + "value": "From <11.8> to <17.9>." + } + ] + }, + { + "video_id": "5FDHU", + "question_id": "5FDHU", + "video_file": "Charades/videos/5FDHU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating a sandwich with the other' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <6.1>." + } + ] + }, + { + "video_id": "5FDHU", + "question_id": "5FDHU", + "video_file": "Charades/videos/5FDHU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <6.1>." + } + ] + }, + { + "video_id": "44Q8A", + "question_id": "44Q8A", + "video_file": "Charades/videos/44Q8A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person awakens suddenly' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.2> to <25.4>." + } + ] + }, + { + "video_id": "44Q8A", + "question_id": "44Q8A", + "video_file": "Charades/videos/44Q8A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws the pillow aside' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.9> to <32.0>." + } + ] + }, + { + "video_id": "ZWPS3", + "question_id": "ZWPS3", + "video_file": "Charades/videos/ZWPS3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sit on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.8> to <18.0>." + } + ] + }, + { + "video_id": "ZWPS3", + "question_id": "ZWPS3", + "video_file": "Charades/videos/ZWPS3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person puts groceries onto a table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.4> to <18.2>." + } + ] + }, + { + "video_id": "ZWPS3", + "question_id": "ZWPS3", + "video_file": "Charades/videos/ZWPS3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person the put the bottle' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <17.3>." + } + ] + }, + { + "video_id": "ZWPS3", + "question_id": "ZWPS3", + "video_file": "Charades/videos/ZWPS3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person walks through a doorway holding a bottle'." + }, + { + "from": "gpt", + "value": "From <5.3> to <11.7>." + } + ] + }, + { + "video_id": "ZWPS3", + "question_id": "ZWPS3", + "video_file": "Charades/videos/ZWPS3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person sits on a chair working on a laptop'." + }, + { + "from": "gpt", + "value": "From <12.8> to <18.0>." + } + ] + }, + { + "video_id": "HFW6H", + "question_id": "HFW6H", + "video_file": "Charades/videos/HFW6H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they both throw the shirts on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <35.1> to <43.3>." + } + ] + }, + { + "video_id": "VSZ4W", + "question_id": "VSZ4W", + "video_file": "Charades/videos/VSZ4W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person drinks from a glass'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "VSZ4W", + "question_id": "VSZ4W", + "video_file": "Charades/videos/VSZ4W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking from a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "VAPVR", + "question_id": "VAPVR", + "video_file": "Charades/videos/VAPVR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person looks out a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.7> to <30.4>." + } + ] + }, + { + "video_id": "W495O", + "question_id": "W495O", + "video_file": "Charades/videos/W495O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts cooking' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.0> to <20.0>." + } + ] + }, + { + "video_id": "WLA52", + "question_id": "WLA52", + "video_file": "Charades/videos/WLA52.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts down the blanket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.8> to <23.0>." + } + ] + }, + { + "video_id": "WLA52", + "question_id": "WLA52", + "video_file": "Charades/videos/WLA52.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts down the blanket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.8> to <23.0>." + } + ] + }, + { + "video_id": "TMATS", + "question_id": "TMATS", + "video_file": "Charades/videos/TMATS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts shoes on pillow'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <17.9>." + } + ] + }, + { + "video_id": "NO0WS", + "question_id": "NO0WS", + "video_file": "Charades/videos/NO0WS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they begin to put away their groceries' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.0> to <18.7>." + } + ] + }, + { + "video_id": "NO0WS", + "question_id": "NO0WS", + "video_file": "Charades/videos/NO0WS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is standing in their pantry eating a cookie' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.2>." + } + ] + }, + { + "video_id": "NO0WS", + "question_id": "NO0WS", + "video_file": "Charades/videos/NO0WS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating something' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.2>." + } + ] + }, + { + "video_id": "UAWXI", + "question_id": "UAWXI", + "video_file": "Charades/videos/UAWXI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person runs into a room holding an envelope' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.5>." + } + ] + }, + { + "video_id": "9OEWM", + "question_id": "9OEWM", + "video_file": "Charades/videos/9OEWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person start turning the light on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <17.2>." + } + ] + }, + { + "video_id": "9OEWM", + "question_id": "9OEWM", + "video_file": "Charades/videos/9OEWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person run in a circle around the table'." + }, + { + "from": "gpt", + "value": "From <6.7> to <16.6>." + } + ] + }, + { + "video_id": "9OEWM", + "question_id": "9OEWM", + "video_file": "Charades/videos/9OEWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns on a light'." + }, + { + "from": "gpt", + "value": "From <12.0> to <17.2>." + } + ] + }, + { + "video_id": "9OEWM", + "question_id": "9OEWM", + "video_file": "Charades/videos/9OEWM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person runs around'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.7> to <16.6>." + } + ] + }, + { + "video_id": "MV37M", + "question_id": "MV37M", + "video_file": "Charades/videos/MV37M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins throwing them onto the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.1> to <18.5>." + } + ] + }, + { + "video_id": "MV37M", + "question_id": "MV37M", + "video_file": "Charades/videos/MV37M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating some sort of food' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "MV37M", + "question_id": "MV37M", + "video_file": "Charades/videos/MV37M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person laughing at how silly they seem' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.4> to <24.0>." + } + ] + }, + { + "video_id": "MV37M", + "question_id": "MV37M", + "video_file": "Charades/videos/MV37M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person in the kitchen is eating some groceries'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "MV37M", + "question_id": "MV37M", + "video_file": "Charades/videos/MV37M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they begin throwing some clothes on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.1> to <18.5>." + } + ] + }, + { + "video_id": "ITXRK", + "question_id": "ITXRK", + "video_file": "Charades/videos/ITXRK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they stand up' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <48.6> to <55.1>." + } + ] + }, + { + "video_id": "ITXRK", + "question_id": "ITXRK", + "video_file": "Charades/videos/ITXRK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is laying on the floor of a kitchen' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.0> to <10.5>." + } + ] + }, + { + "video_id": "ITXRK", + "question_id": "ITXRK", + "video_file": "Charades/videos/ITXRK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person undresses from their jacket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <32.2> to <51.2>." + } + ] + }, + { + "video_id": "ITXRK", + "question_id": "ITXRK", + "video_file": "Charades/videos/ITXRK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person start undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <32.2> to <51.2>." + } + ] + }, + { + "video_id": "ITXRK", + "question_id": "ITXRK", + "video_file": "Charades/videos/ITXRK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'there is a person laying on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <10.5>." + } + ] + }, + { + "video_id": "SXYLN", + "question_id": "SXYLN", + "video_file": "Charades/videos/SXYLN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person throws a bag of groceries on the ground' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.9> to <25.7>." + } + ] + }, + { + "video_id": "SXYLN", + "question_id": "SXYLN", + "video_file": "Charades/videos/SXYLN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they throw the sandwich wrapper in the garbage' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.9> to <25.7>." + } + ] + }, + { + "video_id": "SXYLN", + "question_id": "SXYLN", + "video_file": "Charades/videos/SXYLN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put the bag down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <6.3>." + } + ] + }, + { + "video_id": "S7RGX", + "question_id": "S7RGX", + "video_file": "Charades/videos/S7RGX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put it back on the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.9> to <27.6>." + } + ] + }, + { + "video_id": "FZJTT", + "question_id": "FZJTT", + "video_file": "Charades/videos/FZJTT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakens in the living room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.1> to <11.9>." + } + ] + }, + { + "video_id": "FZJTT", + "question_id": "FZJTT", + "video_file": "Charades/videos/FZJTT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes out their camera' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.0> to <24.0>." + } + ] + }, + { + "video_id": "FZJTT", + "question_id": "FZJTT", + "video_file": "Charades/videos/FZJTT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person putting their shoes on'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <21.6>." + } + ] + }, + { + "video_id": "FZJTT", + "question_id": "FZJTT", + "video_file": "Charades/videos/FZJTT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts their shoes on' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.0> to <21.6>." + } + ] + }, + { + "video_id": "FZJTT", + "question_id": "FZJTT", + "video_file": "Charades/videos/FZJTT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts on their shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.0> to <21.6>." + } + ] + }, + { + "video_id": "0WLCJ", + "question_id": "0WLCJ", + "video_file": "Charades/videos/0WLCJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person throws a blanket on the sofa' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.8> to <7.2>." + } + ] + }, + { + "video_id": "0WLCJ", + "question_id": "0WLCJ", + "video_file": "Charades/videos/0WLCJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws blanket onto couch'." + }, + { + "from": "gpt", + "value": "From <1.8> to <7.2>." + } + ] + }, + { + "video_id": "8WJIR", + "question_id": "8WJIR", + "video_file": "Charades/videos/8WJIR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sitting on a bed watching a television' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.6> to <11.2>." + } + ] + }, + { + "video_id": "8WJIR", + "question_id": "8WJIR", + "video_file": "Charades/videos/8WJIR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person sits on the bed watching television'." + }, + { + "from": "gpt", + "value": "From <0.6> to <11.2>." + } + ] + }, + { + "video_id": "8WJIR", + "question_id": "8WJIR", + "video_file": "Charades/videos/8WJIR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person goes back to the bed sits down'." + }, + { + "from": "gpt", + "value": "From <10.2> to <21.3>." + } + ] + }, + { + "video_id": "8WJIR", + "question_id": "8WJIR", + "video_file": "Charades/videos/8WJIR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they start closing the bedroom door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <17.6>." + } + ] + }, + { + "video_id": "8WJIR", + "question_id": "8WJIR", + "video_file": "Charades/videos/8WJIR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking a cup of coffee' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.9> to <23.5>." + } + ] + }, + { + "video_id": "8WJIR", + "question_id": "8WJIR", + "video_file": "Charades/videos/8WJIR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person proceeds to close the doors' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.4> to <17.6>." + } + ] + }, + { + "video_id": "PDK24", + "question_id": "PDK24", + "video_file": "Charades/videos/PDK24.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person turns on the light in their dining room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "PDK24", + "question_id": "PDK24", + "video_file": "Charades/videos/PDK24.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they sit in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.3> to <25.0>." + } + ] + }, + { + "video_id": "PDK24", + "question_id": "PDK24", + "video_file": "Charades/videos/PDK24.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person start sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.9> to <23.8>." + } + ] + }, + { + "video_id": "PDK24", + "question_id": "PDK24", + "video_file": "Charades/videos/PDK24.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.9> to <23.8>." + } + ] + }, + { + "video_id": "PDK24", + "question_id": "PDK24", + "video_file": "Charades/videos/PDK24.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person turns on the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.6>." + } + ] + }, + { + "video_id": "WT3IG", + "question_id": "WT3IG", + "video_file": "Charades/videos/WT3IG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is holding a camera'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.2>." + } + ] + }, + { + "video_id": "57MEO", + "question_id": "57MEO", + "video_file": "Charades/videos/57MEO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "57MEO", + "question_id": "57MEO", + "video_file": "Charades/videos/57MEO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put the book'." + }, + { + "from": "gpt", + "value": "From <7.3> to <15.0>." + } + ] + }, + { + "video_id": "57MEO", + "question_id": "57MEO", + "video_file": "Charades/videos/57MEO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <15.0>." + } + ] + }, + { + "video_id": "57MEO", + "question_id": "57MEO", + "video_file": "Charades/videos/57MEO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "UFVO8", + "question_id": "UFVO8", + "video_file": "Charades/videos/UFVO8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closed the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.6> to <13.2>." + } + ] + }, + { + "video_id": "UFVO8", + "question_id": "UFVO8", + "video_file": "Charades/videos/UFVO8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a door'." + }, + { + "from": "gpt", + "value": "From <3.0> to <9.8>." + } + ] + }, + { + "video_id": "H6ROK", + "question_id": "H6ROK", + "video_file": "Charades/videos/H6ROK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throwing the blanket to the ground' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.1>." + } + ] + }, + { + "video_id": "4JH6O", + "question_id": "4JH6O", + "video_file": "Charades/videos/4JH6O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks into a door holding a blanket' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.7>." + } + ] + }, + { + "video_id": "4JH6O", + "question_id": "4JH6O", + "video_file": "Charades/videos/4JH6O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs into a room' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "EOELV", + "question_id": "EOELV", + "video_file": "Charades/videos/EOELV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person awakens' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.1>." + } + ] + }, + { + "video_id": "OIZT9", + "question_id": "OIZT9", + "video_file": "Charades/videos/OIZT9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts the groceries on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.8> to <15.8>." + } + ] + }, + { + "video_id": "OIZT9", + "question_id": "OIZT9", + "video_file": "Charades/videos/OIZT9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a picture with a camera' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <26.0>." + } + ] + }, + { + "video_id": "OIZT9", + "question_id": "OIZT9", + "video_file": "Charades/videos/OIZT9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a picture of them doing that'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <26.0>." + } + ] + }, + { + "video_id": "7TJ14", + "question_id": "7TJ14", + "video_file": "Charades/videos/7TJ14.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.1> to <26.6>." + } + ] + }, + { + "video_id": "7TJ14", + "question_id": "7TJ14", + "video_file": "Charades/videos/7TJ14.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person fixes something to eat' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.1> to <26.6>." + } + ] + }, + { + "video_id": "7TJ14", + "question_id": "7TJ14", + "video_file": "Charades/videos/7TJ14.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person runs out eating' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.1> to <26.6>." + } + ] + }, + { + "video_id": "8Q0EY", + "question_id": "8Q0EY", + "video_file": "Charades/videos/8Q0EY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person holds a towel' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.5> to <33.0>." + } + ] + }, + { + "video_id": "P8Y27", + "question_id": "P8Y27", + "video_file": "Charades/videos/P8Y27.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person throws a bag on chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.6> to <5.1>." + } + ] + }, + { + "video_id": "P8Y27", + "question_id": "P8Y27", + "video_file": "Charades/videos/P8Y27.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person throws a bag into the pantry' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <5.1>." + } + ] + }, + { + "video_id": "P8Y27", + "question_id": "P8Y27", + "video_file": "Charades/videos/P8Y27.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person wash the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.4> to <10.6>." + } + ] + }, + { + "video_id": "2SIO0", + "question_id": "2SIO0", + "video_file": "Charades/videos/2SIO0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person smiling at a television in a bedroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.0> to <22.1>." + } + ] + }, + { + "video_id": "2SIO0", + "question_id": "2SIO0", + "video_file": "Charades/videos/2SIO0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'laughing person dressed in a towel turns on a television' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.1> to <17.4>." + } + ] + }, + { + "video_id": "Z6IAE", + "question_id": "Z6IAE", + "video_file": "Charades/videos/Z6IAE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person looks out the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <18.6>." + } + ] + }, + { + "video_id": "Z6IAE", + "question_id": "Z6IAE", + "video_file": "Charades/videos/Z6IAE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looks out a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <18.6>." + } + ] + }, + { + "video_id": "Z6IAE", + "question_id": "Z6IAE", + "video_file": "Charades/videos/Z6IAE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.2> to <32.2>." + } + ] + }, + { + "video_id": "Z6IAE", + "question_id": "Z6IAE", + "video_file": "Charades/videos/Z6IAE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opening a door'." + }, + { + "from": "gpt", + "value": "From <25.2> to <32.2>." + } + ] + }, + { + "video_id": "RKU88", + "question_id": "RKU88", + "video_file": "Charades/videos/RKU88.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws it on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.1> to <24.5>." + } + ] + }, + { + "video_id": "RKU88", + "question_id": "RKU88", + "video_file": "Charades/videos/RKU88.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throwing down their dirty clothes as they come off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.7> to <13.4>." + } + ] + }, + { + "video_id": "RKU88", + "question_id": "RKU88", + "video_file": "Charades/videos/RKU88.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throw it on the floor'." + }, + { + "from": "gpt", + "value": "From <19.1> to <24.5>." + } + ] + }, + { + "video_id": "RKU88", + "question_id": "RKU88", + "video_file": "Charades/videos/RKU88.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks through the doorway into their bedroom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "FVSB0", + "question_id": "FVSB0", + "video_file": "Charades/videos/FVSB0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "FVSB0", + "question_id": "FVSB0", + "video_file": "Charades/videos/FVSB0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person holding clothes puts them in the wardrobe' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.5> to <21.3>." + } + ] + }, + { + "video_id": "7YMK9", + "question_id": "7YMK9", + "video_file": "Charades/videos/7YMK9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the door behind them'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <36.3> to <41.0>." + } + ] + }, + { + "video_id": "7YMK9", + "question_id": "7YMK9", + "video_file": "Charades/videos/7YMK9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <34.1> to <39.9>." + } + ] + }, + { + "video_id": "EVHAZ", + "question_id": "EVHAZ", + "video_file": "Charades/videos/EVHAZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person stands on a chair'." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.4>." + } + ] + }, + { + "video_id": "9D2T4", + "question_id": "9D2T4", + "video_file": "Charades/videos/9D2T4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is in the hallway throwing a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.2> to <9.6>." + } + ] + }, + { + "video_id": "7O2LP", + "question_id": "7O2LP", + "video_file": "Charades/videos/7O2LP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throws the dishes they were carrying onto the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <21.8>." + } + ] + }, + { + "video_id": "CBLSW", + "question_id": "CBLSW", + "video_file": "Charades/videos/CBLSW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts a glass in the sink' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.7> to <31.3>." + } + ] + }, + { + "video_id": "1GQAJ", + "question_id": "1GQAJ", + "video_file": "Charades/videos/1GQAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walks in holding a broom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.8> to <17.7>." + } + ] + }, + { + "video_id": "1GQAJ", + "question_id": "1GQAJ", + "video_file": "Charades/videos/1GQAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks in holding a broom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.8> to <17.7>." + } + ] + }, + { + "video_id": "O7NI6", + "question_id": "O7NI6", + "video_file": "Charades/videos/O7NI6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "VO45S", + "question_id": "VO45S", + "video_file": "Charades/videos/VO45S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person holding a shoe box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "VO45S", + "question_id": "VO45S", + "video_file": "Charades/videos/VO45S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the box down'." + }, + { + "from": "gpt", + "value": "From <9.1> to <14.1>." + } + ] + }, + { + "video_id": "VO45S", + "question_id": "VO45S", + "video_file": "Charades/videos/VO45S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.2>." + } + ] + }, + { + "video_id": "N8R3F", + "question_id": "N8R3F", + "video_file": "Charades/videos/N8R3F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throws a pair of shoes at a shelf'." + }, + { + "from": "gpt", + "value": "From <3.9> to <9.2>." + } + ] + }, + { + "video_id": "N8R3F", + "question_id": "N8R3F", + "video_file": "Charades/videos/N8R3F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws some shoes on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <9.2>." + } + ] + }, + { + "video_id": "705RV", + "question_id": "705RV", + "video_file": "Charades/videos/705RV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.9> to <29.5>." + } + ] + }, + { + "video_id": "705RV", + "question_id": "705RV", + "video_file": "Charades/videos/705RV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person start putting the sandwich on a table'." + }, + { + "from": "gpt", + "value": "From <15.5> to <21.8>." + } + ] + }, + { + "video_id": "705RV", + "question_id": "705RV", + "video_file": "Charades/videos/705RV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take some medicine'." + }, + { + "from": "gpt", + "value": "From <18.1> to <26.9>." + } + ] + }, + { + "video_id": "TE421", + "question_id": "TE421", + "video_file": "Charades/videos/TE421.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.3> to <33.0>." + } + ] + }, + { + "video_id": "TE421", + "question_id": "TE421", + "video_file": "Charades/videos/TE421.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person takes some medicine'." + }, + { + "from": "gpt", + "value": "From <1.2> to <6.5>." + } + ] + }, + { + "video_id": "TE421", + "question_id": "TE421", + "video_file": "Charades/videos/TE421.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person talking on the phone'." + }, + { + "from": "gpt", + "value": "From <19.1> to <33.0>." + } + ] + }, + { + "video_id": "TE421", + "question_id": "TE421", + "video_file": "Charades/videos/TE421.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person with a phone closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.3> to <33.0>." + } + ] + }, + { + "video_id": "82XWA", + "question_id": "82XWA", + "video_file": "Charades/videos/82XWA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is awakening' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.2>." + } + ] + }, + { + "video_id": "0QMGP", + "question_id": "0QMGP", + "video_file": "Charades/videos/0QMGP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person set a laptop computer down on a table'." + }, + { + "from": "gpt", + "value": "From <9.6> to <17.0>." + } + ] + }, + { + "video_id": "0QMGP", + "question_id": "0QMGP", + "video_file": "Charades/videos/0QMGP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person wash their hands in the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.7> to <32.0>." + } + ] + }, + { + "video_id": "0QMGP", + "question_id": "0QMGP", + "video_file": "Charades/videos/0QMGP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person washed hands in water running from a faucet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.7> to <32.0>." + } + ] + }, + { + "video_id": "0QMGP", + "question_id": "0QMGP", + "video_file": "Charades/videos/0QMGP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.3> to <17.0>." + } + ] + }, + { + "video_id": "BMLFA", + "question_id": "BMLFA", + "video_file": "Charades/videos/BMLFA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is holding a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.1> to <22.7>." + } + ] + }, + { + "video_id": "BMLFA", + "question_id": "BMLFA", + "video_file": "Charades/videos/BMLFA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put a box away in the refrigerator'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <22.5>." + } + ] + }, + { + "video_id": "KX90W", + "question_id": "KX90W", + "video_file": "Charades/videos/KX90W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walks through a doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.9>." + } + ] + }, + { + "video_id": "KX90W", + "question_id": "KX90W", + "video_file": "Charades/videos/KX90W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person opens a closet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <9.1>." + } + ] + }, + { + "video_id": "5MSZP", + "question_id": "5MSZP", + "video_file": "Charades/videos/5MSZP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is smiling' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.8> to <17.5>." + } + ] + }, + { + "video_id": "5MSZP", + "question_id": "5MSZP", + "video_file": "Charades/videos/5MSZP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person is sitting in a chair eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.1> to <7.9>." + } + ] + }, + { + "video_id": "5MSZP", + "question_id": "5MSZP", + "video_file": "Charades/videos/5MSZP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting towels away' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.2> to <27.3>." + } + ] + }, + { + "video_id": "LPR6Z", + "question_id": "LPR6Z", + "video_file": "Charades/videos/LPR6Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turning the light off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.1> to <33.0>." + } + ] + }, + { + "video_id": "LPR6Z", + "question_id": "LPR6Z", + "video_file": "Charades/videos/LPR6Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person cuts the light off'." + }, + { + "from": "gpt", + "value": "From <28.1> to <33.0>." + } + ] + }, + { + "video_id": "8R0SZ", + "question_id": "8R0SZ", + "video_file": "Charades/videos/8R0SZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs through the doorway' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "8R0SZ", + "question_id": "8R0SZ", + "video_file": "Charades/videos/8R0SZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person running back out'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "WTFOC", + "question_id": "WTFOC", + "video_file": "Charades/videos/WTFOC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they open a closet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.4> to <23.0>." + } + ] + }, + { + "video_id": "PVVQS", + "question_id": "PVVQS", + "video_file": "Charades/videos/PVVQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person the are laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.0> to <21.1>." + } + ] + }, + { + "video_id": "PVVQS", + "question_id": "PVVQS", + "video_file": "Charades/videos/PVVQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is walking through a doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "PVVQS", + "question_id": "PVVQS", + "video_file": "Charades/videos/PVVQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws their shoes down the hallway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.0> to <29.0>." + } + ] + }, + { + "video_id": "PVVQS", + "question_id": "PVVQS", + "video_file": "Charades/videos/PVVQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns on the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <10.6>." + } + ] + }, + { + "video_id": "PVVQS", + "question_id": "PVVQS", + "video_file": "Charades/videos/PVVQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begin throwing their shoes'." + }, + { + "from": "gpt", + "value": "From <23.0> to <29.0>." + } + ] + }, + { + "video_id": "PVVQS", + "question_id": "PVVQS", + "video_file": "Charades/videos/PVVQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns on a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.0> to <10.6>." + } + ] + }, + { + "video_id": "PVVQS", + "question_id": "PVVQS", + "video_file": "Charades/videos/PVVQS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.0> to <21.1>." + } + ] + }, + { + "video_id": "G6BFY", + "question_id": "G6BFY", + "video_file": "Charades/videos/G6BFY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put it back on the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <10.3>." + } + ] + }, + { + "video_id": "G6BFY", + "question_id": "G6BFY", + "video_file": "Charades/videos/G6BFY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person smiles into a mirror' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.0> to <16.8>." + } + ] + }, + { + "video_id": "O8DER", + "question_id": "O8DER", + "video_file": "Charades/videos/O8DER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they stand up' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.7> to <22.6>." + } + ] + }, + { + "video_id": "O8DER", + "question_id": "O8DER", + "video_file": "Charades/videos/O8DER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put a box on it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.6> to <28.6>." + } + ] + }, + { + "video_id": "O8DER", + "question_id": "O8DER", + "video_file": "Charades/videos/O8DER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts a box on the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.7> to <28.6>." + } + ] + }, + { + "video_id": "O8DER", + "question_id": "O8DER", + "video_file": "Charades/videos/O8DER.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is seen laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.1> to <12.9>." + } + ] + }, + { + "video_id": "I31V9", + "question_id": "I31V9", + "video_file": "Charades/videos/I31V9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns on light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <7.4>." + } + ] + }, + { + "video_id": "I31V9", + "question_id": "I31V9", + "video_file": "Charades/videos/I31V9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person was eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "I31V9", + "question_id": "I31V9", + "video_file": "Charades/videos/I31V9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "I31V9", + "question_id": "I31V9", + "video_file": "Charades/videos/I31V9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turned on a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <7.4>." + } + ] + }, + { + "video_id": "I31V9", + "question_id": "I31V9", + "video_file": "Charades/videos/I31V9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they turn on a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <7.4>." + } + ] + }, + { + "video_id": "I31V9", + "question_id": "I31V9", + "video_file": "Charades/videos/I31V9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.8> to <13.0>." + } + ] + }, + { + "video_id": "ZRRC8", + "question_id": "ZRRC8", + "video_file": "Charades/videos/ZRRC8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.7> to <19.7>." + } + ] + }, + { + "video_id": "ZRRC8", + "question_id": "ZRRC8", + "video_file": "Charades/videos/ZRRC8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person awakens in bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.6> to <8.9>." + } + ] + }, + { + "video_id": "ZRRC8", + "question_id": "ZRRC8", + "video_file": "Charades/videos/ZRRC8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they look at the book they are holding' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.4> to <24.0>." + } + ] + }, + { + "video_id": "PM1VD", + "question_id": "PM1VD", + "video_file": "Charades/videos/PM1VD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'this person is sitting in a chair a his desk'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <100.8> to <109.7>." + } + ] + }, + { + "video_id": "AQ69X", + "question_id": "AQ69X", + "video_file": "Charades/videos/AQ69X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts on a pair of shoes'." + }, + { + "from": "gpt", + "value": "From <0.1> to <12.7>." + } + ] + }, + { + "video_id": "AQ69X", + "question_id": "AQ69X", + "video_file": "Charades/videos/AQ69X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person putting on shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <12.7>." + } + ] + }, + { + "video_id": "OOCOF", + "question_id": "OOCOF", + "video_file": "Charades/videos/OOCOF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person wash their hands in the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.6> to <31.0>." + } + ] + }, + { + "video_id": "U0X4P", + "question_id": "U0X4P", + "video_file": "Charades/videos/U0X4P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person looks out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.2>." + } + ] + }, + { + "video_id": "U0X4P", + "question_id": "U0X4P", + "video_file": "Charades/videos/U0X4P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is watching something outside the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.2>." + } + ] + }, + { + "video_id": "U0X4P", + "question_id": "U0X4P", + "video_file": "Charades/videos/U0X4P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person throw their clothes on the bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.2> to <30.8>." + } + ] + }, + { + "video_id": "U0X4P", + "question_id": "U0X4P", + "video_file": "Charades/videos/U0X4P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person looked out the window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.2>." + } + ] + }, + { + "video_id": "N2GFQ", + "question_id": "N2GFQ", + "video_file": "Charades/videos/N2GFQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts a bag on a shelf' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.9> to <13.5>." + } + ] + }, + { + "video_id": "N2GFQ", + "question_id": "N2GFQ", + "video_file": "Charades/videos/N2GFQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person goes to wash their hands'." + }, + { + "from": "gpt", + "value": "From <14.0> to <30.0>." + } + ] + }, + { + "video_id": "N2GFQ", + "question_id": "N2GFQ", + "video_file": "Charades/videos/N2GFQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'that person takes off their shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <12.0>." + } + ] + }, + { + "video_id": "N2GFQ", + "question_id": "N2GFQ", + "video_file": "Charades/videos/N2GFQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person washes hands'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <30.0>." + } + ] + }, + { + "video_id": "N2GFQ", + "question_id": "N2GFQ", + "video_file": "Charades/videos/N2GFQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person washes their hands'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <30.0>." + } + ] + }, + { + "video_id": "N2GFQ", + "question_id": "N2GFQ", + "video_file": "Charades/videos/N2GFQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes their shoes off' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.0> to <12.0>." + } + ] + }, + { + "video_id": "9Z1A3", + "question_id": "9Z1A3", + "video_file": "Charades/videos/9Z1A3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts shoes on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <12.9>." + } + ] + }, + { + "video_id": "9Z1A3", + "question_id": "9Z1A3", + "video_file": "Charades/videos/9Z1A3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is quickly dressing in the dining room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.9>." + } + ] + }, + { + "video_id": "9Z1A3", + "question_id": "9Z1A3", + "video_file": "Charades/videos/9Z1A3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts on a pair of shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <12.9>." + } + ] + }, + { + "video_id": "NWNUY", + "question_id": "NWNUY", + "video_file": "Charades/videos/NWNUY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts away some food onto a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.5> to <25.8>." + } + ] + }, + { + "video_id": "NWNUY", + "question_id": "NWNUY", + "video_file": "Charades/videos/NWNUY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put some food on it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.5> to <25.8>." + } + ] + }, + { + "video_id": "RY8ZS", + "question_id": "RY8ZS", + "video_file": "Charades/videos/RY8ZS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is washing a mirror with a blanket'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.1>." + } + ] + }, + { + "video_id": "CD11G", + "question_id": "CD11G", + "video_file": "Charades/videos/CD11G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is cooking food on a stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "CD11G", + "question_id": "CD11G", + "video_file": "Charades/videos/CD11G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person open the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.0> to <11.9>." + } + ] + }, + { + "video_id": "CD11G", + "question_id": "CD11G", + "video_file": "Charades/videos/CD11G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is cooking on the stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "CD11G", + "question_id": "CD11G", + "video_file": "Charades/videos/CD11G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating in a kitchen'." + }, + { + "from": "gpt", + "value": "From <9.8> to <16.2>." + } + ] + }, + { + "video_id": "CD11G", + "question_id": "CD11G", + "video_file": "Charades/videos/CD11G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.7> to <19.0>." + } + ] + }, + { + "video_id": "CD11G", + "question_id": "CD11G", + "video_file": "Charades/videos/CD11G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the refrigerator door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <11.9>." + } + ] + }, + { + "video_id": "CD11G", + "question_id": "CD11G", + "video_file": "Charades/videos/CD11G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening a refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.0> to <11.9>." + } + ] + }, + { + "video_id": "43OXO", + "question_id": "43OXO", + "video_file": "Charades/videos/43OXO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person awakens' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "43OXO", + "question_id": "43OXO", + "video_file": "Charades/videos/43OXO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts a small box on top of the notebook' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.9> to <15.1>." + } + ] + }, + { + "video_id": "43OXO", + "question_id": "43OXO", + "video_file": "Charades/videos/43OXO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits up on the couch' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <7.9>." + } + ] + }, + { + "video_id": "O9WFN", + "question_id": "O9WFN", + "video_file": "Charades/videos/O9WFN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is opening a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "O9WFN", + "question_id": "O9WFN", + "video_file": "Charades/videos/O9WFN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pouring coffee into a cup from a thermos' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <16.3>." + } + ] + }, + { + "video_id": "O9WFN", + "question_id": "O9WFN", + "video_file": "Charades/videos/O9WFN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens the door to their closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "FQLS1", + "question_id": "FQLS1", + "video_file": "Charades/videos/FQLS1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person smiling for a camera'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.8> to <13.1>." + } + ] + }, + { + "video_id": "FQAAB", + "question_id": "FQAAB", + "video_file": "Charades/videos/FQAAB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person close the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <28.3> to <37.0>." + } + ] + }, + { + "video_id": "FQAAB", + "question_id": "FQAAB", + "video_file": "Charades/videos/FQAAB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs up the stairs'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "FQAAB", + "question_id": "FQAAB", + "video_file": "Charades/videos/FQAAB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they are closing the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.3> to <37.0>." + } + ] + }, + { + "video_id": "FQAAB", + "question_id": "FQAAB", + "video_file": "Charades/videos/FQAAB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is running through the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "FQAAB", + "question_id": "FQAAB", + "video_file": "Charades/videos/FQAAB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take a phone from their pocket'." + }, + { + "from": "gpt", + "value": "From <24.9> to <32.4>." + } + ] + }, + { + "video_id": "XJH5D", + "question_id": "XJH5D", + "video_file": "Charades/videos/XJH5D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts it on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <25.8> to <30.8>." + } + ] + }, + { + "video_id": "XJH5D", + "question_id": "XJH5D", + "video_file": "Charades/videos/XJH5D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.1> to <14.1>." + } + ] + }, + { + "video_id": "VE52Z", + "question_id": "VE52Z", + "video_file": "Charades/videos/VE52Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks in the room eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.4> to <10.7>." + } + ] + }, + { + "video_id": "VE52Z", + "question_id": "VE52Z", + "video_file": "Charades/videos/VE52Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throw the sandwich on the floor'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.0> to <36.0>." + } + ] + }, + { + "video_id": "VE52Z", + "question_id": "VE52Z", + "video_file": "Charades/videos/VE52Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person throws the food to the floor'." + }, + { + "from": "gpt", + "value": "From <30.7> to <36.0>." + } + ] + }, + { + "video_id": "8E5AD", + "question_id": "8E5AD", + "video_file": "Charades/videos/8E5AD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they begin eating' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.3> to <33.0>." + } + ] + }, + { + "video_id": "8E5AD", + "question_id": "8E5AD", + "video_file": "Charades/videos/8E5AD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.3> to <33.0>." + } + ] + }, + { + "video_id": "8E5AD", + "question_id": "8E5AD", + "video_file": "Charades/videos/8E5AD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts eating sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.3> to <33.0>." + } + ] + }, + { + "video_id": "1KDFH", + "question_id": "1KDFH", + "video_file": "Charades/videos/1KDFH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is sneezing'." + }, + { + "from": "gpt", + "value": "From <13.9> to <23.4>." + } + ] + }, + { + "video_id": "C278V", + "question_id": "C278V", + "video_file": "Charades/videos/C278V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a woman is sitting on a sofa folding laundry'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.6> to <32.0>." + } + ] + }, + { + "video_id": "C278V", + "question_id": "C278V", + "video_file": "Charades/videos/C278V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a man takes a pile of clothes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.6> to <32.0>." + } + ] + }, + { + "video_id": "C278V", + "question_id": "C278V", + "video_file": "Charades/videos/C278V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person is putting clothes on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.6> to <14.8>." + } + ] + }, + { + "video_id": "C278V", + "question_id": "C278V", + "video_file": "Charades/videos/C278V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts them on the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.6> to <14.8>." + } + ] + }, + { + "video_id": "9TL3J", + "question_id": "9TL3J", + "video_file": "Charades/videos/9TL3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they drink a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.2> to <30.3>." + } + ] + }, + { + "video_id": "9TL3J", + "question_id": "9TL3J", + "video_file": "Charades/videos/9TL3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <31.9> to <36.9>." + } + ] + }, + { + "video_id": "9TL3J", + "question_id": "9TL3J", + "video_file": "Charades/videos/9TL3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is washing dishes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "9TL3J", + "question_id": "9TL3J", + "video_file": "Charades/videos/9TL3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turning off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.9> to <36.9>." + } + ] + }, + { + "video_id": "9TL3J", + "question_id": "9TL3J", + "video_file": "Charades/videos/9TL3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turn off the kitchen light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <31.9> to <36.9>." + } + ] + }, + { + "video_id": "9TL3J", + "question_id": "9TL3J", + "video_file": "Charades/videos/9TL3J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person washes dishes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "IWMW1", + "question_id": "IWMW1", + "video_file": "Charades/videos/IWMW1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is drinking a glass of water'." + }, + { + "from": "gpt", + "value": "From <6.5> to <12.3>." + } + ] + }, + { + "video_id": "IWMW1", + "question_id": "IWMW1", + "video_file": "Charades/videos/IWMW1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they tidy some dishes off the floor' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.2> to <25.8>." + } + ] + }, + { + "video_id": "561H5", + "question_id": "561H5", + "video_file": "Charades/videos/561H5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a towel' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "561H5", + "question_id": "561H5", + "video_file": "Charades/videos/561H5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'eating a sandwich as the person does so' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.9> to <18.8>." + } + ] + }, + { + "video_id": "3W31X", + "question_id": "3W31X", + "video_file": "Charades/videos/3W31X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts groceries on table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.7> to <27.0>." + } + ] + }, + { + "video_id": "3W31X", + "question_id": "3W31X", + "video_file": "Charades/videos/3W31X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.0> to <27.0>." + } + ] + }, + { + "video_id": "CH8XJ", + "question_id": "CH8XJ", + "video_file": "Charades/videos/CH8XJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open the closet'." + }, + { + "from": "gpt", + "value": "From <13.5> to <20.1>." + } + ] + }, + { + "video_id": "CH8XJ", + "question_id": "CH8XJ", + "video_file": "Charades/videos/CH8XJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person starting throwing shoes into their closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.9> to <25.4>." + } + ] + }, + { + "video_id": "CH8XJ", + "question_id": "CH8XJ", + "video_file": "Charades/videos/CH8XJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person throw shoes inside'." + }, + { + "from": "gpt", + "value": "From <16.9> to <25.4>." + } + ] + }, + { + "video_id": "E661I", + "question_id": "E661I", + "video_file": "Charades/videos/E661I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person is running in shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <13.8>." + } + ] + }, + { + "video_id": "E661I", + "question_id": "E661I", + "video_file": "Charades/videos/E661I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person runs to another room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <13.8>." + } + ] + }, + { + "video_id": "E661I", + "question_id": "E661I", + "video_file": "Charades/videos/E661I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the first person keeps sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "E661I", + "question_id": "E661I", + "video_file": "Charades/videos/E661I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sneezing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.5>." + } + ] + }, + { + "video_id": "1VMZM", + "question_id": "1VMZM", + "video_file": "Charades/videos/1VMZM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.8> to <8.8>." + } + ] + }, + { + "video_id": "1VMZM", + "question_id": "1VMZM", + "video_file": "Charades/videos/1VMZM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs out' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.8> to <28.3>." + } + ] + }, + { + "video_id": "1VMZM", + "question_id": "1VMZM", + "video_file": "Charades/videos/1VMZM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks into their laundry room holding a vacuum' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <14.3>." + } + ] + }, + { + "video_id": "1VMZM", + "question_id": "1VMZM", + "video_file": "Charades/videos/1VMZM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws some clothes into a bin' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <23.3>." + } + ] + }, + { + "video_id": "1VMZM", + "question_id": "1VMZM", + "video_file": "Charades/videos/1VMZM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person run out of the room'." + }, + { + "from": "gpt", + "value": "From <22.8> to <28.3>." + } + ] + }, + { + "video_id": "V24A5", + "question_id": "V24A5", + "video_file": "Charades/videos/V24A5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts down the food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <23.8>." + } + ] + }, + { + "video_id": "V24A5", + "question_id": "V24A5", + "video_file": "Charades/videos/V24A5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walks through the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <31.0>." + } + ] + }, + { + "video_id": "GP5CT", + "question_id": "GP5CT", + "video_file": "Charades/videos/GP5CT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they drink a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.8> to <31.0>." + } + ] + }, + { + "video_id": "GP5CT", + "question_id": "GP5CT", + "video_file": "Charades/videos/GP5CT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is laughing'." + }, + { + "from": "gpt", + "value": "From <12.3> to <17.4>." + } + ] + }, + { + "video_id": "R4FOQ", + "question_id": "R4FOQ", + "video_file": "Charades/videos/R4FOQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is opening a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.0>." + } + ] + }, + { + "video_id": "GMTL9", + "question_id": "GMTL9", + "video_file": "Charades/videos/GMTL9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.1> to <29.9>." + } + ] + }, + { + "video_id": "GMTL9", + "question_id": "GMTL9", + "video_file": "Charades/videos/GMTL9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person was eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <14.2> to <20.7>." + } + ] + }, + { + "video_id": "NFXPI", + "question_id": "NFXPI", + "video_file": "Charades/videos/NFXPI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person washes a dirty mirror' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.6> to <31.0>." + } + ] + }, + { + "video_id": "WU8I6", + "question_id": "WU8I6", + "video_file": "Charades/videos/WU8I6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pours medicine out of a bottle'." + }, + { + "from": "gpt", + "value": "From <6.5> to <16.0>." + } + ] + }, + { + "video_id": "UF91R", + "question_id": "UF91R", + "video_file": "Charades/videos/UF91R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walk over to the refrigerator to open it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.9> to <14.3>." + } + ] + }, + { + "video_id": "UF91R", + "question_id": "UF91R", + "video_file": "Charades/videos/UF91R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person close the laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <12.1>." + } + ] + }, + { + "video_id": "UF91R", + "question_id": "UF91R", + "video_file": "Charades/videos/UF91R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person started out on their laptop completing homework' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.8>." + } + ] + }, + { + "video_id": "PUY9Q", + "question_id": "PUY9Q", + "video_file": "Charades/videos/PUY9Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is in a kitchen putting food away' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.6> to <19.4>." + } + ] + }, + { + "video_id": "ZS66D", + "question_id": "ZS66D", + "video_file": "Charades/videos/ZS66D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes a drink from a bottle of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.4> to <24.1>." + } + ] + }, + { + "video_id": "ZS66D", + "question_id": "ZS66D", + "video_file": "Charades/videos/ZS66D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "ZS66D", + "question_id": "ZS66D", + "video_file": "Charades/videos/ZS66D.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the bag down'." + }, + { + "from": "gpt", + "value": "From <10.7> to <16.0>." + } + ] + }, + { + "video_id": "10LP3", + "question_id": "10LP3", + "video_file": "Charades/videos/10LP3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws it on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.2> to <16.1>." + } + ] + }, + { + "video_id": "10LP3", + "question_id": "10LP3", + "video_file": "Charades/videos/10LP3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walking through a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "10LP3", + "question_id": "10LP3", + "video_file": "Charades/videos/10LP3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throwing it on the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.2> to <16.1>." + } + ] + }, + { + "video_id": "8CJML", + "question_id": "8CJML", + "video_file": "Charades/videos/8CJML.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person snuggle with a pillow they got from a table'." + }, + { + "from": "gpt", + "value": "From <17.6> to <30.0>." + } + ] + }, + { + "video_id": "ONA0S", + "question_id": "ONA0S", + "video_file": "Charades/videos/ONA0S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opening the doorknob'." + }, + { + "from": "gpt", + "value": "From <9.5> to <19.5>." + } + ] + }, + { + "video_id": "ONA0S", + "question_id": "ONA0S", + "video_file": "Charades/videos/ONA0S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person is opening a wooden door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <19.5>." + } + ] + }, + { + "video_id": "5Q8C7", + "question_id": "5Q8C7", + "video_file": "Charades/videos/5Q8C7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a picture on their phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.9> to <31.0>." + } + ] + }, + { + "video_id": "9A5MP", + "question_id": "9A5MP", + "video_file": "Charades/videos/9A5MP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person eats the snack'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.3> to <25.6>." + } + ] + }, + { + "video_id": "9A5MP", + "question_id": "9A5MP", + "video_file": "Charades/videos/9A5MP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closes the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.5> to <34.0>." + } + ] + }, + { + "video_id": "9A5MP", + "question_id": "9A5MP", + "video_file": "Charades/videos/9A5MP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person tidying the area with a broom stops' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.3>." + } + ] + }, + { + "video_id": "9A5MP", + "question_id": "9A5MP", + "video_file": "Charades/videos/9A5MP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a bite of food removed from a bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.2> to <22.3>." + } + ] + }, + { + "video_id": "9A5MP", + "question_id": "9A5MP", + "video_file": "Charades/videos/9A5MP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person uses a broom to tidy up the garage' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.3>." + } + ] + }, + { + "video_id": "R6H5P", + "question_id": "R6H5P", + "video_file": "Charades/videos/R6H5P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person gets up smiling' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.0> to <13.4>." + } + ] + }, + { + "video_id": "R6H5P", + "question_id": "R6H5P", + "video_file": "Charades/videos/R6H5P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person throws a lot of clothes into a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.0> to <10.0>." + } + ] + }, + { + "video_id": "R6H5P", + "question_id": "R6H5P", + "video_file": "Charades/videos/R6H5P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a second person walks through the door holding dishes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <9.1>." + } + ] + }, + { + "video_id": "R6H5P", + "question_id": "R6H5P", + "video_file": "Charades/videos/R6H5P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person holding dishes comes in the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <9.1>." + } + ] + }, + { + "video_id": "MLDGN", + "question_id": "MLDGN", + "video_file": "Charades/videos/MLDGN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person start watching the picture' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.1> to <11.5>." + } + ] + }, + { + "video_id": "80E3X", + "question_id": "80E3X", + "video_file": "Charades/videos/80E3X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person awakens'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.8>." + } + ] + }, + { + "video_id": "80E3X", + "question_id": "80E3X", + "video_file": "Charades/videos/80E3X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person looks at a picture hanging on the wall'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.1> to <31.0>." + } + ] + }, + { + "video_id": "80E3X", + "question_id": "80E3X", + "video_file": "Charades/videos/80E3X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looking at a picture on a wall'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.1> to <31.0>." + } + ] + }, + { + "video_id": "80E3X", + "question_id": "80E3X", + "video_file": "Charades/videos/80E3X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turn on a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.8> to <14.1>." + } + ] + }, + { + "video_id": "80E3X", + "question_id": "80E3X", + "video_file": "Charades/videos/80E3X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins eating a bag of chips' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.0> to <31.0>." + } + ] + }, + { + "video_id": "80E3X", + "question_id": "80E3X", + "video_file": "Charades/videos/80E3X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person turns on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <14.1>." + } + ] + }, + { + "video_id": "80E3X", + "question_id": "80E3X", + "video_file": "Charades/videos/80E3X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begin eating'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.0> to <31.0>." + } + ] + }, + { + "video_id": "80E3X", + "question_id": "80E3X", + "video_file": "Charades/videos/80E3X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person awakens in bed'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.8>." + } + ] + }, + { + "video_id": "WNW0S", + "question_id": "WNW0S", + "video_file": "Charades/videos/WNW0S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person can run with a broom' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.7>." + } + ] + }, + { + "video_id": "WNW0S", + "question_id": "WNW0S", + "video_file": "Charades/videos/WNW0S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs around the room holding a broom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.8>." + } + ] + }, + { + "video_id": "4HFN2", + "question_id": "4HFN2", + "video_file": "Charades/videos/4HFN2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they begin cooking'." + }, + { + "from": "gpt", + "value": "From <27.5> to <38.0>." + } + ] + }, + { + "video_id": "AFTMI", + "question_id": "AFTMI", + "video_file": "Charades/videos/AFTMI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a pillow off a shelf'." + }, + { + "from": "gpt", + "value": "From <32.3> to <37.8>." + } + ] + }, + { + "video_id": "AFTMI", + "question_id": "AFTMI", + "video_file": "Charades/videos/AFTMI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person takes a pillow from the shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.3> to <37.8>." + } + ] + }, + { + "video_id": "1X9A1", + "question_id": "1X9A1", + "video_file": "Charades/videos/1X9A1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take some medicine' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.6> to <29.4>." + } + ] + }, + { + "video_id": "ZDFOL", + "question_id": "ZDFOL", + "video_file": "Charades/videos/ZDFOL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts the book down very hard on the desk' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <17.1>." + } + ] + }, + { + "video_id": "ZDFOL", + "question_id": "ZDFOL", + "video_file": "Charades/videos/ZDFOL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'holding the book the person decides to put on shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <17.1>." + } + ] + }, + { + "video_id": "ZDFOL", + "question_id": "ZDFOL", + "video_file": "Charades/videos/ZDFOL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person reading a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.3>." + } + ] + }, + { + "video_id": "I0IW5", + "question_id": "I0IW5", + "video_file": "Charades/videos/I0IW5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits on the couch' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.0> to <19.0>." + } + ] + }, + { + "video_id": "I0IW5", + "question_id": "I0IW5", + "video_file": "Charades/videos/I0IW5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person a boy comes into the room eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.4> to <10.3>." + } + ] + }, + { + "video_id": "I0IW5", + "question_id": "I0IW5", + "video_file": "Charades/videos/I0IW5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the sandwich on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <10.3>." + } + ] + }, + { + "video_id": "I0IW5", + "question_id": "I0IW5", + "video_file": "Charades/videos/I0IW5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts it on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <10.3>." + } + ] + }, + { + "video_id": "I0IW5", + "question_id": "I0IW5", + "video_file": "Charades/videos/I0IW5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person walks in eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.4> to <10.3>." + } + ] + }, + { + "video_id": "0NZVB", + "question_id": "0NZVB", + "video_file": "Charades/videos/0NZVB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is drinking from a glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <17.7>." + } + ] + }, + { + "video_id": "0NZVB", + "question_id": "0NZVB", + "video_file": "Charades/videos/0NZVB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person close the door'." + }, + { + "from": "gpt", + "value": "From <36.9> to <40.9>." + } + ] + }, + { + "video_id": "BFW9A", + "question_id": "BFW9A", + "video_file": "Charades/videos/BFW9A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closing the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <14.0>." + } + ] + }, + { + "video_id": "WJNJJ", + "question_id": "WJNJJ", + "video_file": "Charades/videos/WJNJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is watching television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.8> to <12.2>." + } + ] + }, + { + "video_id": "WJNJJ", + "question_id": "WJNJJ", + "video_file": "Charades/videos/WJNJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is standing in a closet holding a doorknob'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <10.6>." + } + ] + }, + { + "video_id": "WJNJJ", + "question_id": "WJNJJ", + "video_file": "Charades/videos/WJNJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they are watching television' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.8> to <12.2>." + } + ] + }, + { + "video_id": "JY6TC", + "question_id": "JY6TC", + "video_file": "Charades/videos/JY6TC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person awakens in their bedroom'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "ZH4JE", + "question_id": "ZH4JE", + "video_file": "Charades/videos/ZH4JE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person reading a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.4>." + } + ] + }, + { + "video_id": "ZH4JE", + "question_id": "ZH4JE", + "video_file": "Charades/videos/ZH4JE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens that book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.9> to <35.0>." + } + ] + }, + { + "video_id": "IJ01G", + "question_id": "IJ01G", + "video_file": "Charades/videos/IJ01G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person opens a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "IJ01G", + "question_id": "IJ01G", + "video_file": "Charades/videos/IJ01G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <0.9> to <5.6>." + } + ] + }, + { + "video_id": "TF6P8", + "question_id": "TF6P8", + "video_file": "Charades/videos/TF6P8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person puts a box in a closet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.8> to <14.0>." + } + ] + }, + { + "video_id": "E5WJJ", + "question_id": "E5WJJ", + "video_file": "Charades/videos/E5WJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting a picture on the wall' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.4> to <28.9>." + } + ] + }, + { + "video_id": "E5WJJ", + "question_id": "E5WJJ", + "video_file": "Charades/videos/E5WJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put their shoes on' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.6> to <41.5>." + } + ] + }, + { + "video_id": "E5WJJ", + "question_id": "E5WJJ", + "video_file": "Charades/videos/E5WJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they run out' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <37.5> to <45.0>." + } + ] + }, + { + "video_id": "E5WJJ", + "question_id": "E5WJJ", + "video_file": "Charades/videos/E5WJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is running toward the sofa'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <37.5> to <45.0>." + } + ] + }, + { + "video_id": "E5WJJ", + "question_id": "E5WJJ", + "video_file": "Charades/videos/E5WJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they sit on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.9> to <41.8>." + } + ] + }, + { + "video_id": "E5WJJ", + "question_id": "E5WJJ", + "video_file": "Charades/videos/E5WJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person running out the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <37.5> to <45.0>." + } + ] + }, + { + "video_id": "E5WJJ", + "question_id": "E5WJJ", + "video_file": "Charades/videos/E5WJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put on their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.6> to <41.5>." + } + ] + }, + { + "video_id": "E5WJJ", + "question_id": "E5WJJ", + "video_file": "Charades/videos/E5WJJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting on some shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <28.6> to <41.5>." + } + ] + }, + { + "video_id": "Z97SD", + "question_id": "Z97SD", + "video_file": "Charades/videos/Z97SD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person walks through a doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <20.1>." + } + ] + }, + { + "video_id": "Z97SD", + "question_id": "Z97SD", + "video_file": "Charades/videos/Z97SD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eats it'." + }, + { + "from": "gpt", + "value": "From <27.4> to <32.0>." + } + ] + }, + { + "video_id": "Z97SD", + "question_id": "Z97SD", + "video_file": "Charades/videos/Z97SD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens the refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.1> to <24.1>." + } + ] + }, + { + "video_id": "5GWS2", + "question_id": "5GWS2", + "video_file": "Charades/videos/5GWS2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person start fixing their hair'." + }, + { + "from": "gpt", + "value": "From <25.2> to <33.0>." + } + ] + }, + { + "video_id": "5GWS2", + "question_id": "5GWS2", + "video_file": "Charades/videos/5GWS2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put down the book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.2> to <24.6>." + } + ] + }, + { + "video_id": "34IUI", + "question_id": "34IUI", + "video_file": "Charades/videos/34IUI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throws back a blanket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.3> to <17.2>." + } + ] + }, + { + "video_id": "GEJPM", + "question_id": "GEJPM", + "video_file": "Charades/videos/GEJPM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <29.1> to <34.3>." + } + ] + }, + { + "video_id": "GEJPM", + "question_id": "GEJPM", + "video_file": "Charades/videos/GEJPM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is opening a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.1> to <7.2>." + } + ] + }, + { + "video_id": "GEJPM", + "question_id": "GEJPM", + "video_file": "Charades/videos/GEJPM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person puts the dishes on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <16.8>." + } + ] + }, + { + "video_id": "BZ4K7", + "question_id": "BZ4K7", + "video_file": "Charades/videos/BZ4K7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'other person with cup in hand drinking' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <12.0>." + } + ] + }, + { + "video_id": "I8LEO", + "question_id": "I8LEO", + "video_file": "Charades/videos/I8LEO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sneezing on a doorknob' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.6> to <20.5>." + } + ] + }, + { + "video_id": "ZI3ED", + "question_id": "ZI3ED", + "video_file": "Charades/videos/ZI3ED.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts a broom down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.2> to <9.7>." + } + ] + }, + { + "video_id": "ZI3ED", + "question_id": "ZI3ED", + "video_file": "Charades/videos/ZI3ED.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person throws the broom aside' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.4> to <10.5>." + } + ] + }, + { + "video_id": "ZI3ED", + "question_id": "ZI3ED", + "video_file": "Charades/videos/ZI3ED.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs their home office carrying a broom'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "ZI3ED", + "question_id": "ZI3ED", + "video_file": "Charades/videos/ZI3ED.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs into a room'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "OR3G4", + "question_id": "OR3G4", + "video_file": "Charades/videos/OR3G4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person grasps a towel to put over a shoulder'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.3>." + } + ] + }, + { + "video_id": "OR3G4", + "question_id": "OR3G4", + "video_file": "Charades/videos/OR3G4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts a towel over their shoulder' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.3>." + } + ] + }, + { + "video_id": "6QSON", + "question_id": "6QSON", + "video_file": "Charades/videos/6QSON.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts it on the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.8> to <18.1>." + } + ] + }, + { + "video_id": "6QSON", + "question_id": "6QSON", + "video_file": "Charades/videos/6QSON.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person open a bag containing groceries' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.5> to <24.3>." + } + ] + }, + { + "video_id": "N8XDV", + "question_id": "N8XDV", + "video_file": "Charades/videos/N8XDV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'this person takes a look out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <16.1>." + } + ] + }, + { + "video_id": "N8XDV", + "question_id": "N8XDV", + "video_file": "Charades/videos/N8XDV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person turns to look out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <16.1>." + } + ] + }, + { + "video_id": "N8XDV", + "question_id": "N8XDV", + "video_file": "Charades/videos/N8XDV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.8> to <7.9>." + } + ] + }, + { + "video_id": "N8XDV", + "question_id": "N8XDV", + "video_file": "Charades/videos/N8XDV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looks out the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <16.1>." + } + ] + }, + { + "video_id": "I9SSV", + "question_id": "I9SSV", + "video_file": "Charades/videos/I9SSV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting on a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.5> to <28.0>." + } + ] + }, + { + "video_id": "I9SSV", + "question_id": "I9SSV", + "video_file": "Charades/videos/I9SSV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts on a pair of shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.2> to <25.8>." + } + ] + }, + { + "video_id": "I9SSV", + "question_id": "I9SSV", + "video_file": "Charades/videos/I9SSV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts their shoes on' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.2> to <25.8>." + } + ] + }, + { + "video_id": "I9SSV", + "question_id": "I9SSV", + "video_file": "Charades/videos/I9SSV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person is sitting on a chair at a desk' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.5> to <28.0>." + } + ] + }, + { + "video_id": "I9SSV", + "question_id": "I9SSV", + "video_file": "Charades/videos/I9SSV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person is sitting in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.5> to <28.0>." + } + ] + }, + { + "video_id": "I9SSV", + "question_id": "I9SSV", + "video_file": "Charades/videos/I9SSV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.8> to <10.8>." + } + ] + }, + { + "video_id": "AM3R2", + "question_id": "AM3R2", + "video_file": "Charades/videos/AM3R2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person throws the pillow down' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.8> to <14.8>." + } + ] + }, + { + "video_id": "AM3R2", + "question_id": "AM3R2", + "video_file": "Charades/videos/AM3R2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person runs around the room with a blanket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.4>." + } + ] + }, + { + "video_id": "AM3R2", + "question_id": "AM3R2", + "video_file": "Charades/videos/AM3R2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs around a room holding a pillow'." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.4>." + } + ] + }, + { + "video_id": "AM3R2", + "question_id": "AM3R2", + "video_file": "Charades/videos/AM3R2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person throwing the pillow on the ground'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <14.8>." + } + ] + }, + { + "video_id": "EYQ6U", + "question_id": "EYQ6U", + "video_file": "Charades/videos/EYQ6U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person is watching television'." + }, + { + "from": "gpt", + "value": "From <78.6> to <99.3>." + } + ] + }, + { + "video_id": "EYQ6U", + "question_id": "EYQ6U", + "video_file": "Charades/videos/EYQ6U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <81.8> to <90.1>." + } + ] + }, + { + "video_id": "EYQ6U", + "question_id": "EYQ6U", + "video_file": "Charades/videos/EYQ6U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person pouring soda into glass' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <32.6> to <43.1>." + } + ] + }, + { + "video_id": "MRSYL", + "question_id": "MRSYL", + "video_file": "Charades/videos/MRSYL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person holding a bag of groceries'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.6>." + } + ] + }, + { + "video_id": "MRSYL", + "question_id": "MRSYL", + "video_file": "Charades/videos/MRSYL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person runs into their study'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.5>." + } + ] + }, + { + "video_id": "FX7ZY", + "question_id": "FX7ZY", + "video_file": "Charades/videos/FX7ZY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person proceeds to throw the box on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <21.0>." + } + ] + }, + { + "video_id": "FX7ZY", + "question_id": "FX7ZY", + "video_file": "Charades/videos/FX7ZY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they begin opening the box'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <7.0>." + } + ] + }, + { + "video_id": "FX7ZY", + "question_id": "FX7ZY", + "video_file": "Charades/videos/FX7ZY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person throwing it to the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <21.0>." + } + ] + }, + { + "video_id": "14A2O", + "question_id": "14A2O", + "video_file": "Charades/videos/14A2O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating one throws a picture on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "14A2O", + "question_id": "14A2O", + "video_file": "Charades/videos/14A2O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person eats' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "14A2O", + "question_id": "14A2O", + "video_file": "Charades/videos/14A2O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person looking at a picture' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.1>." + } + ] + }, + { + "video_id": "14A2O", + "question_id": "14A2O", + "video_file": "Charades/videos/14A2O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person is eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.8>." + } + ] + }, + { + "video_id": "I79Q6", + "question_id": "I79Q6", + "video_file": "Charades/videos/I79Q6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting a box in' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.4> to <8.1>." + } + ] + }, + { + "video_id": "I79Q6", + "question_id": "I79Q6", + "video_file": "Charades/videos/I79Q6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person walks to refrigerator opening it'." + }, + { + "from": "gpt", + "value": "From <1.2> to <8.1>." + } + ] + }, + { + "video_id": "I79Q6", + "question_id": "I79Q6", + "video_file": "Charades/videos/I79Q6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pretends to pour it into cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <21.3>." + } + ] + }, + { + "video_id": "I79Q6", + "question_id": "I79Q6", + "video_file": "Charades/videos/I79Q6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person pour themselves a glass of juice'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.2> to <21.3>." + } + ] + }, + { + "video_id": "I79Q6", + "question_id": "I79Q6", + "video_file": "Charades/videos/I79Q6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the box in the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <8.1>." + } + ] + }, + { + "video_id": "ESTDK", + "question_id": "ESTDK", + "video_file": "Charades/videos/ESTDK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts the box away'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.7> to <33.0>." + } + ] + }, + { + "video_id": "ESTDK", + "question_id": "ESTDK", + "video_file": "Charades/videos/ESTDK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting the box back' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.7> to <33.0>." + } + ] + }, + { + "video_id": "ESTDK", + "question_id": "ESTDK", + "video_file": "Charades/videos/ESTDK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens the cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.2>." + } + ] + }, + { + "video_id": "ESTDK", + "question_id": "ESTDK", + "video_file": "Charades/videos/ESTDK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person opens up the cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.2>." + } + ] + }, + { + "video_id": "ESTDK", + "question_id": "ESTDK", + "video_file": "Charades/videos/ESTDK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the glass down'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.0> to <27.2>." + } + ] + }, + { + "video_id": "D9UO2", + "question_id": "D9UO2", + "video_file": "Charades/videos/D9UO2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person was holding a bag of new clothes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.9> to <10.5>." + } + ] + }, + { + "video_id": "A0CJL", + "question_id": "A0CJL", + "video_file": "Charades/videos/A0CJL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the laptop' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.9> to <21.9>." + } + ] + }, + { + "video_id": "5LH8B", + "question_id": "5LH8B", + "video_file": "Charades/videos/5LH8B.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person smiles as they look into the mirror'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.4> to <10.9>." + } + ] + }, + { + "video_id": "PV1WX", + "question_id": "PV1WX", + "video_file": "Charades/videos/PV1WX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is tidying the clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.1>." + } + ] + }, + { + "video_id": "JZ9L4", + "question_id": "JZ9L4", + "video_file": "Charades/videos/JZ9L4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person begins holding a towel to put away' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <30.6>." + } + ] + }, + { + "video_id": "WJL39", + "question_id": "WJL39", + "video_file": "Charades/videos/WJL39.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person is finishing dressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.0>." + } + ] + }, + { + "video_id": "WJL39", + "question_id": "WJL39", + "video_file": "Charades/videos/WJL39.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opening a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.0> to <29.2>." + } + ] + }, + { + "video_id": "BB5DY", + "question_id": "BB5DY", + "video_file": "Charades/videos/BB5DY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is seen putting their homework on the table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.1> to <20.9>." + } + ] + }, + { + "video_id": "BB5DY", + "question_id": "BB5DY", + "video_file": "Charades/videos/BB5DY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take a blanket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.4> to <25.8>." + } + ] + }, + { + "video_id": "NUSKJ", + "question_id": "NUSKJ", + "video_file": "Charades/videos/NUSKJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person someone walks through their kitchen talking on the phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "NUSKJ", + "question_id": "NUSKJ", + "video_file": "Charades/videos/NUSKJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person talking on the phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "VKLP7", + "question_id": "VKLP7", + "video_file": "Charades/videos/VKLP7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person runs to the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "VKLP7", + "question_id": "VKLP7", + "video_file": "Charades/videos/VKLP7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is running up to the kitchen sink' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.3>." + } + ] + }, + { + "video_id": "7EEF0", + "question_id": "7EEF0", + "video_file": "Charades/videos/7EEF0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person start smiling at a picture that is near by'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.7> to <12.6>." + } + ] + }, + { + "video_id": "2SKOA", + "question_id": "2SKOA", + "video_file": "Charades/videos/2SKOA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts a pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <20.4>." + } + ] + }, + { + "video_id": "2FL0X", + "question_id": "2FL0X", + "video_file": "Charades/videos/2FL0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is running down the hallway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.0> to <6.8>." + } + ] + }, + { + "video_id": "2FL0X", + "question_id": "2FL0X", + "video_file": "Charades/videos/2FL0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.8> to <9.6>." + } + ] + }, + { + "video_id": "2FL0X", + "question_id": "2FL0X", + "video_file": "Charades/videos/2FL0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person runs to the basement' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <6.8>." + } + ] + }, + { + "video_id": "2FL0X", + "question_id": "2FL0X", + "video_file": "Charades/videos/2FL0X.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <9.6>." + } + ] + }, + { + "video_id": "IHLXC", + "question_id": "IHLXC", + "video_file": "Charades/videos/IHLXC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins snuggling up to a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "IHLXC", + "question_id": "IHLXC", + "video_file": "Charades/videos/IHLXC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person lies down on the sofa' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.3> to <7.2>." + } + ] + }, + { + "video_id": "24B2K", + "question_id": "24B2K", + "video_file": "Charades/videos/24B2K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <20.9>." + } + ] + }, + { + "video_id": "24B2K", + "question_id": "24B2K", + "video_file": "Charades/videos/24B2K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <20.9>." + } + ] + }, + { + "video_id": "24B2K", + "question_id": "24B2K", + "video_file": "Charades/videos/24B2K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person snuggles with a blanket'." + }, + { + "from": "gpt", + "value": "From <34.7> to <47.9>." + } + ] + }, + { + "video_id": "24B2K", + "question_id": "24B2K", + "video_file": "Charades/videos/24B2K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person start snuggling up to a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <34.7> to <47.9>." + } + ] + }, + { + "video_id": "MNNDB", + "question_id": "MNNDB", + "video_file": "Charades/videos/MNNDB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.5> to <22.2>." + } + ] + }, + { + "video_id": "MNNDB", + "question_id": "MNNDB", + "video_file": "Charades/videos/MNNDB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they leave holding a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.0> to <23.1>." + } + ] + }, + { + "video_id": "MNNDB", + "question_id": "MNNDB", + "video_file": "Charades/videos/MNNDB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person leave the room holding a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <23.1>." + } + ] + }, + { + "video_id": "8GXNA", + "question_id": "8GXNA", + "video_file": "Charades/videos/8GXNA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <28.0> to <33.0>." + } + ] + }, + { + "video_id": "8GXNA", + "question_id": "8GXNA", + "video_file": "Charades/videos/8GXNA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is holding a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.0>." + } + ] + }, + { + "video_id": "8GXNA", + "question_id": "8GXNA", + "video_file": "Charades/videos/8GXNA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes the clothes out' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.8> to <13.1>." + } + ] + }, + { + "video_id": "J4U86", + "question_id": "J4U86", + "video_file": "Charades/videos/J4U86.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person stops to take some medicine' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.3> to <36.2>." + } + ] + }, + { + "video_id": "J4U86", + "question_id": "J4U86", + "video_file": "Charades/videos/J4U86.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person laughs'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.7> to <27.6>." + } + ] + }, + { + "video_id": "6YZQO", + "question_id": "6YZQO", + "video_file": "Charades/videos/6YZQO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walking to look out a window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.0> to <26.0>." + } + ] + }, + { + "video_id": "6YZQO", + "question_id": "6YZQO", + "video_file": "Charades/videos/6YZQO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the pillow on the bed'." + }, + { + "from": "gpt", + "value": "From <6.0> to <11.0>." + } + ] + }, + { + "video_id": "6YZQO", + "question_id": "6YZQO", + "video_file": "Charades/videos/6YZQO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a picture of the outside with their phone' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.2> to <18.2>." + } + ] + }, + { + "video_id": "6YZQO", + "question_id": "6YZQO", + "video_file": "Charades/videos/6YZQO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holds a phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <15.0>." + } + ] + }, + { + "video_id": "FTG07", + "question_id": "FTG07", + "video_file": "Charades/videos/FTG07.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opened the door'." + }, + { + "from": "gpt", + "value": "From <8.5> to <17.1>." + } + ] + }, + { + "video_id": "FTG07", + "question_id": "FTG07", + "video_file": "Charades/videos/FTG07.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person grabbed dishes opened the door left briefly'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.5> to <17.1>." + } + ] + }, + { + "video_id": "UGF6Q", + "question_id": "UGF6Q", + "video_file": "Charades/videos/UGF6Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person puts the book down on the shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.1> to <18.2>." + } + ] + }, + { + "video_id": "UGF6Q", + "question_id": "UGF6Q", + "video_file": "Charades/videos/UGF6Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person read a book'." + }, + { + "from": "gpt", + "value": "From <9.2> to <14.5>." + } + ] + }, + { + "video_id": "UGF6Q", + "question_id": "UGF6Q", + "video_file": "Charades/videos/UGF6Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put the book in a cupboard'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.1> to <18.2>." + } + ] + }, + { + "video_id": "UGF6Q", + "question_id": "UGF6Q", + "video_file": "Charades/videos/UGF6Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person standing in a room reading a book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.2> to <14.5>." + } + ] + }, + { + "video_id": "UGF6Q", + "question_id": "UGF6Q", + "video_file": "Charades/videos/UGF6Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person take a blanket out'." + }, + { + "from": "gpt", + "value": "From <16.8> to <23.0>." + } + ] + }, + { + "video_id": "EDKXT", + "question_id": "EDKXT", + "video_file": "Charades/videos/EDKXT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they start drinking from a glass of water'." + }, + { + "from": "gpt", + "value": "From <10.6> to <15.0>." + } + ] + }, + { + "video_id": "IM4UD", + "question_id": "IM4UD", + "video_file": "Charades/videos/IM4UD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "8XC4G", + "question_id": "8XC4G", + "video_file": "Charades/videos/8XC4G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person laughing at a book they're reading' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.7> to <8.9>." + } + ] + }, + { + "video_id": "8XC4G", + "question_id": "8XC4G", + "video_file": "Charades/videos/8XC4G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take off blankets' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.1> to <11.8>." + } + ] + }, + { + "video_id": "88TCK", + "question_id": "88TCK", + "video_file": "Charades/videos/88TCK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eat it' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.7> to <32.0>." + } + ] + }, + { + "video_id": "88TCK", + "question_id": "88TCK", + "video_file": "Charades/videos/88TCK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts eating something that is sitting on it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.7> to <32.0>." + } + ] + }, + { + "video_id": "T9BED", + "question_id": "T9BED", + "video_file": "Charades/videos/T9BED.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a closet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.7>." + } + ] + }, + { + "video_id": "T9BED", + "question_id": "T9BED", + "video_file": "Charades/videos/T9BED.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is drinking water out of a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.6> to <22.9>." + } + ] + }, + { + "video_id": "7M1D0", + "question_id": "7M1D0", + "video_file": "Charades/videos/7M1D0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person tried sneezing in the sink' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.2> to <29.0>." + } + ] + }, + { + "video_id": "7M1D0", + "question_id": "7M1D0", + "video_file": "Charades/videos/7M1D0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is putting a towel into a dryer' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.2> to <17.1>." + } + ] + }, + { + "video_id": "099Y1", + "question_id": "099Y1", + "video_file": "Charades/videos/099Y1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.7> to <27.3>." + } + ] + }, + { + "video_id": "099Y1", + "question_id": "099Y1", + "video_file": "Charades/videos/099Y1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opened a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <10.8>." + } + ] + }, + { + "video_id": "5J9HP", + "question_id": "5J9HP", + "video_file": "Charades/videos/5J9HP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person runs into the kitchen' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "5J9HP", + "question_id": "5J9HP", + "video_file": "Charades/videos/5J9HP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins eating it' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.8> to <27.2>." + } + ] + }, + { + "video_id": "5J9HP", + "question_id": "5J9HP", + "video_file": "Charades/videos/5J9HP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person runs into the room'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "5J9HP", + "question_id": "5J9HP", + "video_file": "Charades/videos/5J9HP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes the medicine'." + }, + { + "from": "gpt", + "value": "From <20.6> to <28.1>." + } + ] + }, + { + "video_id": "14JPP", + "question_id": "14JPP", + "video_file": "Charades/videos/14JPP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person undressing in front of a mirror in the bedroom' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.0> to <33.0>." + } + ] + }, + { + "video_id": "7VMDP", + "question_id": "7VMDP", + "video_file": "Charades/videos/7VMDP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts a box on the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.2> to <16.1>." + } + ] + }, + { + "video_id": "A18EX", + "question_id": "A18EX", + "video_file": "Charades/videos/A18EX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person is laughing at a doorknob'." + }, + { + "from": "gpt", + "value": "From <17.3> to <31.0>." + } + ] + }, + { + "video_id": "A18EX", + "question_id": "A18EX", + "video_file": "Charades/videos/A18EX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person #2 stood near the bed smiling at person #1' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.7> to <31.0>." + } + ] + }, + { + "video_id": "T2J3M", + "question_id": "T2J3M", + "video_file": "Charades/videos/T2J3M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts clothes in a wardrobe'." + }, + { + "from": "gpt", + "value": "From <13.1> to <25.4>." + } + ] + }, + { + "video_id": "T2J3M", + "question_id": "T2J3M", + "video_file": "Charades/videos/T2J3M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person putting clothes in the wardrobe' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.1> to <25.4>." + } + ] + }, + { + "video_id": "T2J3M", + "question_id": "T2J3M", + "video_file": "Charades/videos/T2J3M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is opening the door with the doorknob' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.2> to <15.8>." + } + ] + }, + { + "video_id": "T2J3M", + "question_id": "T2J3M", + "video_file": "Charades/videos/T2J3M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens a door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.2> to <15.8>." + } + ] + }, + { + "video_id": "AZC1I", + "question_id": "AZC1I", + "video_file": "Charades/videos/AZC1I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sit in the chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.3> to <32.0>." + } + ] + }, + { + "video_id": "AZC1I", + "question_id": "AZC1I", + "video_file": "Charades/videos/AZC1I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sits on chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.3> to <32.0>." + } + ] + }, + { + "video_id": "U2VK5", + "question_id": "U2VK5", + "video_file": "Charades/videos/U2VK5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person standing by door with groceries starts smiling' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.4> to <27.2>." + } + ] + }, + { + "video_id": "U2VK5", + "question_id": "U2VK5", + "video_file": "Charades/videos/U2VK5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks through a doorway holding a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <24.9>." + } + ] + }, + { + "video_id": "DQGYG", + "question_id": "DQGYG", + "video_file": "Charades/videos/DQGYG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting in a chair smiling' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <16.1>." + } + ] + }, + { + "video_id": "DQGYG", + "question_id": "DQGYG", + "video_file": "Charades/videos/DQGYG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sits on a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <16.1>." + } + ] + }, + { + "video_id": "2LTCY", + "question_id": "2LTCY", + "video_file": "Charades/videos/2LTCY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a smiling person is grasping their falling towel being worn'." + }, + { + "from": "gpt", + "value": "From <14.2> to <22.2>." + } + ] + }, + { + "video_id": "2LTCY", + "question_id": "2LTCY", + "video_file": "Charades/videos/2LTCY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person grabs phone takes selfie pictures' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.6> to <23.1>." + } + ] + }, + { + "video_id": "2LTCY", + "question_id": "2LTCY", + "video_file": "Charades/videos/2LTCY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they take their phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.7> to <13.2>." + } + ] + }, + { + "video_id": "069GJ", + "question_id": "069GJ", + "video_file": "Charades/videos/069GJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is smiling' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.2> to <25.0>." + } + ] + }, + { + "video_id": "0BLSL", + "question_id": "0BLSL", + "video_file": "Charades/videos/0BLSL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person go back to cooking' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "0BLSL", + "question_id": "0BLSL", + "video_file": "Charades/videos/0BLSL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person stands at the stove cooking some food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "0BLSL", + "question_id": "0BLSL", + "video_file": "Charades/videos/0BLSL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is cooking at a stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "2U0QT", + "question_id": "2U0QT", + "video_file": "Charades/videos/2U0QT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they close the cabinet' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.4> to <25.1>." + } + ] + }, + { + "video_id": "2U0QT", + "question_id": "2U0QT", + "video_file": "Charades/videos/2U0QT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person watch some television' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.7> to <17.1>." + } + ] + }, + { + "video_id": "2U0QT", + "question_id": "2U0QT", + "video_file": "Charades/videos/2U0QT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens a cabinet next to a television' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <7.2>." + } + ] + }, + { + "video_id": "2U0QT", + "question_id": "2U0QT", + "video_file": "Charades/videos/2U0QT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they take a glass of water' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.2> to <11.9>." + } + ] + }, + { + "video_id": "2U0QT", + "question_id": "2U0QT", + "video_file": "Charades/videos/2U0QT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person close the cabinet door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.4> to <25.1>." + } + ] + }, + { + "video_id": "2U0QT", + "question_id": "2U0QT", + "video_file": "Charades/videos/2U0QT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <20.4> to <25.1>." + } + ] + }, + { + "video_id": "2U0QT", + "question_id": "2U0QT", + "video_file": "Charades/videos/2U0QT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opens the cabinet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.3> to <7.2>." + } + ] + }, + { + "video_id": "2U0QT", + "question_id": "2U0QT", + "video_file": "Charades/videos/2U0QT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person watching television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.7> to <17.1>." + } + ] + }, + { + "video_id": "2U0QT", + "question_id": "2U0QT", + "video_file": "Charades/videos/2U0QT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.2> to <11.9>." + } + ] + }, + { + "video_id": "39TDM", + "question_id": "39TDM", + "video_file": "Charades/videos/39TDM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begins sneezing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.3> to <33.0>." + } + ] + }, + { + "video_id": "OK2AF", + "question_id": "OK2AF", + "video_file": "Charades/videos/OK2AF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person watches television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.1>." + } + ] + }, + { + "video_id": "OK2AF", + "question_id": "OK2AF", + "video_file": "Charades/videos/OK2AF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting down watching television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.1>." + } + ] + }, + { + "video_id": "OK2AF", + "question_id": "OK2AF", + "video_file": "Charades/videos/OK2AF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person the door opens' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.0> to <20.0>." + } + ] + }, + { + "video_id": "OK2AF", + "question_id": "OK2AF", + "video_file": "Charades/videos/OK2AF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person opens a door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.0> to <20.0>." + } + ] + }, + { + "video_id": "OK2AF", + "question_id": "OK2AF", + "video_file": "Charades/videos/OK2AF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting in a chair watching television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.1>." + } + ] + }, + { + "video_id": "OK2AF", + "question_id": "OK2AF", + "video_file": "Charades/videos/OK2AF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes a book out from a chair'." + }, + { + "from": "gpt", + "value": "From <23.0> to <30.0>." + } + ] + }, + { + "video_id": "WRFZW", + "question_id": "WRFZW", + "video_file": "Charades/videos/WRFZW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person pours water into a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <14.6>." + } + ] + }, + { + "video_id": "WRFZW", + "question_id": "WRFZW", + "video_file": "Charades/videos/WRFZW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pours some water in a cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.3> to <14.6>." + } + ] + }, + { + "video_id": "QY75H", + "question_id": "QY75H", + "video_file": "Charades/videos/QY75H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person proceeds to pour a cup of coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.8> to <21.9>." + } + ] + }, + { + "video_id": "RHVM2", + "question_id": "RHVM2", + "video_file": "Charades/videos/RHVM2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person down to look out the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <16.7>." + } + ] + }, + { + "video_id": "RHVM2", + "question_id": "RHVM2", + "video_file": "Charades/videos/RHVM2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person keeps watching out the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.0> to <16.7>." + } + ] + }, + { + "video_id": "RHVM2", + "question_id": "RHVM2", + "video_file": "Charades/videos/RHVM2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is playing on their phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "RHVM2", + "question_id": "RHVM2", + "video_file": "Charades/videos/RHVM2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person looks out a window twice'." + }, + { + "from": "gpt", + "value": "From <7.0> to <16.7>." + } + ] + }, + { + "video_id": "1Y09V", + "question_id": "1Y09V", + "video_file": "Charades/videos/1Y09V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking coffee from a cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.5> to <10.3>." + } + ] + }, + { + "video_id": "1Y09V", + "question_id": "1Y09V", + "video_file": "Charades/videos/1Y09V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person drinks from a cup of coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.5> to <10.3>." + } + ] + }, + { + "video_id": "0M1X0", + "question_id": "0M1X0", + "video_file": "Charades/videos/0M1X0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person starts to undress'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.1>." + } + ] + }, + { + "video_id": "0M1X0", + "question_id": "0M1X0", + "video_file": "Charades/videos/0M1X0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put their clothes on a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.1> to <14.6>." + } + ] + }, + { + "video_id": "0M1X0", + "question_id": "0M1X0", + "video_file": "Charades/videos/0M1X0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is undressing by a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.1>." + } + ] + }, + { + "video_id": "IKQLJ", + "question_id": "IKQLJ", + "video_file": "Charades/videos/IKQLJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drinks a cup of coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <47.4> to <53.6>." + } + ] + }, + { + "video_id": "DPCGS", + "question_id": "DPCGS", + "video_file": "Charades/videos/DPCGS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eat their sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.4> to <9.8>." + } + ] + }, + { + "video_id": "GO761", + "question_id": "GO761", + "video_file": "Charades/videos/GO761.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns off the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <14.3> to <19.0>." + } + ] + }, + { + "video_id": "2G59W", + "question_id": "2G59W", + "video_file": "Charades/videos/2G59W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they put their phone away' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.8> to <8.6>." + } + ] + }, + { + "video_id": "2G59W", + "question_id": "2G59W", + "video_file": "Charades/videos/2G59W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts the cellphone down' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <8.6>." + } + ] + }, + { + "video_id": "NU0UK", + "question_id": "NU0UK", + "video_file": "Charades/videos/NU0UK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sitting at a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.1>." + } + ] + }, + { + "video_id": "NU0UK", + "question_id": "NU0UK", + "video_file": "Charades/videos/NU0UK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person is sitting at a table drinking coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.1>." + } + ] + }, + { + "video_id": "NU0UK", + "question_id": "NU0UK", + "video_file": "Charades/videos/NU0UK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the other person is sitting at the table drinking coffee' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.1>." + } + ] + }, + { + "video_id": "HETLN", + "question_id": "HETLN", + "video_file": "Charades/videos/HETLN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person is smiling as they drink from the cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.5> to <13.0>." + } + ] + }, + { + "video_id": "JIGAJ", + "question_id": "JIGAJ", + "video_file": "Charades/videos/JIGAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <27.7> to <34.0>." + } + ] + }, + { + "video_id": "JIGAJ", + "question_id": "JIGAJ", + "video_file": "Charades/videos/JIGAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.7> to <34.0>." + } + ] + }, + { + "video_id": "JIGAJ", + "question_id": "JIGAJ", + "video_file": "Charades/videos/JIGAJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turning off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <26.2> to <32.0>." + } + ] + }, + { + "video_id": "WTJQ0", + "question_id": "WTJQ0", + "video_file": "Charades/videos/WTJQ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person opens the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.9> to <12.8>." + } + ] + }, + { + "video_id": "WTJQ0", + "question_id": "WTJQ0", + "video_file": "Charades/videos/WTJQ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts a plate of food on a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <7.3>." + } + ] + }, + { + "video_id": "WTJQ0", + "question_id": "WTJQ0", + "video_file": "Charades/videos/WTJQ0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is putting food on a coffee table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.1> to <7.3>." + } + ] + }, + { + "video_id": "SMLI0", + "question_id": "SMLI0", + "video_file": "Charades/videos/SMLI0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they put the picture down'." + }, + { + "from": "gpt", + "value": "From <2.7> to <8.3>." + } + ] + }, + { + "video_id": "Y6MUU", + "question_id": "Y6MUU", + "video_file": "Charades/videos/Y6MUU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.5> to <37.6>." + } + ] + }, + { + "video_id": "LOSLI", + "question_id": "LOSLI", + "video_file": "Charades/videos/LOSLI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person starts sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <15.5>." + } + ] + }, + { + "video_id": "NLU6W", + "question_id": "NLU6W", + "video_file": "Charades/videos/NLU6W.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is holding a pair of shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.2> to <14.5>." + } + ] + }, + { + "video_id": "I0THD", + "question_id": "I0THD", + "video_file": "Charades/videos/I0THD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sneezing as the turn the doorknob' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.1> to <32.0>." + } + ] + }, + { + "video_id": "AXYF9", + "question_id": "AXYF9", + "video_file": "Charades/videos/AXYF9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person one is eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <2.9> to <13.7>." + } + ] + }, + { + "video_id": "AXYF9", + "question_id": "AXYF9", + "video_file": "Charades/videos/AXYF9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person is eating a sandwich'." + }, + { + "from": "gpt", + "value": "From <2.9> to <13.7>." + } + ] + }, + { + "video_id": "9KGOL", + "question_id": "9KGOL", + "video_file": "Charades/videos/9KGOL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they take a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.8> to <24.7>." + } + ] + }, + { + "video_id": "UMXKN", + "question_id": "UMXKN", + "video_file": "Charades/videos/UMXKN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put the clothes in the hamper' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.2> to <22.3>." + } + ] + }, + { + "video_id": "G0BKH", + "question_id": "G0BKH", + "video_file": "Charades/videos/G0BKH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turn off the light in their closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.8> to <24.3>." + } + ] + }, + { + "video_id": "G0BKH", + "question_id": "G0BKH", + "video_file": "Charades/videos/G0BKH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put a pillow into it'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.9> to <30.7>." + } + ] + }, + { + "video_id": "G0BKH", + "question_id": "G0BKH", + "video_file": "Charades/videos/G0BKH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns off the closet light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.8> to <24.3>." + } + ] + }, + { + "video_id": "SY1JE", + "question_id": "SY1JE", + "video_file": "Charades/videos/SY1JE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person walks out of the room through a doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.5> to <17.2>." + } + ] + }, + { + "video_id": "SAFT3", + "question_id": "SAFT3", + "video_file": "Charades/videos/SAFT3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person sits in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <10.0>." + } + ] + }, + { + "video_id": "SAFT3", + "question_id": "SAFT3", + "video_file": "Charades/videos/SAFT3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they sit down in the chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <10.0>." + } + ] + }, + { + "video_id": "T832F", + "question_id": "T832F", + "video_file": "Charades/videos/T832F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person is seen throwing the folded blanket on the bed'." + }, + { + "from": "gpt", + "value": "From <6.1> to <10.8>." + } + ] + }, + { + "video_id": "QXI52", + "question_id": "QXI52", + "video_file": "Charades/videos/QXI52.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person watching television' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "OUH2U", + "question_id": "OUH2U", + "video_file": "Charades/videos/OUH2U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.3> to <16.7>." + } + ] + }, + { + "video_id": "OUH2U", + "question_id": "OUH2U", + "video_file": "Charades/videos/OUH2U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door behind them'." + }, + { + "from": "gpt", + "value": "From <8.3> to <16.7>." + } + ] + }, + { + "video_id": "42MC3", + "question_id": "42MC3", + "video_file": "Charades/videos/42MC3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eats a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.5> to <25.0>." + } + ] + }, + { + "video_id": "42MC3", + "question_id": "42MC3", + "video_file": "Charades/videos/42MC3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes a door repeatedly' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.2>." + } + ] + }, + { + "video_id": "42MC3", + "question_id": "42MC3", + "video_file": "Charades/videos/42MC3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person off repeatedly' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.5> to <25.0>." + } + ] + }, + { + "video_id": "42MC3", + "question_id": "42MC3", + "video_file": "Charades/videos/42MC3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turning a light off' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.2> to <16.0>." + } + ] + }, + { + "video_id": "42MC3", + "question_id": "42MC3", + "video_file": "Charades/videos/42MC3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns a light switch on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.5> to <16.8>." + } + ] + }, + { + "video_id": "42MC3", + "question_id": "42MC3", + "video_file": "Charades/videos/42MC3.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closing a door'." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.2>." + } + ] + }, + { + "video_id": "IST7I", + "question_id": "IST7I", + "video_file": "Charades/videos/IST7I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating food out of box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.4> to <32.8>." + } + ] + }, + { + "video_id": "IST7I", + "question_id": "IST7I", + "video_file": "Charades/videos/IST7I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats some food'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.4> to <32.8>." + } + ] + }, + { + "video_id": "IST7I", + "question_id": "IST7I", + "video_file": "Charades/videos/IST7I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person begin eating the crackers' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <18.4> to <32.8>." + } + ] + }, + { + "video_id": "IST7I", + "question_id": "IST7I", + "video_file": "Charades/videos/IST7I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they open a box of ritz crackers'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.2>." + } + ] + }, + { + "video_id": "BZ5AU", + "question_id": "BZ5AU", + "video_file": "Charades/videos/BZ5AU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is dressing in front of the mirror' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.2> to <16.0>." + } + ] + }, + { + "video_id": "RV11J", + "question_id": "RV11J", + "video_file": "Charades/videos/RV11J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts their camera away into a bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.1> to <41.0>." + } + ] + }, + { + "video_id": "RV11J", + "question_id": "RV11J", + "video_file": "Charades/videos/RV11J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person put the camera into it's bag'." + }, + { + "from": "gpt", + "value": "From <33.1> to <41.0>." + } + ] + }, + { + "video_id": "2RU6J", + "question_id": "2RU6J", + "video_file": "Charades/videos/2RU6J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'then person a lays down on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <25.0>." + } + ] + }, + { + "video_id": "07LG5", + "question_id": "07LG5", + "video_file": "Charades/videos/07LG5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is running' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <13.3>." + } + ] + }, + { + "video_id": "07LG5", + "question_id": "07LG5", + "video_file": "Charades/videos/07LG5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the fridge door'." + }, + { + "from": "gpt", + "value": "From <7.3> to <21.1>." + } + ] + }, + { + "video_id": "07LG5", + "question_id": "07LG5", + "video_file": "Charades/videos/07LG5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person walks to a closet closing the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <21.1>." + } + ] + }, + { + "video_id": "07LG5", + "question_id": "07LG5", + "video_file": "Charades/videos/07LG5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the cabinet door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <7.3> to <21.1>." + } + ] + }, + { + "video_id": "TS2TD", + "question_id": "TS2TD", + "video_file": "Charades/videos/TS2TD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person open the first book to start'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.4> to <35.0>." + } + ] + }, + { + "video_id": "TS2TD", + "question_id": "TS2TD", + "video_file": "Charades/videos/TS2TD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person read a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.9> to <35.0>." + } + ] + }, + { + "video_id": "0RJ18", + "question_id": "0RJ18", + "video_file": "Charades/videos/0RJ18.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting it on a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <15.8>." + } + ] + }, + { + "video_id": "S6W5T", + "question_id": "S6W5T", + "video_file": "Charades/videos/S6W5T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person takes off their shoes'." + }, + { + "from": "gpt", + "value": "From <17.2> to <27.0>." + } + ] + }, + { + "video_id": "LI5VF", + "question_id": "LI5VF", + "video_file": "Charades/videos/LI5VF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person tidy's some hanging clothes'." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "LI5VF", + "question_id": "LI5VF", + "video_file": "Charades/videos/LI5VF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts bag with pillow in sink' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.7> to <19.3>." + } + ] + }, + { + "video_id": "LI5VF", + "question_id": "LI5VF", + "video_file": "Charades/videos/LI5VF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.1> to <28.2>." + } + ] + }, + { + "video_id": "SG99Q", + "question_id": "SG99Q", + "video_file": "Charades/videos/SG99Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits down in a chair'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <31.0>." + } + ] + }, + { + "video_id": "SG99Q", + "question_id": "SG99Q", + "video_file": "Charades/videos/SG99Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sitting down in a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <31.0>." + } + ] + }, + { + "video_id": "N0KQ7", + "question_id": "N0KQ7", + "video_file": "Charades/videos/N0KQ7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put on their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.9> to <27.0>." + } + ] + }, + { + "video_id": "Z4Y02", + "question_id": "Z4Y02", + "video_file": "Charades/videos/Z4Y02.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they close the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.3> to <18.2>." + } + ] + }, + { + "video_id": "01O27", + "question_id": "01O27", + "video_file": "Charades/videos/01O27.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person but the door won't open'." + }, + { + "from": "gpt", + "value": "From <0.5> to <6.6>." + } + ] + }, + { + "video_id": "01O27", + "question_id": "01O27", + "video_file": "Charades/videos/01O27.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door behind them'." + }, + { + "from": "gpt", + "value": "From <5.1> to <11.5>." + } + ] + }, + { + "video_id": "01O27", + "question_id": "01O27", + "video_file": "Charades/videos/01O27.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens the bathroom door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.5> to <6.6>." + } + ] + }, + { + "video_id": "C3JAS", + "question_id": "C3JAS", + "video_file": "Charades/videos/C3JAS.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person start sneezing' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.6> to <22.3>." + } + ] + }, + { + "video_id": "JRSPH", + "question_id": "JRSPH", + "video_file": "Charades/videos/JRSPH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person sitting in a chair takes off their shoes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.2>." + } + ] + }, + { + "video_id": "1WWJZ", + "question_id": "1WWJZ", + "video_file": "Charades/videos/1WWJZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person drink out of a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.7> to <15.7>." + } + ] + }, + { + "video_id": "AI6KJ", + "question_id": "AI6KJ", + "video_file": "Charades/videos/AI6KJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts their bag down'." + }, + { + "from": "gpt", + "value": "From <5.0> to <12.3>." + } + ] + }, + { + "video_id": "AI6KJ", + "question_id": "AI6KJ", + "video_file": "Charades/videos/AI6KJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closing the door behind them' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.6> to <18.4>." + } + ] + }, + { + "video_id": "AI6KJ", + "question_id": "AI6KJ", + "video_file": "Charades/videos/AI6KJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person puts a bag onto the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.2> to <12.3>." + } + ] + }, + { + "video_id": "KENG1", + "question_id": "KENG1", + "video_file": "Charades/videos/KENG1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person pours a glass of water'." + }, + { + "from": "gpt", + "value": "From <29.3> to <38.0>." + } + ] + }, + { + "video_id": "KENG1", + "question_id": "KENG1", + "video_file": "Charades/videos/KENG1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person puts on their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <12.1>." + } + ] + }, + { + "video_id": "KENG1", + "question_id": "KENG1", + "video_file": "Charades/videos/KENG1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts the shoes on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.3> to <12.1>." + } + ] + }, + { + "video_id": "3H09L", + "question_id": "3H09L", + "video_file": "Charades/videos/3H09L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <24.4>." + } + ] + }, + { + "video_id": "3H09L", + "question_id": "3H09L", + "video_file": "Charades/videos/3H09L.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person one holds a pair of shoes'." + }, + { + "from": "gpt", + "value": "From <2.9> to <15.0>." + } + ] + }, + { + "video_id": "UMOE4", + "question_id": "UMOE4", + "video_file": "Charades/videos/UMOE4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person suddenly comes running through'." + }, + { + "from": "gpt", + "value": "From <6.2> to <11.7>." + } + ] + }, + { + "video_id": "UMOE4", + "question_id": "UMOE4", + "video_file": "Charades/videos/UMOE4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'another person runs across the room' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.2> to <11.7>." + } + ] + }, + { + "video_id": "IK8PA", + "question_id": "IK8PA", + "video_file": "Charades/videos/IK8PA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts it on a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.2> to <31.7>." + } + ] + }, + { + "video_id": "IK8PA", + "question_id": "IK8PA", + "video_file": "Charades/videos/IK8PA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sneezing' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <21.6>." + } + ] + }, + { + "video_id": "ZCOSY", + "question_id": "ZCOSY", + "video_file": "Charades/videos/ZCOSY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turn the light on' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <31.0>." + } + ] + }, + { + "video_id": "ZCOSY", + "question_id": "ZCOSY", + "video_file": "Charades/videos/ZCOSY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person fixing a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.3> to <31.0>." + } + ] + }, + { + "video_id": "ZCOSY", + "question_id": "ZCOSY", + "video_file": "Charades/videos/ZCOSY.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person turns on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.3> to <31.0>." + } + ] + }, + { + "video_id": "FTWBB", + "question_id": "FTWBB", + "video_file": "Charades/videos/FTWBB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sitting on a chair laughing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "FTWBB", + "question_id": "FTWBB", + "video_file": "Charades/videos/FTWBB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'laughs as another person plays' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.1>." + } + ] + }, + { + "video_id": "VWN13", + "question_id": "VWN13", + "video_file": "Charades/videos/VWN13.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holds their phone as they grab the closet doorknob' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <14.0>." + } + ] + }, + { + "video_id": "VWN13", + "question_id": "VWN13", + "video_file": "Charades/videos/VWN13.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person open/close the door as if playing hide'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.5> to <14.0>." + } + ] + }, + { + "video_id": "CNSMI", + "question_id": "CNSMI", + "video_file": "Charades/videos/CNSMI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <30.8> to <37.7>." + } + ] + }, + { + "video_id": "PIJCA", + "question_id": "PIJCA", + "video_file": "Charades/videos/PIJCA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.1> to <18.2>." + } + ] + }, + { + "video_id": "J4VCR", + "question_id": "J4VCR", + "video_file": "Charades/videos/J4VCR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts eating them'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <21.5>." + } + ] + }, + { + "video_id": "J4VCR", + "question_id": "J4VCR", + "video_file": "Charades/videos/J4VCR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.4> to <21.5>." + } + ] + }, + { + "video_id": "NDLY5", + "question_id": "NDLY5", + "video_file": "Charades/videos/NDLY5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person snuggling a pillow close to their chest'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.1>." + } + ] + }, + { + "video_id": "NDLY5", + "question_id": "NDLY5", + "video_file": "Charades/videos/NDLY5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begin undressing out of their pajamas'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.3> to <29.7>." + } + ] + }, + { + "video_id": "JO6GB", + "question_id": "JO6GB", + "video_file": "Charades/videos/JO6GB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person put it in a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <39.1> to <43.0>." + } + ] + }, + { + "video_id": "JO6GB", + "question_id": "JO6GB", + "video_file": "Charades/videos/JO6GB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open their bag' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <37.6> to <43.0>." + } + ] + }, + { + "video_id": "B17UQ", + "question_id": "B17UQ", + "video_file": "Charades/videos/B17UQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sitting on a sofa watching television' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.0> to <11.8>." + } + ] + }, + { + "video_id": "B17UQ", + "question_id": "B17UQ", + "video_file": "Charades/videos/B17UQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the remote on the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.3> to <12.3>." + } + ] + }, + { + "video_id": "CUB69", + "question_id": "CUB69", + "video_file": "Charades/videos/CUB69.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sitting in a chair eating food out a plate' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "CUB69", + "question_id": "CUB69", + "video_file": "Charades/videos/CUB69.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is eating at a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "NAWO1", + "question_id": "NAWO1", + "video_file": "Charades/videos/NAWO1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person is holding their phone' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <10.2>." + } + ] + }, + { + "video_id": "M7Y6V", + "question_id": "M7Y6V", + "video_file": "Charades/videos/M7Y6V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person eating food is watching something outside the window' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.9> to <33.0>." + } + ] + }, + { + "video_id": "M7Y6V", + "question_id": "M7Y6V", + "video_file": "Charades/videos/M7Y6V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating the food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.9> to <33.0>." + } + ] + }, + { + "video_id": "M7Y6V", + "question_id": "M7Y6V", + "video_file": "Charades/videos/M7Y6V.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person was eating the food that was in hand' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.9> to <33.0>." + } + ] + }, + { + "video_id": "I8BC5", + "question_id": "I8BC5", + "video_file": "Charades/videos/I8BC5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks into the entryway eating a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.7> to <16.7>." + } + ] + }, + { + "video_id": "I8BC5", + "question_id": "I8BC5", + "video_file": "Charades/videos/I8BC5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person walks in eating something' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <17.0>." + } + ] + }, + { + "video_id": "ZLRFB", + "question_id": "ZLRFB", + "video_file": "Charades/videos/ZLRFB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins to dress' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.0> to <37.0>." + } + ] + }, + { + "video_id": "ZLRFB", + "question_id": "ZLRFB", + "video_file": "Charades/videos/ZLRFB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts getting dressed'." + }, + { + "from": "gpt", + "value": "From <21.0> to <37.0>." + } + ] + }, + { + "video_id": "ZLRFB", + "question_id": "ZLRFB", + "video_file": "Charades/videos/ZLRFB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person watching television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <20.5>." + } + ] + }, + { + "video_id": "ZLRFB", + "question_id": "ZLRFB", + "video_file": "Charades/videos/ZLRFB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person laughs at something on the tv' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.0> to <17.2>." + } + ] + }, + { + "video_id": "8V97G", + "question_id": "8V97G", + "video_file": "Charades/videos/8V97G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opens a nearby door'." + }, + { + "from": "gpt", + "value": "From <17.9> to <26.0>." + } + ] + }, + { + "video_id": "BZ0YN", + "question_id": "BZ0YN", + "video_file": "Charades/videos/BZ0YN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns to start cooking at the stove' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.2> to <31.0>." + } + ] + }, + { + "video_id": "P9R08", + "question_id": "P9R08", + "video_file": "Charades/videos/P9R08.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person fixes hair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.0> to <18.0>." + } + ] + }, + { + "video_id": "P9R08", + "question_id": "P9R08", + "video_file": "Charades/videos/P9R08.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person holding a picture looks into a mirror' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.4> to <32.0>." + } + ] + }, + { + "video_id": "P9R08", + "question_id": "P9R08", + "video_file": "Charades/videos/P9R08.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put the picture on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.8> to <22.1>." + } + ] + }, + { + "video_id": "YUOVC", + "question_id": "YUOVC", + "video_file": "Charades/videos/YUOVC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person gets dressed' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.8> to <18.8>." + } + ] + }, + { + "video_id": "YUOVC", + "question_id": "YUOVC", + "video_file": "Charades/videos/YUOVC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person is frantically dressing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.8> to <18.8>." + } + ] + }, + { + "video_id": "YUOVC", + "question_id": "YUOVC", + "video_file": "Charades/videos/YUOVC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes some clothes off the floor' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.9> to <7.7>." + } + ] + }, + { + "video_id": "YUOVC", + "question_id": "YUOVC", + "video_file": "Charades/videos/YUOVC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person redress in another item'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.8> to <18.8>." + } + ] + }, + { + "video_id": "YUOVC", + "question_id": "YUOVC", + "video_file": "Charades/videos/YUOVC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they undress' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.8> to <18.8>." + } + ] + }, + { + "video_id": "RJDH2", + "question_id": "RJDH2", + "video_file": "Charades/videos/RJDH2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person throws the pillow against the wall' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.0> to <26.6>." + } + ] + }, + { + "video_id": "3EDV7", + "question_id": "3EDV7", + "video_file": "Charades/videos/3EDV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person tosses clothes into their washer'." + }, + { + "from": "gpt", + "value": "From <2.6> to <10.4>." + } + ] + }, + { + "video_id": "3EDV7", + "question_id": "3EDV7", + "video_file": "Charades/videos/3EDV7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person puts some clothes in the washing machine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.6> to <10.4>." + } + ] + }, + { + "video_id": "JOLLV", + "question_id": "JOLLV", + "video_file": "Charades/videos/JOLLV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is snuggling under a blanket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "EKPAQ", + "question_id": "EKPAQ", + "video_file": "Charades/videos/EKPAQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is sneezing by a shelf' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.0>." + } + ] + }, + { + "video_id": "NSUHV", + "question_id": "NSUHV", + "video_file": "Charades/videos/NSUHV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts a sandwich into a bag on a table' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.4> to <17.8>." + } + ] + }, + { + "video_id": "NSUHV", + "question_id": "NSUHV", + "video_file": "Charades/videos/NSUHV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take a bite of sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.0> to <15.7>." + } + ] + }, + { + "video_id": "52FNN", + "question_id": "52FNN", + "video_file": "Charades/videos/52FNN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pours water into a cup' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.1> to <21.9>." + } + ] + }, + { + "video_id": "52FNN", + "question_id": "52FNN", + "video_file": "Charades/videos/52FNN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is holding a sandwich they' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <24.6> to <29.9>." + } + ] + }, + { + "video_id": "52FNN", + "question_id": "52FNN", + "video_file": "Charades/videos/52FNN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walks through a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.7> to <33.0>." + } + ] + }, + { + "video_id": "52FNN", + "question_id": "52FNN", + "video_file": "Charades/videos/52FNN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person puts a sandwich on a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.4> to <6.8>." + } + ] + }, + { + "video_id": "H8BTR", + "question_id": "H8BTR", + "video_file": "Charades/videos/H8BTR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person pours water into a cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.4> to <23.8>." + } + ] + }, + { + "video_id": "H8BTR", + "question_id": "H8BTR", + "video_file": "Charades/videos/H8BTR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person takes a blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <8.0>." + } + ] + }, + { + "video_id": "H8BTR", + "question_id": "H8BTR", + "video_file": "Charades/videos/H8BTR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'another person is drinking a glass of water' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.4> to <31.0>." + } + ] + }, + { + "video_id": "WQQ4Q", + "question_id": "WQQ4Q", + "video_file": "Charades/videos/WQQ4Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take some food out of the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.8> to <20.8>." + } + ] + }, + { + "video_id": "WQQ4Q", + "question_id": "WQQ4Q", + "video_file": "Charades/videos/WQQ4Q.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person washes some dishes in a sink' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.8>." + } + ] + }, + { + "video_id": "5RY3S", + "question_id": "5RY3S", + "video_file": "Charades/videos/5RY3S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person holds a sandwich' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.6> to <15.7>." + } + ] + }, + { + "video_id": "YT9AL", + "question_id": "YT9AL", + "video_file": "Charades/videos/YT9AL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts on shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.9> to <15.6>." + } + ] + }, + { + "video_id": "YT9AL", + "question_id": "YT9AL", + "video_file": "Charades/videos/YT9AL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person laughing on their phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <46.1> to <53.8>." + } + ] + }, + { + "video_id": "YT9AL", + "question_id": "YT9AL", + "video_file": "Charades/videos/YT9AL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person talking on the phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.7> to <29.2>." + } + ] + }, + { + "video_id": "YT9AL", + "question_id": "YT9AL", + "video_file": "Charades/videos/YT9AL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they are smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <41.3> to <47.6>." + } + ] + }, + { + "video_id": "YT9AL", + "question_id": "YT9AL", + "video_file": "Charades/videos/YT9AL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'there the person puts on shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.9> to <15.6>." + } + ] + }, + { + "video_id": "VZY0C", + "question_id": "VZY0C", + "video_file": "Charades/videos/VZY0C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person running to a bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.0> to <9.0>." + } + ] + }, + { + "video_id": "VZY0C", + "question_id": "VZY0C", + "video_file": "Charades/videos/VZY0C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person was running to go get the groceries' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.0> to <9.0>." + } + ] + }, + { + "video_id": "VZY0C", + "question_id": "VZY0C", + "video_file": "Charades/videos/VZY0C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person runs into the bedroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <9.0>." + } + ] + }, + { + "video_id": "7ELBG", + "question_id": "7ELBG", + "video_file": "Charades/videos/7ELBG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person takes some picture' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.6> to <31.2>." + } + ] + }, + { + "video_id": "7ELBG", + "question_id": "7ELBG", + "video_file": "Charades/videos/7ELBG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person a flashlight from the box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.1> to <31.0>." + } + ] + }, + { + "video_id": "7ELBG", + "question_id": "7ELBG", + "video_file": "Charades/videos/7ELBG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pulls out a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.1> to <31.0>." + } + ] + }, + { + "video_id": "7ELBG", + "question_id": "7ELBG", + "video_file": "Charades/videos/7ELBG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they turn on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.1> to <31.0>." + } + ] + }, + { + "video_id": "7ELBG", + "question_id": "7ELBG", + "video_file": "Charades/videos/7ELBG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts a box on the table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <18.4>." + } + ] + }, + { + "video_id": "XKPLB", + "question_id": "XKPLB", + "video_file": "Charades/videos/XKPLB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person quickly open the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <5.2>." + } + ] + }, + { + "video_id": "XKPLB", + "question_id": "XKPLB", + "video_file": "Charades/videos/XKPLB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts running' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.1> to <23.2>." + } + ] + }, + { + "video_id": "CH0PM", + "question_id": "CH0PM", + "video_file": "Charades/videos/CH0PM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they turn the light off' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.3> to <9.9>." + } + ] + }, + { + "video_id": "CH0PM", + "question_id": "CH0PM", + "video_file": "Charades/videos/CH0PM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they take their shoes off'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.9> to <30.3>." + } + ] + }, + { + "video_id": "CH0PM", + "question_id": "CH0PM", + "video_file": "Charades/videos/CH0PM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <9.3>." + } + ] + }, + { + "video_id": "CH0PM", + "question_id": "CH0PM", + "video_file": "Charades/videos/CH0PM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they open the closet door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <9.3>." + } + ] + }, + { + "video_id": "CH0PM", + "question_id": "CH0PM", + "video_file": "Charades/videos/CH0PM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person begin undressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.8> to <24.0>." + } + ] + }, + { + "video_id": "CH0PM", + "question_id": "CH0PM", + "video_file": "Charades/videos/CH0PM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is undressing in the closet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <8.8> to <24.0>." + } + ] + }, + { + "video_id": "7Z69Y", + "question_id": "7Z69Y", + "video_file": "Charades/videos/7Z69Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person hold a bag in a closet'." + }, + { + "from": "gpt", + "value": "From <20.9> to <29.7>." + } + ] + }, + { + "video_id": "7Z69Y", + "question_id": "7Z69Y", + "video_file": "Charades/videos/7Z69Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person gets dressed in a closet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <38.5> to <58.0>." + } + ] + }, + { + "video_id": "7Z69Y", + "question_id": "7Z69Y", + "video_file": "Charades/videos/7Z69Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person starts dressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <38.5> to <58.0>." + } + ] + }, + { + "video_id": "7Z69Y", + "question_id": "7Z69Y", + "video_file": "Charades/videos/7Z69Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sit down on a chair'." + }, + { + "from": "gpt", + "value": "From <36.2> to <42.5>." + } + ] + }, + { + "video_id": "7Z69Y", + "question_id": "7Z69Y", + "video_file": "Charades/videos/7Z69Y.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sitting on a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <36.2> to <42.5>." + } + ] + }, + { + "video_id": "ICALV", + "question_id": "ICALV", + "video_file": "Charades/videos/ICALV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'one person is eating the remains of a sandwich'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "ICALV", + "question_id": "ICALV", + "video_file": "Charades/videos/ICALV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'one person is eating' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "ICALV", + "question_id": "ICALV", + "video_file": "Charades/videos/ICALV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person is eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.2>." + } + ] + }, + { + "video_id": "W8WRK", + "question_id": "W8WRK", + "video_file": "Charades/videos/W8WRK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person cooking food in a pot'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "W8WRK", + "question_id": "W8WRK", + "video_file": "Charades/videos/W8WRK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is cooking on a stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "W8WRK", + "question_id": "W8WRK", + "video_file": "Charades/videos/W8WRK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person during the cooking process'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "W8WRK", + "question_id": "W8WRK", + "video_file": "Charades/videos/W8WRK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person returns to cooking' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "W8WRK", + "question_id": "W8WRK", + "video_file": "Charades/videos/W8WRK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person standing at a stove cooking' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "DC94A", + "question_id": "DC94A", + "video_file": "Charades/videos/DC94A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person putting away groceries'." + }, + { + "from": "gpt", + "value": "From <2.1> to <8.1>." + } + ] + }, + { + "video_id": "DC94A", + "question_id": "DC94A", + "video_file": "Charades/videos/DC94A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is opening a cabinet' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.5>." + } + ] + }, + { + "video_id": "DC94A", + "question_id": "DC94A", + "video_file": "Charades/videos/DC94A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits in the chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.2> to <26.7>." + } + ] + }, + { + "video_id": "DC94A", + "question_id": "DC94A", + "video_file": "Charades/videos/DC94A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person puts groceries inside a cupboard' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <2.1> to <8.1>." + } + ] + }, + { + "video_id": "VKTYD", + "question_id": "VKTYD", + "video_file": "Charades/videos/VKTYD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.8> to <11.0>." + } + ] + }, + { + "video_id": "VKTYD", + "question_id": "VKTYD", + "video_file": "Charades/videos/VKTYD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person a glass sits down on the couch' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.1> to <19.8>." + } + ] + }, + { + "video_id": "3YC35", + "question_id": "3YC35", + "video_file": "Charades/videos/3YC35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person hold th edoorknob' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.7> to <36.0>." + } + ] + }, + { + "video_id": "3YC35", + "question_id": "3YC35", + "video_file": "Charades/videos/3YC35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.4> to <26.1>." + } + ] + }, + { + "video_id": "3YC35", + "question_id": "3YC35", + "video_file": "Charades/videos/3YC35.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is tidying the floor with a broom'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.8> to <17.4>." + } + ] + }, + { + "video_id": "LKQQO", + "question_id": "LKQQO", + "video_file": "Charades/videos/LKQQO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is seen putting coffee on a table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.0> to <11.2>." + } + ] + }, + { + "video_id": "B0MFE", + "question_id": "B0MFE", + "video_file": "Charades/videos/B0MFE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks through a doorway holding some towels' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.4>." + } + ] + }, + { + "video_id": "6PYRZ", + "question_id": "6PYRZ", + "video_file": "Charades/videos/6PYRZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person throws a pillow on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.8> to <6.0>." + } + ] + }, + { + "video_id": "6PYRZ", + "question_id": "6PYRZ", + "video_file": "Charades/videos/6PYRZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a dish' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "6PYRZ", + "question_id": "6PYRZ", + "video_file": "Charades/videos/6PYRZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sneezing unto a plate of food in a hallway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <23.0>." + } + ] + }, + { + "video_id": "6PYRZ", + "question_id": "6PYRZ", + "video_file": "Charades/videos/6PYRZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is throwing a pillow' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.8> to <6.0>." + } + ] + }, + { + "video_id": "974PN", + "question_id": "974PN", + "video_file": "Charades/videos/974PN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person smiling as they watch something on their laptop'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.4>." + } + ] + }, + { + "video_id": "TR0TO", + "question_id": "TR0TO", + "video_file": "Charades/videos/TR0TO.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person puts the sandwich down' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <14.8>." + } + ] + }, + { + "video_id": "T5MLJ", + "question_id": "T5MLJ", + "video_file": "Charades/videos/T5MLJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person start laughing out of uncontrollable rage' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <24.6> to <33.7>." + } + ] + }, + { + "video_id": "6AHBW", + "question_id": "6AHBW", + "video_file": "Charades/videos/6AHBW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is drinking a glass of water' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.8> to <14.5>." + } + ] + }, + { + "video_id": "6AHBW", + "question_id": "6AHBW", + "video_file": "Charades/videos/6AHBW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is sitting at a table eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "6AHBW", + "question_id": "6AHBW", + "video_file": "Charades/videos/6AHBW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinking water from a glass'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.8> to <14.5>." + } + ] + }, + { + "video_id": "6AHBW", + "question_id": "6AHBW", + "video_file": "Charades/videos/6AHBW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating something' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "6AHBW", + "question_id": "6AHBW", + "video_file": "Charades/videos/6AHBW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person eating a sandwich' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.5>." + } + ] + }, + { + "video_id": "T5E2T", + "question_id": "T5E2T", + "video_file": "Charades/videos/T5E2T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a muffin' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <37.6> to <44.6>." + } + ] + }, + { + "video_id": "T5E2T", + "question_id": "T5E2T", + "video_file": "Charades/videos/T5E2T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is awakening from a bed' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.4> to <21.5>." + } + ] + }, + { + "video_id": "YH85Z", + "question_id": "YH85Z", + "video_file": "Charades/videos/YH85Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes out a bag of food'." + }, + { + "from": "gpt", + "value": "From <15.4> to <25.1>." + } + ] + }, + { + "video_id": "4BSR0", + "question_id": "4BSR0", + "video_file": "Charades/videos/4BSR0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person pouring coffee from one cup to another' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <10.3>." + } + ] + }, + { + "video_id": "TVCQF", + "question_id": "TVCQF", + "video_file": "Charades/videos/TVCQF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'another person is opening a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <12.0> to <17.0>." + } + ] + }, + { + "video_id": "AAJI0", + "question_id": "AAJI0", + "video_file": "Charades/videos/AAJI0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opening the window in the garage' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.8> to <20.6>." + } + ] + }, + { + "video_id": "AAJI0", + "question_id": "AAJI0", + "video_file": "Charades/videos/AAJI0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person open a window'." + }, + { + "from": "gpt", + "value": "From <10.8> to <20.6>." + } + ] + }, + { + "video_id": "AAJI0", + "question_id": "AAJI0", + "video_file": "Charades/videos/AAJI0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes off shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <11.8>." + } + ] + }, + { + "video_id": "AAJI0", + "question_id": "AAJI0", + "video_file": "Charades/videos/AAJI0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they take off their shoes'." + }, + { + "from": "gpt", + "value": "From <1.0> to <11.8>." + } + ] + }, + { + "video_id": "J5ZZQ", + "question_id": "J5ZZQ", + "video_file": "Charades/videos/J5ZZQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turns the light on by the doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.3> to <16.5>." + } + ] + }, + { + "video_id": "3064K", + "question_id": "3064K", + "video_file": "Charades/videos/3064K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is sitting on a couch folding laundry' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.3>." + } + ] + }, + { + "video_id": "RE9O1", + "question_id": "RE9O1", + "video_file": "Charades/videos/RE9O1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closing the door'." + }, + { + "from": "gpt", + "value": "From <14.4> to <24.3>." + } + ] + }, + { + "video_id": "RE9O1", + "question_id": "RE9O1", + "video_file": "Charades/videos/RE9O1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person walks through a doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.9> to <22.2>." + } + ] + }, + { + "video_id": "0XRD1", + "question_id": "0XRD1", + "video_file": "Charades/videos/0XRD1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person closes the book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.4>." + } + ] + }, + { + "video_id": "SLIH4", + "question_id": "SLIH4", + "video_file": "Charades/videos/SLIH4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closing the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <13.0> to <17.9>." + } + ] + }, + { + "video_id": "SLIH4", + "question_id": "SLIH4", + "video_file": "Charades/videos/SLIH4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they walk over to the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.8> to <15.5>." + } + ] + }, + { + "video_id": "SLIH4", + "question_id": "SLIH4", + "video_file": "Charades/videos/SLIH4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the door behind them' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <17.9>." + } + ] + }, + { + "video_id": "QLWQ4", + "question_id": "QLWQ4", + "video_file": "Charades/videos/QLWQ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is putting clothes in a box'." + }, + { + "from": "gpt", + "value": "From <21.4> to <43.4>." + } + ] + }, + { + "video_id": "QLWQ4", + "question_id": "QLWQ4", + "video_file": "Charades/videos/QLWQ4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put a blanket in a basket' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <4.2> to <14.2>." + } + ] + }, + { + "video_id": "GT53K", + "question_id": "GT53K", + "video_file": "Charades/videos/GT53K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns on the light as keeps eating' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.0> to <12.0>." + } + ] + }, + { + "video_id": "GT53K", + "question_id": "GT53K", + "video_file": "Charades/videos/GT53K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks to a light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <6.0> to <12.0>." + } + ] + }, + { + "video_id": "66J84", + "question_id": "66J84", + "video_file": "Charades/videos/66J84.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is laughing in the garage'." + }, + { + "from": "gpt", + "value": "From <15.5> to <26.4>." + } + ] + }, + { + "video_id": "66J84", + "question_id": "66J84", + "video_file": "Charades/videos/66J84.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walks into the room laughing' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <15.5> to <26.4>." + } + ] + }, + { + "video_id": "FD9A5", + "question_id": "FD9A5", + "video_file": "Charades/videos/FD9A5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person pours coffee into a cup'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <19.0>." + } + ] + }, + { + "video_id": "P2JL6", + "question_id": "P2JL6", + "video_file": "Charades/videos/P2JL6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person takes a picture with their camera' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <6.9>." + } + ] + }, + { + "video_id": "P2JL6", + "question_id": "P2JL6", + "video_file": "Charades/videos/P2JL6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person watching something out the window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.6>." + } + ] + }, + { + "video_id": "PT0D1", + "question_id": "PT0D1", + "video_file": "Charades/videos/PT0D1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turning a light on' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <17.0>." + } + ] + }, + { + "video_id": "PT0D1", + "question_id": "PT0D1", + "video_file": "Charades/videos/PT0D1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person turns on the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.4> to <17.0>." + } + ] + }, + { + "video_id": "PT0D1", + "question_id": "PT0D1", + "video_file": "Charades/videos/PT0D1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person turns on a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <10.4> to <17.0>." + } + ] + }, + { + "video_id": "WGJ8T", + "question_id": "WGJ8T", + "video_file": "Charades/videos/WGJ8T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person gets dressed'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "WGJ8T", + "question_id": "WGJ8T", + "video_file": "Charades/videos/WGJ8T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is in a hallway dressing by a table'." + }, + { + "from": "gpt", + "value": "From <0.0> to <6.0>." + } + ] + }, + { + "video_id": "MMGTX", + "question_id": "MMGTX", + "video_file": "Charades/videos/MMGTX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is holding up clothes for inspection' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.0> to <11.0>." + } + ] + }, + { + "video_id": "MMGTX", + "question_id": "MMGTX", + "video_file": "Charades/videos/MMGTX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they have taken them out of a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "JBJGX", + "question_id": "JBJGX", + "video_file": "Charades/videos/JBJGX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opens the laptop back up' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <24.6> to <31.5>." + } + ] + }, + { + "video_id": "U4XV9", + "question_id": "U4XV9", + "video_file": "Charades/videos/U4XV9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <17.9>." + } + ] + }, + { + "video_id": "U4XV9", + "question_id": "U4XV9", + "video_file": "Charades/videos/U4XV9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person awakens'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.5> to <10.4>." + } + ] + }, + { + "video_id": "U4XV9", + "question_id": "U4XV9", + "video_file": "Charades/videos/U4XV9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person put on some shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <26.1>." + } + ] + }, + { + "video_id": "U4XV9", + "question_id": "U4XV9", + "video_file": "Charades/videos/U4XV9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person lying on a bed awakens' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <2.5> to <10.4>." + } + ] + }, + { + "video_id": "U4XV9", + "question_id": "U4XV9", + "video_file": "Charades/videos/U4XV9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts on their shoes' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.4> to <26.1>." + } + ] + }, + { + "video_id": "U4XV9", + "question_id": "U4XV9", + "video_file": "Charades/videos/U4XV9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person puts on shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.4> to <26.1>." + } + ] + }, + { + "video_id": "M704T", + "question_id": "M704T", + "video_file": "Charades/videos/M704T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sneezing into it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.3> to <18.6>." + } + ] + }, + { + "video_id": "EGVGL", + "question_id": "EGVGL", + "video_file": "Charades/videos/EGVGL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person look at a picture' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.5> to <26.9>." + } + ] + }, + { + "video_id": "EGVGL", + "question_id": "EGVGL", + "video_file": "Charades/videos/EGVGL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they open a bag'." + }, + { + "from": "gpt", + "value": "From <17.5> to <23.0>." + } + ] + }, + { + "video_id": "55V6Z", + "question_id": "55V6Z", + "video_file": "Charades/videos/55V6Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person putting dishes in a box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <2.2> to <11.0>." + } + ] + }, + { + "video_id": "55V6Z", + "question_id": "55V6Z", + "video_file": "Charades/videos/55V6Z.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is watching a television' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.5> to <31.0>." + } + ] + }, + { + "video_id": "NB0S4", + "question_id": "NB0S4", + "video_file": "Charades/videos/NB0S4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts a bag on a shelf in the wardrobe'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.9> to <29.2>." + } + ] + }, + { + "video_id": "NB0S4", + "question_id": "NB0S4", + "video_file": "Charades/videos/NB0S4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'another person puts something on a shelf'." + }, + { + "from": "gpt", + "value": "From <20.9> to <29.2>." + } + ] + }, + { + "video_id": "TIEA9", + "question_id": "TIEA9", + "video_file": "Charades/videos/TIEA9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a woman is sitting on a bed smiling' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "WCERV", + "question_id": "WCERV", + "video_file": "Charades/videos/WCERV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person start working on their laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.4> to <32.0>." + } + ] + }, + { + "video_id": "3R7NX", + "question_id": "3R7NX", + "video_file": "Charades/videos/3R7NX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens the door to leave' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.1> to <31.0>." + } + ] + }, + { + "video_id": "3R7NX", + "question_id": "3R7NX", + "video_file": "Charades/videos/3R7NX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person opens a cabinet'." + }, + { + "from": "gpt", + "value": "From <26.5> to <31.0>." + } + ] + }, + { + "video_id": "3R7NX", + "question_id": "3R7NX", + "video_file": "Charades/videos/3R7NX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.1> to <31.0>." + } + ] + }, + { + "video_id": "ZS9IX", + "question_id": "ZS9IX", + "video_file": "Charades/videos/ZS9IX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person put their laptop on the table in the entryway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.3> to <21.7>." + } + ] + }, + { + "video_id": "ZS9IX", + "question_id": "ZS9IX", + "video_file": "Charades/videos/ZS9IX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person proceeds to undress from their winter gear' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <18.2> to <32.0>." + } + ] + }, + { + "video_id": "ZS9IX", + "question_id": "ZS9IX", + "video_file": "Charades/videos/ZS9IX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes off their shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <15.9>." + } + ] + }, + { + "video_id": "H8F1C", + "question_id": "H8F1C", + "video_file": "Charades/videos/H8F1C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts down a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <4.9> to <11.0>." + } + ] + }, + { + "video_id": "H8F1C", + "question_id": "H8F1C", + "video_file": "Charades/videos/H8F1C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is holding a book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.3>." + } + ] + }, + { + "video_id": "H8F1C", + "question_id": "H8F1C", + "video_file": "Charades/videos/H8F1C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person she looks at the picture on the wall,smiles'." + }, + { + "from": "gpt", + "value": "From <0.7> to <6.1>." + } + ] + }, + { + "video_id": "H8F1C", + "question_id": "H8F1C", + "video_file": "Charades/videos/H8F1C.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person begins undressing'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.3> to <18.0>." + } + ] + }, + { + "video_id": "XDT45", + "question_id": "XDT45", + "video_file": "Charades/videos/XDT45.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens'." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.7>." + } + ] + }, + { + "video_id": "AC9NI", + "question_id": "AC9NI", + "video_file": "Charades/videos/AC9NI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person walks through the doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.7> to <27.9>." + } + ] + }, + { + "video_id": "AC9NI", + "question_id": "AC9NI", + "video_file": "Charades/videos/AC9NI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person laughing at something on television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.0> to <10.0>." + } + ] + }, + { + "video_id": "AC9NI", + "question_id": "AC9NI", + "video_file": "Charades/videos/AC9NI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person sits in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.8> to <12.7>." + } + ] + }, + { + "video_id": "AC9NI", + "question_id": "AC9NI", + "video_file": "Charades/videos/AC9NI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person walk through a doorway'." + }, + { + "from": "gpt", + "value": "From <21.7> to <27.9>." + } + ] + }, + { + "video_id": "AC9NI", + "question_id": "AC9NI", + "video_file": "Charades/videos/AC9NI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person watches television'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.8>." + } + ] + }, + { + "video_id": "AC9NI", + "question_id": "AC9NI", + "video_file": "Charades/videos/AC9NI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sitting in a chair' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.8> to <12.7>." + } + ] + }, + { + "video_id": "N0DXI", + "question_id": "N0DXI", + "video_file": "Charades/videos/N0DXI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.2> to <19.5>." + } + ] + }, + { + "video_id": "HS3XP", + "question_id": "HS3XP", + "video_file": "Charades/videos/HS3XP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns off the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.6> to <14.0>." + } + ] + }, + { + "video_id": "HS3XP", + "question_id": "HS3XP", + "video_file": "Charades/videos/HS3XP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person turns off the light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.6> to <14.0>." + } + ] + }, + { + "video_id": "IC4D5", + "question_id": "IC4D5", + "video_file": "Charades/videos/IC4D5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person starts undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <18.1> to <25.3>." + } + ] + }, + { + "video_id": "QCIV0", + "question_id": "QCIV0", + "video_file": "Charades/videos/QCIV0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is sitting by the stairs eating some food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.7>." + } + ] + }, + { + "video_id": "QCIV0", + "question_id": "QCIV0", + "video_file": "Charades/videos/QCIV0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turn on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.2> to <21.0>." + } + ] + }, + { + "video_id": "QCIV0", + "question_id": "QCIV0", + "video_file": "Charades/videos/QCIV0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turn on a light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.2> to <21.0>." + } + ] + }, + { + "video_id": "4FAWP", + "question_id": "4FAWP", + "video_file": "Charades/videos/4FAWP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person take a picture of the desk' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.5> to <32.0>." + } + ] + }, + { + "video_id": "4FAWP", + "question_id": "4FAWP", + "video_file": "Charades/videos/4FAWP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person takes a picture of their work' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <32.0>." + } + ] + }, + { + "video_id": "KMWCM", + "question_id": "KMWCM", + "video_file": "Charades/videos/KMWCM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating from a bowl of food in their lap' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.0> to <36.2>." + } + ] + }, + { + "video_id": "KMWCM", + "question_id": "KMWCM", + "video_file": "Charades/videos/KMWCM.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.0> to <36.2>." + } + ] + }, + { + "video_id": "LECWG", + "question_id": "LECWG", + "video_file": "Charades/videos/LECWG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person opening a door'." + }, + { + "from": "gpt", + "value": "From <24.4> to <32.0>." + } + ] + }, + { + "video_id": "LECWG", + "question_id": "LECWG", + "video_file": "Charades/videos/LECWG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person dresses into a pair of shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.5>." + } + ] + }, + { + "video_id": "0O022", + "question_id": "0O022", + "video_file": "Charades/videos/0O022.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'another person is pouring water into a cup back' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.0> to <21.8>." + } + ] + }, + { + "video_id": "0O022", + "question_id": "0O022", + "video_file": "Charades/videos/0O022.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person is closing the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.9> to <10.1>." + } + ] + }, + { + "video_id": "0O022", + "question_id": "0O022", + "video_file": "Charades/videos/0O022.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is closing a cabinet door'." + }, + { + "from": "gpt", + "value": "From <0.9> to <10.1>." + } + ] + }, + { + "video_id": "0O022", + "question_id": "0O022", + "video_file": "Charades/videos/0O022.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person closes a cabinet'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.7>." + } + ] + }, + { + "video_id": "0O022", + "question_id": "0O022", + "video_file": "Charades/videos/0O022.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pours water from a glass' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.0> to <21.8>." + } + ] + }, + { + "video_id": "937EW", + "question_id": "937EW", + "video_file": "Charades/videos/937EW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person watches picture' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.9> to <24.2>." + } + ] + }, + { + "video_id": "937EW", + "question_id": "937EW", + "video_file": "Charades/videos/937EW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is putting clothes in a closet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "937EW", + "question_id": "937EW", + "video_file": "Charades/videos/937EW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts clothes away'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.0>." + } + ] + }, + { + "video_id": "937EW", + "question_id": "937EW", + "video_file": "Charades/videos/937EW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes dirty clothes away' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "JCLES", + "question_id": "JCLES", + "video_file": "Charades/videos/JCLES.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes a book out of the cabinet' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.6> to <29.5>." + } + ] + }, + { + "video_id": "JCLES", + "question_id": "JCLES", + "video_file": "Charades/videos/JCLES.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person opening the cabinet gives a smile' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <16.3>." + } + ] + }, + { + "video_id": "LX54K", + "question_id": "LX54K", + "video_file": "Charades/videos/LX54K.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holds his food' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.9> to <33.0>." + } + ] + }, + { + "video_id": "0OQVD", + "question_id": "0OQVD", + "video_file": "Charades/videos/0OQVD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person drinking coffee from a cup' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.6>." + } + ] + }, + { + "video_id": "42E9I", + "question_id": "42E9I", + "video_file": "Charades/videos/42E9I.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <31.0> to <34.0>." + } + ] + }, + { + "video_id": "JCGWC", + "question_id": "JCGWC", + "video_file": "Charades/videos/JCGWC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is closing the refrigerator' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "JCGWC", + "question_id": "JCGWC", + "video_file": "Charades/videos/JCGWC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they close the refrigerator door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "06OKE", + "question_id": "06OKE", + "video_file": "Charades/videos/06OKE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person they sit on the sofa' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <12.0>." + } + ] + }, + { + "video_id": "06OKE", + "question_id": "06OKE", + "video_file": "Charades/videos/06OKE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinks bottle of water'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.0> to <20.0>." + } + ] + }, + { + "video_id": "06OKE", + "question_id": "06OKE", + "video_file": "Charades/videos/06OKE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person put medicine back on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.1> to <13.7>." + } + ] + }, + { + "video_id": "06OKE", + "question_id": "06OKE", + "video_file": "Charades/videos/06OKE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person is putting medicine on towel'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "06OKE", + "question_id": "06OKE", + "video_file": "Charades/videos/06OKE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts a towel on a table'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "06OKE", + "question_id": "06OKE", + "video_file": "Charades/videos/06OKE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person starts smiling'." + }, + { + "from": "gpt", + "value": "From <2.8> to <10.1>." + } + ] + }, + { + "video_id": "JNWX7", + "question_id": "JNWX7", + "video_file": "Charades/videos/JNWX7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person closes the door'." + }, + { + "from": "gpt", + "value": "From <37.0> to <42.8>." + } + ] + }, + { + "video_id": "NVF2S", + "question_id": "NVF2S", + "video_file": "Charades/videos/NVF2S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person take the medicine'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.5> to <25.0>." + } + ] + }, + { + "video_id": "NVF2S", + "question_id": "NVF2S", + "video_file": "Charades/videos/NVF2S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they take a medicine bottle from the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.5> to <25.0>." + } + ] + }, + { + "video_id": "X5QMG", + "question_id": "X5QMG", + "video_file": "Charades/videos/X5QMG.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the sitting person fixes the positioning of the chairs' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.4>." + } + ] + }, + { + "video_id": "9AFSH", + "question_id": "9AFSH", + "video_file": "Charades/videos/9AFSH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is running towards a blanket'." + }, + { + "from": "gpt", + "value": "From <11.7> to <22.0>." + } + ] + }, + { + "video_id": "9AFSH", + "question_id": "9AFSH", + "video_file": "Charades/videos/9AFSH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is throwing a pillow onto a chair' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <10.7>." + } + ] + }, + { + "video_id": "9AFSH", + "question_id": "9AFSH", + "video_file": "Charades/videos/9AFSH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs around' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.7> to <22.0>." + } + ] + }, + { + "video_id": "9AFSH", + "question_id": "9AFSH", + "video_file": "Charades/videos/9AFSH.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person throws a pillow on a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.5> to <10.7>." + } + ] + }, + { + "video_id": "JIWY2", + "question_id": "JIWY2", + "video_file": "Charades/videos/JIWY2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is seen eating some food' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.2>." + } + ] + }, + { + "video_id": "JIWY2", + "question_id": "JIWY2", + "video_file": "Charades/videos/JIWY2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person look out the window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.5> to <29.0>." + } + ] + }, + { + "video_id": "JIWY2", + "question_id": "JIWY2", + "video_file": "Charades/videos/JIWY2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person eats cereal' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.2>." + } + ] + }, + { + "video_id": "JIWY2", + "question_id": "JIWY2", + "video_file": "Charades/videos/JIWY2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person drinks a cup of coffee at the kitchen table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <18.8>." + } + ] + }, + { + "video_id": "JIWY2", + "question_id": "JIWY2", + "video_file": "Charades/videos/JIWY2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating food out of a plastic container'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.2>." + } + ] + }, + { + "video_id": "2PMYI", + "question_id": "2PMYI", + "video_file": "Charades/videos/2PMYI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person they awaken'." + }, + { + "from": "gpt", + "value": "From <2.5> to <9.9>." + } + ] + }, + { + "video_id": "XHQT1", + "question_id": "XHQT1", + "video_file": "Charades/videos/XHQT1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person walks through a doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <26.4> to <34.2>." + } + ] + }, + { + "video_id": "XHQT1", + "question_id": "XHQT1", + "video_file": "Charades/videos/XHQT1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person opens a door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.7> to <33.4>." + } + ] + }, + { + "video_id": "F207R", + "question_id": "F207R", + "video_file": "Charades/videos/F207R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person takes out a book' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <5.7> to <12.8>." + } + ] + }, + { + "video_id": "F207R", + "question_id": "F207R", + "video_file": "Charades/videos/F207R.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts it on the shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <34.0> to <42.0>." + } + ] + }, + { + "video_id": "8UQVB", + "question_id": "8UQVB", + "video_file": "Charades/videos/8UQVB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is watching their laptop screen' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.6> to <33.7>." + } + ] + }, + { + "video_id": "8UQVB", + "question_id": "8UQVB", + "video_file": "Charades/videos/8UQVB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is sneezing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.8> to <6.4>." + } + ] + }, + { + "video_id": "KTQEL", + "question_id": "KTQEL", + "video_file": "Charades/videos/KTQEL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding a picture opens the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.6> to <11.9>." + } + ] + }, + { + "video_id": "KTQEL", + "question_id": "KTQEL", + "video_file": "Charades/videos/KTQEL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts a picture on the shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <8.0> to <23.0>." + } + ] + }, + { + "video_id": "004QE", + "question_id": "004QE", + "video_file": "Charades/videos/004QE.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a drink from the cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.2> to <22.5>." + } + ] + }, + { + "video_id": "45D5U", + "question_id": "45D5U", + "video_file": "Charades/videos/45D5U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts phone on table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.1> to <30.7>." + } + ] + }, + { + "video_id": "ZSSNJ", + "question_id": "ZSSNJ", + "video_file": "Charades/videos/ZSSNJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person turns on the light in their pantry' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <2.9>." + } + ] + }, + { + "video_id": "ZSSNJ", + "question_id": "ZSSNJ", + "video_file": "Charades/videos/ZSSNJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person start laughing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <7.0> to <16.5>." + } + ] + }, + { + "video_id": "ZSSNJ", + "question_id": "ZSSNJ", + "video_file": "Charades/videos/ZSSNJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person sweeps floor takes phone out of pocket laughing texting' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.0> to <16.5>." + } + ] + }, + { + "video_id": "B2NBR", + "question_id": "B2NBR", + "video_file": "Charades/videos/B2NBR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person runs back'." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "B2NBR", + "question_id": "B2NBR", + "video_file": "Charades/videos/B2NBR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a man is running through the hallway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <8.3>." + } + ] + }, + { + "video_id": "59CDL", + "question_id": "59CDL", + "video_file": "Charades/videos/59CDL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sits down to look out the window' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.7> to <33.0>." + } + ] + }, + { + "video_id": "59CDL", + "question_id": "59CDL", + "video_file": "Charades/videos/59CDL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sits in the chair to stare outside' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.1> to <33.0>." + } + ] + }, + { + "video_id": "8J5S4", + "question_id": "8J5S4", + "video_file": "Charades/videos/8J5S4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is seen eating' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <15.4>." + } + ] + }, + { + "video_id": "8J5S4", + "question_id": "8J5S4", + "video_file": "Charades/videos/8J5S4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns off the lights' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <19.4> to <24.8>." + } + ] + }, + { + "video_id": "8J5S4", + "question_id": "8J5S4", + "video_file": "Charades/videos/8J5S4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is eating' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <15.4>." + } + ] + }, + { + "video_id": "8J5S4", + "question_id": "8J5S4", + "video_file": "Charades/videos/8J5S4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turn off the nearby light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.4> to <24.8>." + } + ] + }, + { + "video_id": "W7EHF", + "question_id": "W7EHF", + "video_file": "Charades/videos/W7EHF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is standing in a doorway.the open a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "W7EHF", + "question_id": "W7EHF", + "video_file": "Charades/videos/W7EHF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person put a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.7> to <24.0>." + } + ] + }, + { + "video_id": "W7EHF", + "question_id": "W7EHF", + "video_file": "Charades/videos/W7EHF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is opening a box'." + }, + { + "from": "gpt", + "value": "From <0.0> to <9.9>." + } + ] + }, + { + "video_id": "ZOBZV", + "question_id": "ZOBZV", + "video_file": "Charades/videos/ZOBZV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person putting their phone on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <34.1> to <40.0>." + } + ] + }, + { + "video_id": "ZOBZV", + "question_id": "ZOBZV", + "video_file": "Charades/videos/ZOBZV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person puts the phone down on a shelf'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <34.1> to <40.0>." + } + ] + }, + { + "video_id": "RYK18", + "question_id": "RYK18", + "video_file": "Charades/videos/RYK18.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is in a bedroom sitting on a chair'." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.0>." + } + ] + }, + { + "video_id": "RYK18", + "question_id": "RYK18", + "video_file": "Charades/videos/RYK18.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turns off a light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.2> to <33.0>." + } + ] + }, + { + "video_id": "RYK18", + "question_id": "RYK18", + "video_file": "Charades/videos/RYK18.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turning off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.2> to <33.0>." + } + ] + }, + { + "video_id": "RYK18", + "question_id": "RYK18", + "video_file": "Charades/videos/RYK18.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person start undressing from their clothes'." + }, + { + "from": "gpt", + "value": "From <16.4> to <30.2>." + } + ] + }, + { + "video_id": "RYK18", + "question_id": "RYK18", + "video_file": "Charades/videos/RYK18.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins to undress' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.4> to <30.2>." + } + ] + }, + { + "video_id": "ABHC6", + "question_id": "ABHC6", + "video_file": "Charades/videos/ABHC6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person turns off the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.5> to <23.9>." + } + ] + }, + { + "video_id": "50TLF", + "question_id": "50TLF", + "video_file": "Charades/videos/50TLF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person takes a bottle of medicine from a box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.2> to <10.0>." + } + ] + }, + { + "video_id": "13IJA", + "question_id": "13IJA", + "video_file": "Charades/videos/13IJA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person throws the blanket on the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.9> to <10.9>." + } + ] + }, + { + "video_id": "13IJA", + "question_id": "13IJA", + "video_file": "Charades/videos/13IJA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person is sitting down holding a blanket' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <7.0>." + } + ] + }, + { + "video_id": "5S2ZU", + "question_id": "5S2ZU", + "video_file": "Charades/videos/5S2ZU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is in the kitchen cooking'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <18.5>." + } + ] + }, + { + "video_id": "5S2ZU", + "question_id": "5S2ZU", + "video_file": "Charades/videos/5S2ZU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is seen cooking on the stove' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <18.5>." + } + ] + }, + { + "video_id": "F8A4G", + "question_id": "F8A4G", + "video_file": "Charades/videos/F8A4G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person turns the doorknob to open the garage door'." + }, + { + "from": "gpt", + "value": "From <1.8> to <9.7>." + } + ] + }, + { + "video_id": "F8A4G", + "question_id": "F8A4G", + "video_file": "Charades/videos/F8A4G.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person opened a door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.8> to <9.7>." + } + ] + }, + { + "video_id": "GBPPR", + "question_id": "GBPPR", + "video_file": "Charades/videos/GBPPR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person stops to take a drink out of their cup' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.3> to <31.0>." + } + ] + }, + { + "video_id": "GBPPR", + "question_id": "GBPPR", + "video_file": "Charades/videos/GBPPR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person eating some food'." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.0>." + } + ] + }, + { + "video_id": "GBPPR", + "question_id": "GBPPR", + "video_file": "Charades/videos/GBPPR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person sitting on the couch eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <13.0>." + } + ] + }, + { + "video_id": "HR6MF", + "question_id": "HR6MF", + "video_file": "Charades/videos/HR6MF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person holding a broom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <8.2>." + } + ] + }, + { + "video_id": "HR6MF", + "question_id": "HR6MF", + "video_file": "Charades/videos/HR6MF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is running laughing at a desk' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "HR6MF", + "question_id": "HR6MF", + "video_file": "Charades/videos/HR6MF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a laughing person runs through the hallway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.8>." + } + ] + }, + { + "video_id": "HR6MF", + "question_id": "HR6MF", + "video_file": "Charades/videos/HR6MF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person starts tidying a table with a laptop on it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <15.7>." + } + ] + }, + { + "video_id": "GIA8M", + "question_id": "GIA8M", + "video_file": "Charades/videos/GIA8M.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'another person is smiling at a pile of clothes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.3>." + } + ] + }, + { + "video_id": "8NW4J", + "question_id": "8NW4J", + "video_file": "Charades/videos/8NW4J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting away dirty clothes in a bedroom' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.5> to <16.6>." + } + ] + }, + { + "video_id": "8NW4J", + "question_id": "8NW4J", + "video_file": "Charades/videos/8NW4J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is lying on a bed holding a pillow' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.1>." + } + ] + }, + { + "video_id": "8NW4J", + "question_id": "8NW4J", + "video_file": "Charades/videos/8NW4J.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is snuggling with a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.9>." + } + ] + }, + { + "video_id": "YHPLU", + "question_id": "YHPLU", + "video_file": "Charades/videos/YHPLU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eats them' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <24.3>." + } + ] + }, + { + "video_id": "YHPLU", + "question_id": "YHPLU", + "video_file": "Charades/videos/YHPLU.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'one person is eating in the doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <24.3>." + } + ] + }, + { + "video_id": "X8DZV", + "question_id": "X8DZV", + "video_file": "Charades/videos/X8DZV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <15.6> to <21.9>." + } + ] + }, + { + "video_id": "X8DZV", + "question_id": "X8DZV", + "video_file": "Charades/videos/X8DZV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person turns on a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.4> to <8.7>." + } + ] + }, + { + "video_id": "X8DZV", + "question_id": "X8DZV", + "video_file": "Charades/videos/X8DZV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinks from a glass'." + }, + { + "from": "gpt", + "value": "From <8.6> to <17.8>." + } + ] + }, + { + "video_id": "X8DZV", + "question_id": "X8DZV", + "video_file": "Charades/videos/X8DZV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person opens a box to grab a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <16.4> to <28.6>." + } + ] + }, + { + "video_id": "X8DZV", + "question_id": "X8DZV", + "video_file": "Charades/videos/X8DZV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person sits down on a chair drinking some water'." + }, + { + "from": "gpt", + "value": "From <6.5> to <12.9>." + } + ] + }, + { + "video_id": "R3O7U", + "question_id": "R3O7U", + "video_file": "Charades/videos/R3O7U.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is washing a window' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.4> to <19.7>." + } + ] + }, + { + "video_id": "81YF0", + "question_id": "81YF0", + "video_file": "Charades/videos/81YF0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is holding a towel in the hallway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.3> to <18.2>." + } + ] + }, + { + "video_id": "81YF0", + "question_id": "81YF0", + "video_file": "Charades/videos/81YF0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person throws a towel onto the floor' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.2> to <28.3>." + } + ] + }, + { + "video_id": "CAND1", + "question_id": "CAND1", + "video_file": "Charades/videos/CAND1.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person is washing a glass window' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.1> to <23.9>." + } + ] + }, + { + "video_id": "M1SRW", + "question_id": "M1SRW", + "video_file": "Charades/videos/M1SRW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they put the camera on a table' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <21.1> to <27.1>." + } + ] + }, + { + "video_id": "ZV4JV", + "question_id": "ZV4JV", + "video_file": "Charades/videos/ZV4JV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person holding their phone' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <11.2>." + } + ] + }, + { + "video_id": "HA1XJ", + "question_id": "HA1XJ", + "video_file": "Charades/videos/HA1XJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person looks out door window'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.4>." + } + ] + }, + { + "video_id": "HA1XJ", + "question_id": "HA1XJ", + "video_file": "Charades/videos/HA1XJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking from a glass' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <1.0> to <7.6>." + } + ] + }, + { + "video_id": "HA1XJ", + "question_id": "HA1XJ", + "video_file": "Charades/videos/HA1XJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinks from cup'." + }, + { + "from": "gpt", + "value": "From <1.0> to <7.6>." + } + ] + }, + { + "video_id": "RS9D9", + "question_id": "RS9D9", + "video_file": "Charades/videos/RS9D9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person reads a book' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.3> to <36.8>." + } + ] + }, + { + "video_id": "RS9D9", + "question_id": "RS9D9", + "video_file": "Charades/videos/RS9D9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person reading a book' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <27.3> to <36.8>." + } + ] + }, + { + "video_id": "RS9D9", + "question_id": "RS9D9", + "video_file": "Charades/videos/RS9D9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person sits on a chair' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <6.4>." + } + ] + }, + { + "video_id": "RS9D9", + "question_id": "RS9D9", + "video_file": "Charades/videos/RS9D9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person sitting in a chair' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.1> to <6.4>." + } + ] + }, + { + "video_id": "WI6NB", + "question_id": "WI6NB", + "video_file": "Charades/videos/WI6NB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats a snack' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.4> to <26.1>." + } + ] + }, + { + "video_id": "WI6NB", + "question_id": "WI6NB", + "video_file": "Charades/videos/WI6NB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eats cookies' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <13.4> to <26.1>." + } + ] + }, + { + "video_id": "1SJQX", + "question_id": "1SJQX", + "video_file": "Charades/videos/1SJQX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person holding a phone picks up some homework' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.3>." + } + ] + }, + { + "video_id": "1SJQX", + "question_id": "1SJQX", + "video_file": "Charades/videos/1SJQX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'one person walks in holding a phone'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.3>." + } + ] + }, + { + "video_id": "ZXN6A", + "question_id": "ZXN6A", + "video_file": "Charades/videos/ZXN6A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person drinking a cup of coffee'." + }, + { + "from": "gpt", + "value": "From <3.8> to <12.5>." + } + ] + }, + { + "video_id": "ZPNW7", + "question_id": "ZPNW7", + "video_file": "Charades/videos/ZPNW7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is walking through a doorway'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <23.7> to <28.8>." + } + ] + }, + { + "video_id": "ZPNW7", + "question_id": "ZPNW7", + "video_file": "Charades/videos/ZPNW7.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is snuggling with a pillow' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <32.2> to <53.4>." + } + ] + }, + { + "video_id": "ECXEX", + "question_id": "ECXEX", + "video_file": "Charades/videos/ECXEX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person puts them on the table' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <12.5>." + } + ] + }, + { + "video_id": "ECXEX", + "question_id": "ECXEX", + "video_file": "Charades/videos/ECXEX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person takes towel off of shoulder'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.7> to <32.0>." + } + ] + }, + { + "video_id": "ECXEX", + "question_id": "ECXEX", + "video_file": "Charades/videos/ECXEX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person the takes dishes from a shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <14.5> to <24.0>." + } + ] + }, + { + "video_id": "ECXEX", + "question_id": "ECXEX", + "video_file": "Charades/videos/ECXEX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person puts groceries on the shelf' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <4.9> to <12.3>." + } + ] + }, + { + "video_id": "ECXEX", + "question_id": "ECXEX", + "video_file": "Charades/videos/ECXEX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts a plate of food on a table' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.0> to <12.5>." + } + ] + }, + { + "video_id": "4CV3F", + "question_id": "4CV3F", + "video_file": "Charades/videos/4CV3F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts on a pair of shoes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "4CV3F", + "question_id": "4CV3F", + "video_file": "Charades/videos/4CV3F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person puts their shoes on the floor' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "4CV3F", + "question_id": "4CV3F", + "video_file": "Charades/videos/4CV3F.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person puts the shoes on' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.7>." + } + ] + }, + { + "video_id": "4VODV", + "question_id": "4VODV", + "video_file": "Charades/videos/4VODV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person who is lying on the floor awakens'." + }, + { + "from": "gpt", + "value": "From <0.0> to <15.6>." + } + ] + }, + { + "video_id": "4VODV", + "question_id": "4VODV", + "video_file": "Charades/videos/4VODV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eating a sandwich out of the box' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <29.9> to <36.0>." + } + ] + }, + { + "video_id": "4VODV", + "question_id": "4VODV", + "video_file": "Charades/videos/4VODV.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'the person began to undress' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <21.2> to <29.1>." + } + ] + }, + { + "video_id": "P4HXN", + "question_id": "P4HXN", + "video_file": "Charades/videos/P4HXN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person puts the laptop back on the clothes'." + }, + { + "from": "gpt", + "value": "From <26.8> to <33.3>." + } + ] + }, + { + "video_id": "TNOBD", + "question_id": "TNOBD", + "video_file": "Charades/videos/TNOBD.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person eating lunch' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.7> to <31.0>." + } + ] + }, + { + "video_id": "SYMIR", + "question_id": "SYMIR", + "video_file": "Charades/videos/SYMIR.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is standing in the hall undressing' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.3> to <14.6>." + } + ] + }, + { + "video_id": "UTLNI", + "question_id": "UTLNI", + "video_file": "Charades/videos/UTLNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person open a cabinet' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.1> to <16.9>." + } + ] + }, + { + "video_id": "UTLNI", + "question_id": "UTLNI", + "video_file": "Charades/videos/UTLNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is holding a phone'." + }, + { + "from": "gpt", + "value": "From <0.6> to <10.1>." + } + ] + }, + { + "video_id": "UTLNI", + "question_id": "UTLNI", + "video_file": "Charades/videos/UTLNI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is playing on their phone' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <5.4>." + } + ] + }, + { + "video_id": "VKKIQ", + "question_id": "VKKIQ", + "video_file": "Charades/videos/VKKIQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person closes the door'." + }, + { + "from": "gpt", + "value": "From <11.4> to <17.3>." + } + ] + }, + { + "video_id": "VKKIQ", + "question_id": "VKKIQ", + "video_file": "Charades/videos/VKKIQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a refrigerator'." + }, + { + "from": "gpt", + "value": "From <0.2> to <8.4>." + } + ] + }, + { + "video_id": "VKKIQ", + "question_id": "VKKIQ", + "video_file": "Charades/videos/VKKIQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'the person takes out a bottle of iced coffee' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <3.6> to <11.0>." + } + ] + }, + { + "video_id": "VKKIQ", + "question_id": "VKKIQ", + "video_file": "Charades/videos/VKKIQ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the refrigerator' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <11.2> to <17.0>." + } + ] + }, + { + "video_id": "5TG5P", + "question_id": "5TG5P", + "video_file": "Charades/videos/5TG5P.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'encounters person b smiling'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.3> to <10.0>." + } + ] + }, + { + "video_id": "YXPTI", + "question_id": "YXPTI", + "video_file": "Charades/videos/YXPTI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person puts some shoes'." + }, + { + "from": "gpt", + "value": "From <1.3> to <12.4>." + } + ] + }, + { + "video_id": "WXMTI", + "question_id": "WXMTI", + "video_file": "Charades/videos/WXMTI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person putting shoes inside' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.8> to <8.6>." + } + ] + }, + { + "video_id": "WXMTI", + "question_id": "WXMTI", + "video_file": "Charades/videos/WXMTI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'one person puts a pair of shoes into a bag' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.8> to <8.6>." + } + ] + }, + { + "video_id": "WXMTI", + "question_id": "WXMTI", + "video_file": "Charades/videos/WXMTI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is holding a bag'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <3.3>." + } + ] + }, + { + "video_id": "J1EA0", + "question_id": "J1EA0", + "video_file": "Charades/videos/J1EA0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person holds a bag open' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.5>." + } + ] + }, + { + "video_id": "G4E4S", + "question_id": "G4E4S", + "video_file": "Charades/videos/G4E4S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person goes over to wash the clothes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.8> to <57.5>." + } + ] + }, + { + "video_id": "G4E4S", + "question_id": "G4E4S", + "video_file": "Charades/videos/G4E4S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person goes to sit down to take off shoes' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <17.4> to <27.8>." + } + ] + }, + { + "video_id": "G4E4S", + "question_id": "G4E4S", + "video_file": "Charades/videos/G4E4S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person take off their shoes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.4> to <27.8>." + } + ] + }, + { + "video_id": "G4E4S", + "question_id": "G4E4S", + "video_file": "Charades/videos/G4E4S.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begin washing their clothes' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <33.8> to <57.5>." + } + ] + }, + { + "video_id": "EAWW4", + "question_id": "EAWW4", + "video_file": "Charades/videos/EAWW4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person puts shoes on a shelf' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <7.6> to <15.0>." + } + ] + }, + { + "video_id": "EAWW4", + "question_id": "EAWW4", + "video_file": "Charades/videos/EAWW4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person opens a laundry room door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.0> to <9.4>." + } + ] + }, + { + "video_id": "MVL2O", + "question_id": "MVL2O", + "video_file": "Charades/videos/MVL2O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they close the laptop' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.3> to <24.4>." + } + ] + }, + { + "video_id": "MVL2O", + "question_id": "MVL2O", + "video_file": "Charades/videos/MVL2O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person stand up' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <20.2> to <32.5>." + } + ] + }, + { + "video_id": "MVL2O", + "question_id": "MVL2O", + "video_file": "Charades/videos/MVL2O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person finally turning off a light on the stairs' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <35.4> to <39.0>." + } + ] + }, + { + "video_id": "MVL2O", + "question_id": "MVL2O", + "video_file": "Charades/videos/MVL2O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person goes up the stairs turn the light off' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <35.4> to <39.0>." + } + ] + }, + { + "video_id": "5OCXJ", + "question_id": "5OCXJ", + "video_file": "Charades/videos/5OCXJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens in the living room'." + }, + { + "from": "gpt", + "value": "From <2.8> to <15.2>." + } + ] + }, + { + "video_id": "8BA61", + "question_id": "8BA61", + "video_file": "Charades/videos/8BA61.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eat a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.3> to <30.5>." + } + ] + }, + { + "video_id": "8BA61", + "question_id": "8BA61", + "video_file": "Charades/videos/8BA61.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eats a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.3> to <30.5>." + } + ] + }, + { + "video_id": "8BA61", + "question_id": "8BA61", + "video_file": "Charades/videos/8BA61.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person eat a snack' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.3> to <30.5>." + } + ] + }, + { + "video_id": "8AZRX", + "question_id": "8AZRX", + "video_file": "Charades/videos/8AZRX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person turn off the light' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <28.5> to <33.8>." + } + ] + }, + { + "video_id": "8AZRX", + "question_id": "8AZRX", + "video_file": "Charades/videos/8AZRX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws some clothes in a bag' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <15.3> to <23.2>." + } + ] + }, + { + "video_id": "H4YTC", + "question_id": "H4YTC", + "video_file": "Charades/videos/H4YTC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns off the light'." + }, + { + "from": "gpt", + "value": "From <27.3> to <32.9>." + } + ] + }, + { + "video_id": "7ICIA", + "question_id": "7ICIA", + "video_file": "Charades/videos/7ICIA.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.8> to <16.5>." + } + ] + }, + { + "video_id": "EIKUN", + "question_id": "EIKUN", + "video_file": "Charades/videos/EIKUN.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person puts a box of groceries onto the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <5.5> to <11.7>." + } + ] + }, + { + "video_id": "2XSK6", + "question_id": "2XSK6", + "video_file": "Charades/videos/2XSK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they sit in a chair at a desk' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.5> to <30.8>." + } + ] + }, + { + "video_id": "2XSK6", + "question_id": "2XSK6", + "video_file": "Charades/videos/2XSK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person closes the door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <9.7> to <18.2>." + } + ] + }, + { + "video_id": "2XSK6", + "question_id": "2XSK6", + "video_file": "Charades/videos/2XSK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person closes the door to the office' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <18.2>." + } + ] + }, + { + "video_id": "2XSK6", + "question_id": "2XSK6", + "video_file": "Charades/videos/2XSK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open the door again' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.4> to <6.2>." + } + ] + }, + { + "video_id": "2XSK6", + "question_id": "2XSK6", + "video_file": "Charades/videos/2XSK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person opens a door'." + }, + { + "from": "gpt", + "value": "From <0.4> to <6.2>." + } + ] + }, + { + "video_id": "2XSK6", + "question_id": "2XSK6", + "video_file": "Charades/videos/2XSK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person close the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.7> to <18.2>." + } + ] + }, + { + "video_id": "2XSK6", + "question_id": "2XSK6", + "video_file": "Charades/videos/2XSK6.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person sits down in a chair at a desk'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.5> to <30.8>." + } + ] + }, + { + "video_id": "QQUU5", + "question_id": "QQUU5", + "video_file": "Charades/videos/QQUU5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person close the front door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <3.5> to <10.1>." + } + ] + }, + { + "video_id": "QQUU5", + "question_id": "QQUU5", + "video_file": "Charades/videos/QQUU5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person closes the door' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <3.5> to <10.1>." + } + ] + }, + { + "video_id": "QQUU5", + "question_id": "QQUU5", + "video_file": "Charades/videos/QQUU5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person is drinking a cup of coffee' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.8>." + } + ] + }, + { + "video_id": "QQUU5", + "question_id": "QQUU5", + "video_file": "Charades/videos/QQUU5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the coffee down on the table'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <6.4>." + } + ] + }, + { + "video_id": "58VZI", + "question_id": "58VZI", + "video_file": "Charades/videos/58VZI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person puts a pillow on the bench'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.5> to <23.4>." + } + ] + }, + { + "video_id": "58VZI", + "question_id": "58VZI", + "video_file": "Charades/videos/58VZI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person drinking a glass of water' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <0.8> to <14.0>." + } + ] + }, + { + "video_id": "75ZOJ", + "question_id": "75ZOJ", + "video_file": "Charades/videos/75ZOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person looking through papers/book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.0>." + } + ] + }, + { + "video_id": "75ZOJ", + "question_id": "75ZOJ", + "video_file": "Charades/videos/75ZOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'the person drinks from a water bottle' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <20.6>." + } + ] + }, + { + "video_id": "75ZOJ", + "question_id": "75ZOJ", + "video_file": "Charades/videos/75ZOJ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person drinks from a water bottle'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.1> to <20.6>." + } + ] + }, + { + "video_id": "W25MK", + "question_id": "W25MK", + "video_file": "Charades/videos/W25MK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they close the book' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.5> to <25.4>." + } + ] + }, + { + "video_id": "MBSW4", + "question_id": "MBSW4", + "video_file": "Charades/videos/MBSW4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they open the box' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.4> to <12.4>." + } + ] + }, + { + "video_id": "MBSW4", + "question_id": "MBSW4", + "video_file": "Charades/videos/MBSW4.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person take out clothes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <8.4> to <14.6>." + } + ] + }, + { + "video_id": "FN71T", + "question_id": "FN71T", + "video_file": "Charades/videos/FN71T.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person they take a pile of clothes' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <20.0> to <28.4>." + } + ] + }, + { + "video_id": "18FTZ", + "question_id": "18FTZ", + "video_file": "Charades/videos/18FTZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person goes into the kitchen to eat something' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <25.0> to <33.0>." + } + ] + }, + { + "video_id": "18FTZ", + "question_id": "18FTZ", + "video_file": "Charades/videos/18FTZ.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a person is eating over a sink' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.0> to <33.0>." + } + ] + }, + { + "video_id": "4ZJSI", + "question_id": "4ZJSI", + "video_file": "Charades/videos/4ZJSI.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person opens a door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.5> to <17.4>." + } + ] + }, + { + "video_id": "JWPS5", + "question_id": "JWPS5", + "video_file": "Charades/videos/JWPS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person eating a sandwich they stand up'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.6>." + } + ] + }, + { + "video_id": "JWPS5", + "question_id": "JWPS5", + "video_file": "Charades/videos/JWPS5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <14.6>." + } + ] + }, + { + "video_id": "P4AEF", + "question_id": "P4AEF", + "video_file": "Charades/videos/P4AEF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person awakens from sleep'." + }, + { + "from": "gpt", + "value": "From <0.2> to <8.5>." + } + ] + }, + { + "video_id": "P4AEF", + "question_id": "P4AEF", + "video_file": "Charades/videos/P4AEF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person covered in a blanket awakens in a bedroom' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.2> to <8.5>." + } + ] + }, + { + "video_id": "JCBDT", + "question_id": "JCBDT", + "video_file": "Charades/videos/JCBDT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a other person is holding a picture laughing at it' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <25.9> to <31.0>." + } + ] + }, + { + "video_id": "JCBDT", + "question_id": "JCBDT", + "video_file": "Charades/videos/JCBDT.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person sitting on a bed' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.5> to <32.0>." + } + ] + }, + { + "video_id": "UBDB5", + "question_id": "UBDB5", + "video_file": "Charades/videos/UBDB5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person throws a blanket on the ground' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <6.2>." + } + ] + }, + { + "video_id": "UBDB5", + "question_id": "UBDB5", + "video_file": "Charades/videos/UBDB5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person throws a blanket into the corner'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.1> to <6.2>." + } + ] + }, + { + "video_id": "D6D85", + "question_id": "D6D85", + "video_file": "Charades/videos/D6D85.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'a person is eating a sandwich on a chair'." + }, + { + "from": "gpt", + "value": "From <19.0> to <27.4>." + } + ] + }, + { + "video_id": "RWPS8", + "question_id": "RWPS8", + "video_file": "Charades/videos/RWPS8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they put the shoe on a counter'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <17.8> to <24.3>." + } + ] + }, + { + "video_id": "RWPS8", + "question_id": "RWPS8", + "video_file": "Charades/videos/RWPS8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person starts smiling' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <17.7> to <24.3>." + } + ] + }, + { + "video_id": "RWPS8", + "question_id": "RWPS8", + "video_file": "Charades/videos/RWPS8.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person begin to laugh when they see a picture'." + }, + { + "from": "gpt", + "value": "From <20.5> to <29.0>." + } + ] + }, + { + "video_id": "0H5VC", + "question_id": "0H5VC", + "video_file": "Charades/videos/0H5VC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person begins snuggling into their blanket' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <21.4> to <33.0>." + } + ] + }, + { + "video_id": "0H5VC", + "question_id": "0H5VC", + "video_file": "Charades/videos/0H5VC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they close the book'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <14.7> to <19.8>." + } + ] + }, + { + "video_id": "44HW5", + "question_id": "44HW5", + "video_file": "Charades/videos/44HW5.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'a man runs into the room' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <16.7>." + } + ] + }, + { + "video_id": "J6P6H", + "question_id": "J6P6H", + "video_file": "Charades/videos/J6P6H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'the person starts eating a sandwich'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.5> to <17.4>." + } + ] + }, + { + "video_id": "J6P6H", + "question_id": "J6P6H", + "video_file": "Charades/videos/J6P6H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eating a sandwich' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <9.5> to <17.4>." + } + ] + }, + { + "video_id": "J6P6H", + "question_id": "J6P6H", + "video_file": "Charades/videos/J6P6H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'the person stands up to turn off their light'." + }, + { + "from": "gpt", + "value": "From <17.2> to <23.0>." + } + ] + }, + { + "video_id": "HJH7H", + "question_id": "HJH7H", + "video_file": "Charades/videos/HJH7H.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'throws a pillow at a person' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <11.0> to <16.0>." + } + ] + }, + { + "video_id": "WK5PL", + "question_id": "WK5PL", + "video_file": "Charades/videos/WK5PL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is closing a closet door' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <6.4> to <12.4>." + } + ] + }, + { + "video_id": "WK5PL", + "question_id": "WK5PL", + "video_file": "Charades/videos/WK5PL.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person closing the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <6.4> to <12.4>." + } + ] + }, + { + "video_id": "4AQZ2", + "question_id": "4AQZ2", + "video_file": "Charades/videos/4AQZ2.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'one person puts groceries in a small refrigerator' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <0.4> to <14.8>." + } + ] + }, + { + "video_id": "1ELWC", + "question_id": "1ELWC", + "video_file": "Charades/videos/1ELWC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person runs through the bathroom door'." + }, + { + "from": "gpt", + "value": "From <22.3> to <28.0>." + } + ] + }, + { + "video_id": "1ELWC", + "question_id": "1ELWC", + "video_file": "Charades/videos/1ELWC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person runs away' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <22.3> to <28.0>." + } + ] + }, + { + "video_id": "1ELWC", + "question_id": "1ELWC", + "video_file": "Charades/videos/1ELWC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person closes medicine cabinet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.6> to <9.4>." + } + ] + }, + { + "video_id": "1ELWC", + "question_id": "1ELWC", + "video_file": "Charades/videos/1ELWC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person opens medicine cabinet door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "1ELWC", + "question_id": "1ELWC", + "video_file": "Charades/videos/1ELWC.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is opening a cabinet in the bathroom'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <4.7>." + } + ] + }, + { + "video_id": "E8468", + "question_id": "E8468", + "video_file": "Charades/videos/E8468.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person pours coffee in a cup' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <22.1> to <32.0>." + } + ] + }, + { + "video_id": "E8468", + "question_id": "E8468", + "video_file": "Charades/videos/E8468.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person closes the door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <20.7>." + } + ] + }, + { + "video_id": "E8468", + "question_id": "E8468", + "video_file": "Charades/videos/E8468.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they pour a cup of coffee' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <22.1> to <32.0>." + } + ] + }, + { + "video_id": "E8468", + "question_id": "E8468", + "video_file": "Charades/videos/E8468.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is holding a picture'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <0.0> to <12.9>." + } + ] + }, + { + "video_id": "E8468", + "question_id": "E8468", + "video_file": "Charades/videos/E8468.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person close the door' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.5> to <20.7>." + } + ] + }, + { + "video_id": "HQUHP", + "question_id": "HQUHP", + "video_file": "Charades/videos/HQUHP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns on the light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <16.7>." + } + ] + }, + { + "video_id": "HQUHP", + "question_id": "HQUHP", + "video_file": "Charades/videos/HQUHP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turned on a light' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <9.7> to <16.7>." + } + ] + }, + { + "video_id": "HQUHP", + "question_id": "HQUHP", + "video_file": "Charades/videos/HQUHP.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person turns on a light'." + }, + { + "from": "gpt", + "value": "From <9.7> to <16.7>." + } + ] + }, + { + "video_id": "MZOPX", + "question_id": "MZOPX", + "video_file": "Charades/videos/MZOPX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person washes a glass they used earlier' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <23.6> to <37.0>." + } + ] + }, + { + "video_id": "MZOPX", + "question_id": "MZOPX", + "video_file": "Charades/videos/MZOPX.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person begins washing a glass' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <23.6> to <37.0>." + } + ] + }, + { + "video_id": "Q6IU9", + "question_id": "Q6IU9", + "video_file": "Charades/videos/Q6IU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person they put the sandwich back on the ledge' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <28.6> to <33.0>." + } + ] + }, + { + "video_id": "Q6IU9", + "question_id": "Q6IU9", + "video_file": "Charades/videos/Q6IU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they close the laptop' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <12.0> to <18.3>." + } + ] + }, + { + "video_id": "Q6IU9", + "question_id": "Q6IU9", + "video_file": "Charades/videos/Q6IU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person eat it' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <19.3> to <27.3>." + } + ] + }, + { + "video_id": "Q6IU9", + "question_id": "Q6IU9", + "video_file": "Charades/videos/Q6IU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person which they began eating' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <19.3> to <27.3>." + } + ] + }, + { + "video_id": "Q6IU9", + "question_id": "Q6IU9", + "video_file": "Charades/videos/Q6IU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person takes out a sandwich'." + }, + { + "from": "gpt", + "value": "From <15.1> to <22.3>." + } + ] + }, + { + "video_id": "Q6IU9", + "question_id": "Q6IU9", + "video_file": "Charades/videos/Q6IU9.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person closes their laptop' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <12.0> to <18.3>." + } + ] + }, + { + "video_id": "7UV4O", + "question_id": "7UV4O", + "video_file": "Charades/videos/7UV4O.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person puts two pillows on a bed' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <5.4> to <13.5>." + } + ] + }, + { + "video_id": "DZNYK", + "question_id": "DZNYK", + "video_file": "Charades/videos/DZNYK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nDetect and report the start and end timestamps of the video segment that semantically matches the given textual query 'person laughs at what is inside'." + }, + { + "from": "gpt", + "value": "From <17.8> to <29.0>." + } + ] + }, + { + "video_id": "DZNYK", + "question_id": "DZNYK", + "video_file": "Charades/videos/DZNYK.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'the person opens the box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <11.9> to <21.6>." + } + ] + }, + { + "video_id": "DM2HW", + "question_id": "DM2HW", + "video_file": "Charades/videos/DM2HW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person takes a box' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.0> to <17.9>." + } + ] + }, + { + "video_id": "DM2HW", + "question_id": "DM2HW", + "video_file": "Charades/videos/DM2HW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'person they take a box' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <10.0> to <17.9>." + } + ] + }, + { + "video_id": "DM2HW", + "question_id": "DM2HW", + "video_file": "Charades/videos/DM2HW.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'person turn off the light' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <13.6> to <22.0>." + } + ] + }, + { + "video_id": "JPUUF", + "question_id": "JPUUF", + "video_file": "Charades/videos/JPUUF.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is in a kitchen holding some dishes'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <1.5> to <16.7>." + } + ] + }, + { + "video_id": "2H9YB", + "question_id": "2H9YB", + "video_file": "Charades/videos/2H9YB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nThe given natural language query 'a person walks to a doorway' is semantically aligned with a video moment, please give the start time and end time of the video moment." + }, + { + "from": "gpt", + "value": "From <20.8> to <25.8>." + } + ] + }, + { + "video_id": "2H9YB", + "question_id": "2H9YB", + "video_file": "Charades/videos/2H9YB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'a person is fixing a light'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <21.6>." + } + ] + }, + { + "video_id": "2H9YB", + "question_id": "2H9YB", + "video_file": "Charades/videos/2H9YB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nGive you a textual query: 'person they open another door'. When does the described content occur in the video? Please return the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <26.9> to <31.6>." + } + ] + }, + { + "video_id": "2H9YB", + "question_id": "2H9YB", + "video_file": "Charades/videos/2H9YB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person turns on the light' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <16.7> to <21.6>." + } + ] + }, + { + "video_id": "2H9YB", + "question_id": "2H9YB", + "video_file": "Charades/videos/2H9YB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person turns a light on above the door' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <21.6>." + } + ] + }, + { + "video_id": "2H9YB", + "question_id": "2H9YB", + "video_file": "Charades/videos/2H9YB.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocate the visual content mentioned in the text query 'a person replaces a lightbulb above a doorway' within the video, including start and end timestamps." + }, + { + "from": "gpt", + "value": "From <16.7> to <21.6>." + } + ] + }, + { + "video_id": "O0R1A", + "question_id": "O0R1A", + "video_file": "Charades/videos/O0R1A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'a person is throwing clothes into a laundry basket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.1> to <8.8>." + } + ] + }, + { + "video_id": "O0R1A", + "question_id": "O0R1A", + "video_file": "Charades/videos/O0R1A.mp4", + "conversation": [ + { + "from": "human", + "value": "\nFind the video segment that corresponds to the given textual query 'person throws clothes in bucket' and determine its start and end positions." + }, + { + "from": "gpt", + "value": "From <1.1> to <8.8>." + } + ] + }, + { + "video_id": "PRVP0", + "question_id": "PRVP0", + "video_file": "Charades/videos/PRVP0.mp4", + "conversation": [ + { + "from": "human", + "value": "\nLocalize the visual content described by the given textual query 'person walks through the doorway' in the video, and output the start and end timestamps." + }, + { + "from": "gpt", + "value": "From <10.8> to <17.1>." + } + ] + } +] \ No newline at end of file