image
stringlengths
15
15
id
stringlengths
8
12
conversations
stringlengths
79
268
00000007292.jpg
7292_evjvqa
[{"from": "human", "value": "which hand is the girl dressed in purple using to hold the grocery bag?\n<image>"}, {"from": "gpt", "value": "the girl dressed in purple is using her right hand to hold the grocery bag."}]
00000007296.jpg
7296_evjvqa
[{"from": "human", "value": "which area is in front of the reception area?\n<image>"}, {"from": "gpt", "value": "the washing machine area. "}]
00000007298.jpg
7298_evjvqa
[{"from": "human", "value": "how many roofs does the temple have?\n<image>"}, {"from": "gpt", "value": "the temple has two roofs."}]
00000007298.jpg
7298_evjvqa
[{"from": "human", "value": "where are they standing for the photograph?\n<image>"}, {"from": "gpt", "value": "they are standing in front of the temple."}]
00000007298.jpg
7298_evjvqa
[{"from": "human", "value": "how many flights of stairs that leads to the temple?\n<image>"}, {"from": "gpt", "value": "there are two flights of stairs leading to the temple. "}]
00000004603.jpg
4603_evjvqa
[{"from": "human", "value": "what is the baby holding in her hand?\n<image>"}, {"from": "gpt", "value": "holding a bag of vegetables in her hand"}]
00000004603.jpg
4603_evjvqa
[{"from": "human", "value": "what fruit does the stand sell to the left of the kid?\n<image>"}, {"from": "gpt", "value": "on the left is the bitter gourd stand"}]
00000004603.jpg
4603_evjvqa
[{"from": "human", "value": "what color shirt is the man behind the child wearing?\n<image>"}, {"from": "gpt", "value": "wearing a black t-shirt"}]
00000004604.jpg
4604_evjvqa
[{"from": "human", "value": "what color shirt does the staff member wear?\n<image>"}, {"from": "gpt", "value": "wear a black shirt"}]
00000004604.jpg
4604_evjvqa
[{"from": "human", "value": "what is the employee pointing at?\n<image>"}, {"from": "gpt", "value": "pointing at the gift packages"}]
00000004604.jpg
4604_evjvqa
[{"from": "human", "value": "which hand is the employee spreading ahead?\n<image>"}, {"from": "gpt", "value": "spreading her left hand out in front of her"}]
00000004606.jpg
4606_evjvqa
[{"from": "human", "value": "where is everyone?\n<image>"}, {"from": "gpt", "value": "in the park"}]
00000004606.jpg
4606_evjvqa
[{"from": "human", "value": "what are the children doing?\n<image>"}, {"from": "gpt", "value": "playing together"}]
00000004606.jpg
4606_evjvqa
[{"from": "human", "value": "how many children wear red tops?\n<image>"}, {"from": "gpt", "value": "there are two kids wearing red tops"}]
00000004608.jpg
4608_evjvqa
[{"from": "human", "value": "on which side of the wooden slats is the birdcage pinned?\n<image>"}, {"from": "gpt", "value": "the birdcage is pinned on the right side of the wooden slats"}]
00000004608.jpg
4608_evjvqa
[{"from": "human", "value": "how many lamps are on the ground?\n<image>"}, {"from": "gpt", "value": "there are two lamps on the ground"}]
00000004608.jpg
4608_evjvqa
[{"from": "human", "value": "how many text lines are on the wooden slats?\n<image>"}, {"from": "gpt", "value": "there are four lines"}]
00000004609.jpg
4609_evjvqa
[{"from": "human", "value": "what color shirt is the woman holding the basket wearing?\n<image>"}, {"from": "gpt", "value": "wearing a blue t-shirt"}]
00000004609.jpg
4609_evjvqa
[{"from": "human", "value": "which hand is the woman holding the phone in?\n<image>"}, {"from": "gpt", "value": "holding it in the left hand"}]
00000004609.jpg
4609_evjvqa
[{"from": "human", "value": "what is in front of two women?\n<image>"}, {"from": "gpt", "value": "is a tray of sponge cake"}]
00000004610.jpg
4610_evjvqa
[{"from": "human", "value": "how many floors does the building have?\n<image>"}, {"from": "gpt", "value": "the building has three floors"}]
00000004610.jpg
4610_evjvqa
[{"from": "human", "value": "how many vehicles are parked by the curb?\n<image>"}, {"from": "gpt", "value": "there are two vehicles parked by the curb"}]
00000004610.jpg
4610_evjvqa
[{"from": "human", "value": "how many trash cans are there off the pavement?\n<image>"}, {"from": "gpt", "value": "there are two trash cans off the pavement"}]
00000004611.jpg
4611_evjvqa
[{"from": "human", "value": "how many people are on the right side of the road?\n<image>"}, {"from": "gpt", "value": "there are two people walking on the right roadside"}]
00000004611.jpg
4611_evjvqa
[{"from": "human", "value": "which side of the man does the woman walk?\n<image>"}, {"from": "gpt", "value": "walk behind the man"}]
00000004613.jpg
4613_evjvqa
[{"from": "human", "value": "what is the medical staff member doing?\n<image>"}, {"from": "gpt", "value": "injecting medicine into the syringe"}]
00000004613.jpg
4613_evjvqa
[{"from": "human", "value": "which hand is the medical staff holding the syringe in?\n<image>"}, {"from": "gpt", "value": "holding in her right hand"}]
00000004613.jpg
4613_evjvqa
[{"from": "human", "value": "what color is the medical staff's protective gear?\n<image>"}, {"from": "gpt", "value": "the protective gear is blue"}]
00000004615.jpg
4615_evjvqa
[{"from": "human", "value": "what are people doing?\n<image>"}, {"from": "gpt", "value": "shopping at the supermarket"}]
00000004615.jpg
4615_evjvqa
[{"from": "human", "value": "what is the sitting woman doing?\n<image>"}, {"from": "gpt", "value": "viewing an item of goods"}]
00000004615.jpg
4615_evjvqa
[{"from": "human", "value": "what color shirt is the woman wearing behind the sitting person?\n<image>"}, {"from": "gpt", "value": "wearing a white t-shirt"}]
00000004616.jpg
4616_evjvqa
[{"from": "human", "value": "what time of the day is the scene?\n<image>"}, {"from": "gpt", "value": "is currently afternoon"}]
00000004616.jpg
4616_evjvqa
[{"from": "human", "value": "how many bags is the woman crossing the street bringing?\n<image>"}, {"from": "gpt", "value": "bringing two bags"}]
00000004617.jpg
4617_evjvqa
[{"from": "human", "value": "what store is this?\n<image>"}, {"from": "gpt", "value": "this is a cellphone shop"}]
00000004617.jpg
4617_evjvqa
[{"from": "human", "value": "what color uniform does the staff wear?\n<image>"}, {"from": "gpt", "value": "wear a black uniform"}]
00000004618.jpg
4618_evjvqa
[{"from": "human", "value": "how many women are smiling?\n<image>"}, {"from": "gpt", "value": "there are two women smiling"}]
00000004618.jpg
4618_evjvqa
[{"from": "human", "value": "where is everyone?\n<image>"}, {"from": "gpt", "value": "in a market"}]
00000004618.jpg
4618_evjvqa
[{"from": "human", "value": "what are the baskets next to the woman?\n<image>"}, {"from": "gpt", "value": "baskets of dragon fruit"}]
00000004619.jpg
4619_evjvqa
[{"from": "human", "value": "how many people are dressed as santa claus?\n<image>"}, {"from": "gpt", "value": "there are two people dressed as santa claus"}]
00000004619.jpg
4619_evjvqa
[{"from": "human", "value": "what game is everyone playing?\n<image>"}, {"from": "gpt", "value": "playing snowboarding"}]
00000004619.jpg
4619_evjvqa
[{"from": "human", "value": "where is everyone playing?\n<image>"}, {"from": "gpt", "value": "playing in an entertainment centre"}]
00000004620.jpg
4620_evjvqa
[{"from": "human", "value": "what time is the neighborhood?\n<image>"}, {"from": "gpt", "value": "it is twilight"}]
00000004620.jpg
4620_evjvqa
[{"from": "human", "value": "what color are the houses painted?\n<image>"}, {"from": "gpt", "value": "painted yellow"}]
00000004620.jpg
4620_evjvqa
[{"from": "human", "value": "what color shirt is the person walking on the street wearing?\n<image>"}, {"from": "gpt", "value": "wearing a yellow shirt"}]
00000004623.jpg
4623_evjvqa
[{"from": "human", "value": "how many mannequins are placed in front of the shop?\n<image>"}, {"from": "gpt", "value": "there are five mannequins placed in front of the shop"}]
00000004624.jpg
4624_evjvqa
[{"from": "human", "value": "what is the girl doing?\n<image>"}, {"from": "gpt", "value": "looking at antiquities"}]
00000004624.jpg
4624_evjvqa
[{"from": "human", "value": "where is the girl?\n<image>"}, {"from": "gpt", "value": "in a museum"}]
00000004624.jpg
4624_evjvqa
[{"from": "human", "value": "what objects are displayed in the room the girl standing in?\n<image>"}, {"from": "gpt", "value": "displays musical instruments and traditional costumes"}]
00000004626.jpg
4626_evjvqa
[{"from": "human", "value": "what is the woman doing on the curb?\n<image>"}, {"from": "gpt", "value": "cooking a pot of food"}]
00000004626.jpg
4626_evjvqa
[{"from": "human", "value": "how many people are leading bicycles on the road?\n<image>"}, {"from": "gpt", "value": "there is a person leading a bicycle on the road"}]
00000004626.jpg
4626_evjvqa
[{"from": "human", "value": "what is hanging on the bike?\n<image>"}, {"from": "gpt", "value": "big bags are hung on the bike"}]
00000004627.jpg
4627_evjvqa
[{"from": "human", "value": "how many flags were raised in front of the shrine?\n<image>"}, {"from": "gpt", "value": "there are four flags raised in front of the shrine"}]
00000004627.jpg
4627_evjvqa
[{"from": "human", "value": "what animal is carved on the roof of the shrine?\n<image>"}, {"from": "gpt", "value": "the roof is carved with dragon images"}]
00000004627.jpg
4627_evjvqa
[{"from": "human", "value": "how many pots are placed in front of the building?\n<image>"}, {"from": "gpt", "value": "there is a pot placed in front of the building"}]
00000004629.jpg
4629_evjvqa
[{"from": "human", "value": "where is this?\n<image>"}, {"from": "gpt", "value": "this is a flower market"}]
00000004629.jpg
4629_evjvqa
[{"from": "human", "value": "what do the motorbikes carry?\n<image>"}, {"from": "gpt", "value": "the motorcycles carry bouquets of flowers"}]
00000004629.jpg
4629_evjvqa
[{"from": "human", "value": "what color are the flowers transported to the market?\n<image>"}, {"from": "gpt", "value": "flowers are yellow"}]
00000004630.jpg
4630_evjvqa
[{"from": "human", "value": "how many people are shopping in the store?\n<image>"}, {"from": "gpt", "value": "there is a person shopping"}]
00000004630.jpg
4630_evjvqa
[{"from": "human", "value": "where is the condiment shelf located?\n<image>"}, {"from": "gpt", "value": "placed in the center next to the dry food shelf"}]
00000004631.jpg
4631_evjvqa
[{"from": "human", "value": "which hand is the girl holding the conical hat?\n<image>"}, {"from": "gpt", "value": "holding the hat in her right hand"}]
00000004631.jpg
4631_evjvqa
[{"from": "human", "value": "what is behind the girl?\n<image>"}, {"from": "gpt", "value": "behind the girl is a bridge"}]
00000004631.jpg
4631_evjvqa
[{"from": "human", "value": "what color is the girl wearing?\n<image>"}, {"from": "gpt", "value": "wearing a yellow top"}]
00000004634.jpg
4634_evjvqa
[{"from": "human", "value": "wow many white vehicles are going in the opposite direction of the young lady?\n<image>"}, {"from": "gpt", "value": "there are two white vehicles"}]
00000004634.jpg
4634_evjvqa
[{"from": "human", "value": "what color is the girl wearing?\n<image>"}, {"from": "gpt", "value": "wearing a white top"}]
00000004634.jpg
4634_evjvqa
[{"from": "human", "value": "what time of the day is the scene?\n<image>"}, {"from": "gpt", "value": "it is noon"}]
00000004636.jpg
4636_evjvqa
[{"from": "human", "value": "how many people are wearing green?\n<image>"}, {"from": "gpt", "value": "there are four people dressed in green"}]
00000004636.jpg
4636_evjvqa
[{"from": "human", "value": "what are the people wearing behind the green people?\n<image>"}, {"from": "gpt", "value": "behind are people wearing blue"}]
00000004636.jpg
4636_evjvqa
[{"from": "human", "value": "how many people are there in each row?\n<image>"}, {"from": "gpt", "value": "there are two people in each row"}]
00000004638.jpg
4638_evjvqa
[{"from": "human", "value": "what color hat is the woman wearing the handbag wearing?\n<image>"}, {"from": "gpt", "value": "wearing a yellow hat"}]
00000004638.jpg
4638_evjvqa
[{"from": "human", "value": "where are people standing?\n<image>"}, {"from": "gpt", "value": "standing in a tet fair"}]
00000004639.jpg
4639_evjvqa
[{"from": "human", "value": "where was the photo taken?\n<image>"}, {"from": "gpt", "value": "taken in a supermarket"}]
00000004639.jpg
4639_evjvqa
[{"from": "human", "value": "what products do the stalls on the right sell?\n<image>"}, {"from": "gpt", "value": "selling snacks"}]
00000004643.jpg
4643_evjvqa
[{"from": "human", "value": "how many chairs are there in front of the store?\n<image>"}, {"from": "gpt", "value": "there are five chairs in front of the store"}]
00000004643.jpg
4643_evjvqa
[{"from": "human", "value": "how many people are in the store?\n<image>"}, {"from": "gpt", "value": "there's a person in the store"}]
00000004644.jpg
4644_evjvqa
[{"from": "human", "value": "what are people doing?\n<image>"}, {"from": "gpt", "value": "cutting the ribbon to inaugurate the supermarket"}]
00000004644.jpg
4644_evjvqa
[{"from": "human", "value": "how many girls are there holding ribbons?\n<image>"}, {"from": "gpt", "value": "there are two girls holding ribbons"}]
00000004646.jpg
4646_evjvqa
[{"from": "human", "value": "what area is this?\n<image>"}, {"from": "gpt", "value": "this is a mausoleum"}]
00000004646.jpg
4646_evjvqa
[{"from": "human", "value": "how many horse statues are there?\n<image>"}, {"from": "gpt", "value": "there is one statue of a horse"}]
00000004646.jpg
4646_evjvqa
[{"from": "human", "value": "how many functionary figurines are there?\n<image>"}, {"from": "gpt", "value": "there are four statues of functionaries"}]
00000004647.jpg
4647_evjvqa
[{"from": "human", "value": "what is the store behind the blue pants man?\n<image>"}, {"from": "gpt", "value": "christmas gift shop"}]
00000004647.jpg
4647_evjvqa
[{"from": "human", "value": "what color are the costumes for christmas?\n<image>"}, {"from": "gpt", "value": "the costumes are red"}]
00000004648.jpg
4648_evjvqa
[{"from": "human", "value": "what color is the mannequin wearing?\n<image>"}, {"from": "gpt", "value": "wearing a white shirt"}]
00000004648.jpg
4648_evjvqa
[{"from": "human", "value": "what is the mannequin holding?\n<image>"}, {"from": "gpt", "value": "holding a billboard"}]
00000004649.jpg
4649_evjvqa
[{"from": "human", "value": "how is the traffic in the neighborhood?\n<image>"}, {"from": "gpt", "value": "the neighborhood is pretty deserted"}]
00000004649.jpg
4649_evjvqa
[{"from": "human", "value": "how many flag rows are hung across the road?\n<image>"}, {"from": "gpt", "value": "there is a flag row hanging across the street"}]
00000004649.jpg
4649_evjvqa
[{"from": "human", "value": "what time of day is the neighborhood view?\n<image>"}, {"from": "gpt", "value": "the neighborhood is in the late afternoon"}]
00000004650.jpg
4650_evjvqa
[{"from": "human", "value": "which side of the tall man are the group of people sitting?\n<image>"}, {"from": "gpt", "value": "the group of people are sitting to the left of the man"}]
00000004650.jpg
4650_evjvqa
[{"from": "human", "value": "what is the symbol stretched across the road?\n<image>"}, {"from": "gpt", "value": "the symbol stretched across the road is the symbol of quoc tu giam temple of literature in the middle of the lotus branches"}]
00000004651.jpg
4651_evjvqa
[{"from": "human", "value": "how many people are wearing black?\n<image>"}, {"from": "gpt", "value": "there are four people wearing black"}]
00000004651.jpg
4651_evjvqa
[{"from": "human", "value": "what color top is the woman wearing on the left of the three in black?\n<image>"}, {"from": "gpt", "value": "wearing a blue t-shirt"}]
00000004651.jpg
4651_evjvqa
[{"from": "human", "value": "which side of the three people in black is the woman in blue sitting on?\n<image>"}, {"from": "gpt", "value": "sitting to the left of three people in black"}]
00000004652.jpg
4652_evjvqa
[{"from": "human", "value": "what area is this?\n<image>"}, {"from": "gpt", "value": "this is a spontaneous market"}]
00000004652.jpg
4652_evjvqa
[{"from": "human", "value": "how is the market scene?\n<image>"}, {"from": "gpt", "value": "the market is very crowded"}]
00000004652.jpg
4652_evjvqa
[{"from": "human", "value": "what color is the trash can?\n<image>"}, {"from": "gpt", "value": "the trash can is blue"}]
00000004654.jpg
4654_evjvqa
[{"from": "human", "value": "what color shirt is the man wearing?\n<image>"}, {"from": "gpt", "value": "wearing a white shirt"}]
00000004654.jpg
4654_evjvqa
[{"from": "human", "value": "what is the man doing?\n<image>"}, {"from": "gpt", "value": "sitting guarding the isolation area"}]
00000004654.jpg
4654_evjvqa
[{"from": "human", "value": "how many glasses of water are there at the man's feet?\n<image>"}, {"from": "gpt", "value": "there are two glasses of water at the man's feet"}]
00000004659.jpg
4659_evjvqa
[{"from": "human", "value": "where are people standing?\n<image>"}, {"from": "gpt", "value": "standing in a supermarket"}]
00000004661.jpg
4661_evjvqa
[{"from": "human", "value": "what time of day is the view?\n<image>"}, {"from": "gpt", "value": "it is evening"}]
00000004661.jpg
4661_evjvqa
[{"from": "human", "value": "which direction is the group going?\n<image>"}, {"from": "gpt", "value": "walking outwards the market"}]