image
stringlengths 15
15
| id
stringlengths 8
12
| conversations
stringlengths 79
268
|
---|---|---|
00000007292.jpg | 7292_evjvqa | [{"from": "human", "value": "which hand is the girl dressed in purple using to hold the grocery bag?\n<image>"}, {"from": "gpt", "value": "the girl dressed in purple is using her right hand to hold the grocery bag."}] |
00000007296.jpg | 7296_evjvqa | [{"from": "human", "value": "which area is in front of the reception area?\n<image>"}, {"from": "gpt", "value": "the washing machine area. "}] |
00000007298.jpg | 7298_evjvqa | [{"from": "human", "value": "how many roofs does the temple have?\n<image>"}, {"from": "gpt", "value": "the temple has two roofs."}] |
00000007298.jpg | 7298_evjvqa | [{"from": "human", "value": "where are they standing for the photograph?\n<image>"}, {"from": "gpt", "value": "they are standing in front of the temple."}] |
00000007298.jpg | 7298_evjvqa | [{"from": "human", "value": "how many flights of stairs that leads to the temple?\n<image>"}, {"from": "gpt", "value": "there are two flights of stairs leading to the temple. "}] |
00000004603.jpg | 4603_evjvqa | [{"from": "human", "value": "what is the baby holding in her hand?\n<image>"}, {"from": "gpt", "value": "holding a bag of vegetables in her hand"}] |
00000004603.jpg | 4603_evjvqa | [{"from": "human", "value": "what fruit does the stand sell to the left of the kid?\n<image>"}, {"from": "gpt", "value": "on the left is the bitter gourd stand"}] |
00000004603.jpg | 4603_evjvqa | [{"from": "human", "value": "what color shirt is the man behind the child wearing?\n<image>"}, {"from": "gpt", "value": "wearing a black t-shirt"}] |
00000004604.jpg | 4604_evjvqa | [{"from": "human", "value": "what color shirt does the staff member wear?\n<image>"}, {"from": "gpt", "value": "wear a black shirt"}] |
00000004604.jpg | 4604_evjvqa | [{"from": "human", "value": "what is the employee pointing at?\n<image>"}, {"from": "gpt", "value": "pointing at the gift packages"}] |
00000004604.jpg | 4604_evjvqa | [{"from": "human", "value": "which hand is the employee spreading ahead?\n<image>"}, {"from": "gpt", "value": "spreading her left hand out in front of her"}] |
00000004606.jpg | 4606_evjvqa | [{"from": "human", "value": "where is everyone?\n<image>"}, {"from": "gpt", "value": "in the park"}] |
00000004606.jpg | 4606_evjvqa | [{"from": "human", "value": "what are the children doing?\n<image>"}, {"from": "gpt", "value": "playing together"}] |
00000004606.jpg | 4606_evjvqa | [{"from": "human", "value": "how many children wear red tops?\n<image>"}, {"from": "gpt", "value": "there are two kids wearing red tops"}] |
00000004608.jpg | 4608_evjvqa | [{"from": "human", "value": "on which side of the wooden slats is the birdcage pinned?\n<image>"}, {"from": "gpt", "value": "the birdcage is pinned on the right side of the wooden slats"}] |
00000004608.jpg | 4608_evjvqa | [{"from": "human", "value": "how many lamps are on the ground?\n<image>"}, {"from": "gpt", "value": "there are two lamps on the ground"}] |
00000004608.jpg | 4608_evjvqa | [{"from": "human", "value": "how many text lines are on the wooden slats?\n<image>"}, {"from": "gpt", "value": "there are four lines"}] |
00000004609.jpg | 4609_evjvqa | [{"from": "human", "value": "what color shirt is the woman holding the basket wearing?\n<image>"}, {"from": "gpt", "value": "wearing a blue t-shirt"}] |
00000004609.jpg | 4609_evjvqa | [{"from": "human", "value": "which hand is the woman holding the phone in?\n<image>"}, {"from": "gpt", "value": "holding it in the left hand"}] |
00000004609.jpg | 4609_evjvqa | [{"from": "human", "value": "what is in front of two women?\n<image>"}, {"from": "gpt", "value": "is a tray of sponge cake"}] |
00000004610.jpg | 4610_evjvqa | [{"from": "human", "value": "how many floors does the building have?\n<image>"}, {"from": "gpt", "value": "the building has three floors"}] |
00000004610.jpg | 4610_evjvqa | [{"from": "human", "value": "how many vehicles are parked by the curb?\n<image>"}, {"from": "gpt", "value": "there are two vehicles parked by the curb"}] |
00000004610.jpg | 4610_evjvqa | [{"from": "human", "value": "how many trash cans are there off the pavement?\n<image>"}, {"from": "gpt", "value": "there are two trash cans off the pavement"}] |
00000004611.jpg | 4611_evjvqa | [{"from": "human", "value": "how many people are on the right side of the road?\n<image>"}, {"from": "gpt", "value": "there are two people walking on the right roadside"}] |
00000004611.jpg | 4611_evjvqa | [{"from": "human", "value": "which side of the man does the woman walk?\n<image>"}, {"from": "gpt", "value": "walk behind the man"}] |
00000004613.jpg | 4613_evjvqa | [{"from": "human", "value": "what is the medical staff member doing?\n<image>"}, {"from": "gpt", "value": "injecting medicine into the syringe"}] |
00000004613.jpg | 4613_evjvqa | [{"from": "human", "value": "which hand is the medical staff holding the syringe in?\n<image>"}, {"from": "gpt", "value": "holding in her right hand"}] |
00000004613.jpg | 4613_evjvqa | [{"from": "human", "value": "what color is the medical staff's protective gear?\n<image>"}, {"from": "gpt", "value": "the protective gear is blue"}] |
00000004615.jpg | 4615_evjvqa | [{"from": "human", "value": "what are people doing?\n<image>"}, {"from": "gpt", "value": "shopping at the supermarket"}] |
00000004615.jpg | 4615_evjvqa | [{"from": "human", "value": "what is the sitting woman doing?\n<image>"}, {"from": "gpt", "value": "viewing an item of goods"}] |
00000004615.jpg | 4615_evjvqa | [{"from": "human", "value": "what color shirt is the woman wearing behind the sitting person?\n<image>"}, {"from": "gpt", "value": "wearing a white t-shirt"}] |
00000004616.jpg | 4616_evjvqa | [{"from": "human", "value": "what time of the day is the scene?\n<image>"}, {"from": "gpt", "value": "is currently afternoon"}] |
00000004616.jpg | 4616_evjvqa | [{"from": "human", "value": "how many bags is the woman crossing the street bringing?\n<image>"}, {"from": "gpt", "value": "bringing two bags"}] |
00000004617.jpg | 4617_evjvqa | [{"from": "human", "value": "what store is this?\n<image>"}, {"from": "gpt", "value": "this is a cellphone shop"}] |
00000004617.jpg | 4617_evjvqa | [{"from": "human", "value": "what color uniform does the staff wear?\n<image>"}, {"from": "gpt", "value": "wear a black uniform"}] |
00000004618.jpg | 4618_evjvqa | [{"from": "human", "value": "how many women are smiling?\n<image>"}, {"from": "gpt", "value": "there are two women smiling"}] |
00000004618.jpg | 4618_evjvqa | [{"from": "human", "value": "where is everyone?\n<image>"}, {"from": "gpt", "value": "in a market"}] |
00000004618.jpg | 4618_evjvqa | [{"from": "human", "value": "what are the baskets next to the woman?\n<image>"}, {"from": "gpt", "value": "baskets of dragon fruit"}] |
00000004619.jpg | 4619_evjvqa | [{"from": "human", "value": "how many people are dressed as santa claus?\n<image>"}, {"from": "gpt", "value": "there are two people dressed as santa claus"}] |
00000004619.jpg | 4619_evjvqa | [{"from": "human", "value": "what game is everyone playing?\n<image>"}, {"from": "gpt", "value": "playing snowboarding"}] |
00000004619.jpg | 4619_evjvqa | [{"from": "human", "value": "where is everyone playing?\n<image>"}, {"from": "gpt", "value": "playing in an entertainment centre"}] |
00000004620.jpg | 4620_evjvqa | [{"from": "human", "value": "what time is the neighborhood?\n<image>"}, {"from": "gpt", "value": "it is twilight"}] |
00000004620.jpg | 4620_evjvqa | [{"from": "human", "value": "what color are the houses painted?\n<image>"}, {"from": "gpt", "value": "painted yellow"}] |
00000004620.jpg | 4620_evjvqa | [{"from": "human", "value": "what color shirt is the person walking on the street wearing?\n<image>"}, {"from": "gpt", "value": "wearing a yellow shirt"}] |
00000004623.jpg | 4623_evjvqa | [{"from": "human", "value": "how many mannequins are placed in front of the shop?\n<image>"}, {"from": "gpt", "value": "there are five mannequins placed in front of the shop"}] |
00000004624.jpg | 4624_evjvqa | [{"from": "human", "value": "what is the girl doing?\n<image>"}, {"from": "gpt", "value": "looking at antiquities"}] |
00000004624.jpg | 4624_evjvqa | [{"from": "human", "value": "where is the girl?\n<image>"}, {"from": "gpt", "value": "in a museum"}] |
00000004624.jpg | 4624_evjvqa | [{"from": "human", "value": "what objects are displayed in the room the girl standing in?\n<image>"}, {"from": "gpt", "value": "displays musical instruments and traditional costumes"}] |
00000004626.jpg | 4626_evjvqa | [{"from": "human", "value": "what is the woman doing on the curb?\n<image>"}, {"from": "gpt", "value": "cooking a pot of food"}] |
00000004626.jpg | 4626_evjvqa | [{"from": "human", "value": "how many people are leading bicycles on the road?\n<image>"}, {"from": "gpt", "value": "there is a person leading a bicycle on the road"}] |
00000004626.jpg | 4626_evjvqa | [{"from": "human", "value": "what is hanging on the bike?\n<image>"}, {"from": "gpt", "value": "big bags are hung on the bike"}] |
00000004627.jpg | 4627_evjvqa | [{"from": "human", "value": "how many flags were raised in front of the shrine?\n<image>"}, {"from": "gpt", "value": "there are four flags raised in front of the shrine"}] |
00000004627.jpg | 4627_evjvqa | [{"from": "human", "value": "what animal is carved on the roof of the shrine?\n<image>"}, {"from": "gpt", "value": "the roof is carved with dragon images"}] |
00000004627.jpg | 4627_evjvqa | [{"from": "human", "value": "how many pots are placed in front of the building?\n<image>"}, {"from": "gpt", "value": "there is a pot placed in front of the building"}] |
00000004629.jpg | 4629_evjvqa | [{"from": "human", "value": "where is this?\n<image>"}, {"from": "gpt", "value": "this is a flower market"}] |
00000004629.jpg | 4629_evjvqa | [{"from": "human", "value": "what do the motorbikes carry?\n<image>"}, {"from": "gpt", "value": "the motorcycles carry bouquets of flowers"}] |
00000004629.jpg | 4629_evjvqa | [{"from": "human", "value": "what color are the flowers transported to the market?\n<image>"}, {"from": "gpt", "value": "flowers are yellow"}] |
00000004630.jpg | 4630_evjvqa | [{"from": "human", "value": "how many people are shopping in the store?\n<image>"}, {"from": "gpt", "value": "there is a person shopping"}] |
00000004630.jpg | 4630_evjvqa | [{"from": "human", "value": "where is the condiment shelf located?\n<image>"}, {"from": "gpt", "value": "placed in the center next to the dry food shelf"}] |
00000004631.jpg | 4631_evjvqa | [{"from": "human", "value": "which hand is the girl holding the conical hat?\n<image>"}, {"from": "gpt", "value": "holding the hat in her right hand"}] |
00000004631.jpg | 4631_evjvqa | [{"from": "human", "value": "what is behind the girl?\n<image>"}, {"from": "gpt", "value": "behind the girl is a bridge"}] |
00000004631.jpg | 4631_evjvqa | [{"from": "human", "value": "what color is the girl wearing?\n<image>"}, {"from": "gpt", "value": "wearing a yellow top"}] |
00000004634.jpg | 4634_evjvqa | [{"from": "human", "value": "wow many white vehicles are going in the opposite direction of the young lady?\n<image>"}, {"from": "gpt", "value": "there are two white vehicles"}] |
00000004634.jpg | 4634_evjvqa | [{"from": "human", "value": "what color is the girl wearing?\n<image>"}, {"from": "gpt", "value": "wearing a white top"}] |
00000004634.jpg | 4634_evjvqa | [{"from": "human", "value": "what time of the day is the scene?\n<image>"}, {"from": "gpt", "value": "it is noon"}] |
00000004636.jpg | 4636_evjvqa | [{"from": "human", "value": "how many people are wearing green?\n<image>"}, {"from": "gpt", "value": "there are four people dressed in green"}] |
00000004636.jpg | 4636_evjvqa | [{"from": "human", "value": "what are the people wearing behind the green people?\n<image>"}, {"from": "gpt", "value": "behind are people wearing blue"}] |
00000004636.jpg | 4636_evjvqa | [{"from": "human", "value": "how many people are there in each row?\n<image>"}, {"from": "gpt", "value": "there are two people in each row"}] |
00000004638.jpg | 4638_evjvqa | [{"from": "human", "value": "what color hat is the woman wearing the handbag wearing?\n<image>"}, {"from": "gpt", "value": "wearing a yellow hat"}] |
00000004638.jpg | 4638_evjvqa | [{"from": "human", "value": "where are people standing?\n<image>"}, {"from": "gpt", "value": "standing in a tet fair"}] |
00000004639.jpg | 4639_evjvqa | [{"from": "human", "value": "where was the photo taken?\n<image>"}, {"from": "gpt", "value": "taken in a supermarket"}] |
00000004639.jpg | 4639_evjvqa | [{"from": "human", "value": "what products do the stalls on the right sell?\n<image>"}, {"from": "gpt", "value": "selling snacks"}] |
00000004643.jpg | 4643_evjvqa | [{"from": "human", "value": "how many chairs are there in front of the store?\n<image>"}, {"from": "gpt", "value": "there are five chairs in front of the store"}] |
00000004643.jpg | 4643_evjvqa | [{"from": "human", "value": "how many people are in the store?\n<image>"}, {"from": "gpt", "value": "there's a person in the store"}] |
00000004644.jpg | 4644_evjvqa | [{"from": "human", "value": "what are people doing?\n<image>"}, {"from": "gpt", "value": "cutting the ribbon to inaugurate the supermarket"}] |
00000004644.jpg | 4644_evjvqa | [{"from": "human", "value": "how many girls are there holding ribbons?\n<image>"}, {"from": "gpt", "value": "there are two girls holding ribbons"}] |
00000004646.jpg | 4646_evjvqa | [{"from": "human", "value": "what area is this?\n<image>"}, {"from": "gpt", "value": "this is a mausoleum"}] |
00000004646.jpg | 4646_evjvqa | [{"from": "human", "value": "how many horse statues are there?\n<image>"}, {"from": "gpt", "value": "there is one statue of a horse"}] |
00000004646.jpg | 4646_evjvqa | [{"from": "human", "value": "how many functionary figurines are there?\n<image>"}, {"from": "gpt", "value": "there are four statues of functionaries"}] |
00000004647.jpg | 4647_evjvqa | [{"from": "human", "value": "what is the store behind the blue pants man?\n<image>"}, {"from": "gpt", "value": "christmas gift shop"}] |
00000004647.jpg | 4647_evjvqa | [{"from": "human", "value": "what color are the costumes for christmas?\n<image>"}, {"from": "gpt", "value": "the costumes are red"}] |
00000004648.jpg | 4648_evjvqa | [{"from": "human", "value": "what color is the mannequin wearing?\n<image>"}, {"from": "gpt", "value": "wearing a white shirt"}] |
00000004648.jpg | 4648_evjvqa | [{"from": "human", "value": "what is the mannequin holding?\n<image>"}, {"from": "gpt", "value": "holding a billboard"}] |
00000004649.jpg | 4649_evjvqa | [{"from": "human", "value": "how is the traffic in the neighborhood?\n<image>"}, {"from": "gpt", "value": "the neighborhood is pretty deserted"}] |
00000004649.jpg | 4649_evjvqa | [{"from": "human", "value": "how many flag rows are hung across the road?\n<image>"}, {"from": "gpt", "value": "there is a flag row hanging across the street"}] |
00000004649.jpg | 4649_evjvqa | [{"from": "human", "value": "what time of day is the neighborhood view?\n<image>"}, {"from": "gpt", "value": "the neighborhood is in the late afternoon"}] |
00000004650.jpg | 4650_evjvqa | [{"from": "human", "value": "which side of the tall man are the group of people sitting?\n<image>"}, {"from": "gpt", "value": "the group of people are sitting to the left of the man"}] |
00000004650.jpg | 4650_evjvqa | [{"from": "human", "value": "what is the symbol stretched across the road?\n<image>"}, {"from": "gpt", "value": "the symbol stretched across the road is the symbol of quoc tu giam temple of literature in the middle of the lotus branches"}] |
00000004651.jpg | 4651_evjvqa | [{"from": "human", "value": "how many people are wearing black?\n<image>"}, {"from": "gpt", "value": "there are four people wearing black"}] |
00000004651.jpg | 4651_evjvqa | [{"from": "human", "value": "what color top is the woman wearing on the left of the three in black?\n<image>"}, {"from": "gpt", "value": "wearing a blue t-shirt"}] |
00000004651.jpg | 4651_evjvqa | [{"from": "human", "value": "which side of the three people in black is the woman in blue sitting on?\n<image>"}, {"from": "gpt", "value": "sitting to the left of three people in black"}] |
00000004652.jpg | 4652_evjvqa | [{"from": "human", "value": "what area is this?\n<image>"}, {"from": "gpt", "value": "this is a spontaneous market"}] |
00000004652.jpg | 4652_evjvqa | [{"from": "human", "value": "how is the market scene?\n<image>"}, {"from": "gpt", "value": "the market is very crowded"}] |
00000004652.jpg | 4652_evjvqa | [{"from": "human", "value": "what color is the trash can?\n<image>"}, {"from": "gpt", "value": "the trash can is blue"}] |
00000004654.jpg | 4654_evjvqa | [{"from": "human", "value": "what color shirt is the man wearing?\n<image>"}, {"from": "gpt", "value": "wearing a white shirt"}] |
00000004654.jpg | 4654_evjvqa | [{"from": "human", "value": "what is the man doing?\n<image>"}, {"from": "gpt", "value": "sitting guarding the isolation area"}] |
00000004654.jpg | 4654_evjvqa | [{"from": "human", "value": "how many glasses of water are there at the man's feet?\n<image>"}, {"from": "gpt", "value": "there are two glasses of water at the man's feet"}] |
00000004659.jpg | 4659_evjvqa | [{"from": "human", "value": "where are people standing?\n<image>"}, {"from": "gpt", "value": "standing in a supermarket"}] |
00000004661.jpg | 4661_evjvqa | [{"from": "human", "value": "what time of day is the view?\n<image>"}, {"from": "gpt", "value": "it is evening"}] |
00000004661.jpg | 4661_evjvqa | [{"from": "human", "value": "which direction is the group going?\n<image>"}, {"from": "gpt", "value": "walking outwards the market"}] |
Subsets and Splits
No saved queries yet
Save your SQL queries to embed, download, and access them later. Queries will appear here once saved.