image
stringlengths
15
15
id
stringlengths
8
12
conversations
stringlengths
79
268
00000002301.jpg
2301_evjvqa
[{"from": "human", "value": "what color is the shirt of the girl wearing glasses?\n<image>"}, {"from": "gpt", "value": "the girl wearing glasses wears a red shirt"}]
00000002301.jpg
2301_evjvqa
[{"from": "human", "value": "what fruit is put on the table next to the girl in red t-shirt?\n<image>"}, {"from": "gpt", "value": "the girl in red t-shirt is standing next to the lychees table"}]
00000002301.jpg
2301_evjvqa
[{"from": "human", "value": "what are put on the table next to the woman in red shirt?\n<image>"}, {"from": "gpt", "value": "bunches of lychee"}]
00000002302.jpg
2302_evjvqa
[{"from": "human", "value": "what is this woman looking at?\n<image>"}, {"from": "gpt", "value": "she is looking at the milk"}]
00000002302.jpg
2302_evjvqa
[{"from": "human", "value": "what is the woman in white shirt doing?\n<image>"}, {"from": "gpt", "value": "the woman in white shirt is holding and looking at the bottle of milk"}]
00000002302.jpg
2302_evjvqa
[{"from": "human", "value": "what is the girl about to buy at this stall?\n<image>"}, {"from": "gpt", "value": "bottles of milk"}]
00000002304.jpg
2304_evjvqa
[{"from": "human", "value": "what is the girl in pink shirt wearing on her shoulder?\n<image>"}, {"from": "gpt", "value": "she is wearing a black purse"}]
00000002304.jpg
2304_evjvqa
[{"from": "human", "value": "where are those two peoples at together?\n<image>"}, {"from": "gpt", "value": "they are together in a supermarket"}]
00000002304.jpg
2304_evjvqa
[{"from": "human", "value": "what are the girl and the boy come to the market for?\n<image>"}, {"from": "gpt", "value": "shopping"}]
00000002308.jpg
2308_evjvqa
[{"from": "human", "value": "what is the girl in blue jacket coming to this store for?\n<image>"}, {"from": "gpt", "value": "she is coming for christmas decorations"}]
00000002308.jpg
2308_evjvqa
[{"from": "human", "value": "who is coming with the girl in blue jacket?\n<image>"}, {"from": "gpt", "value": "the girl in pink jacket"}]
00000002308.jpg
2308_evjvqa
[{"from": "human", "value": "how does the store owner look like?\n<image>"}, {"from": "gpt", "value": "the woman wearing blue t-shirt at the checkout"}]
00000002308.jpg
2308_evjvqa
[{"from": "human", "value": "what does the woman in blue t-shirt sell?\n<image>"}, {"from": "gpt", "value": "christmas decorations"}]
00000002310.jpg
2310_evjvqa
[{"from": "human", "value": "what is she doing at the supermarket?\n<image>"}, {"from": "gpt", "value": "scanning the qr code"}]
00000002311.jpg
2311_evjvqa
[{"from": "human", "value": "what would we want buy if we come to this store?\n<image>"}, {"from": "gpt", "value": "we would buy a motorbike if we come to this store"}]
00000002314.jpg
2314_evjvqa
[{"from": "human", "value": "what color is the board of the store?\n<image>"}, {"from": "gpt", "value": "the board of the store is painted orange and white"}]
00000002318.jpg
2318_evjvqa
[{"from": "human", "value": "how was the ambulance's license plate look like?\n<image>"}, {"from": "gpt", "value": "the ambulance plate is rectangular, blue background and white text"}]
00000002319.jpg
2319_evjvqa
[{"from": "human", "value": "what products are sold on the right aisle?\n<image>"}, {"from": "gpt", "value": "fresh vegestables"}]
00000002319.jpg
2319_evjvqa
[{"from": "human", "value": "what are the customers doing here?\n<image>"}, {"from": "gpt", "value": "they are shopping in the supermarket"}]
00000002319.jpg
2319_evjvqa
[{"from": "human", "value": "what color is the mask of the girl in white t-shirt?\n<image>"}, {"from": "gpt", "value": "the girl in white t-shirt wears a white mask"}]
00000002328.jpg
2328_evjvqa
[{"from": "human", "value": "what color is the masks of those three people?\n<image>"}, {"from": "gpt", "value": "they are wearing white masks"}]
00000002328.jpg
2328_evjvqa
[{"from": "human", "value": "how was the woman's shirt look like?\n<image>"}, {"from": "gpt", "value": "the woman wears a white shirt with black polka dots"}]
00000002328.jpg
2328_evjvqa
[{"from": "human", "value": "who is standing in this store?\n<image>"}, {"from": "gpt", "value": "there are three people, a woman and two men"}]
00000002333.jpg
2333_evjvqa
[{"from": "human", "value": "who is making a speech at this event?\n<image>"}, {"from": "gpt", "value": "the man with white hair"}]
00000002333.jpg
2333_evjvqa
[{"from": "human", "value": "what is the man with white hair standing for?\n<image>"}, {"from": "gpt", "value": "he is making a speech at the event?"}]
00000002333.jpg
2333_evjvqa
[{"from": "human", "value": "where is the man with white hair standing for the speech?\n<image>"}, {"from": "gpt", "value": "he is standing on the rostrum"}]
00000002338.jpg
2338_evjvqa
[{"from": "human", "value": "how were the cashier costumes look like?\n<image>"}, {"from": "gpt", "value": "white shirt and white hat"}]
00000002338.jpg
2338_evjvqa
[{"from": "human", "value": "what is the man in black t-shirt doing at the checkout?\n<image>"}, {"from": "gpt", "value": "he is packaging the goods into the plastic bag"}]
00000002338.jpg
2338_evjvqa
[{"from": "human", "value": "what color is the shirt of the man packaging the goods into the plastic bag?\n<image>"}, {"from": "gpt", "value": "the man packaging the goods into the plastic bag wears black t-shirt"}]
00000002339.jpg
2339_evjvqa
[{"from": "human", "value": "what is this store mainly sold?\n<image>"}, {"from": "gpt", "value": "this store mainly sells different kinds of dress"}]
00000002340.jpg
2340_evjvqa
[{"from": "human", "value": "which kid is wearing the bag?\n<image>"}, {"from": "gpt", "value": "the boy wearing black shirt is wearing the bag"}]
00000002340.jpg
2340_evjvqa
[{"from": "human", "value": "what are the kids doing at the book aisle?\n<image>"}, {"from": "gpt", "value": "they are looking for books"}]
00000002345.jpg
2345_evjvqa
[{"from": "human", "value": "what are those two man doing?\n<image>"}, {"from": "gpt", "value": "they are fixing the board for the store"}]
00000002345.jpg
2345_evjvqa
[{"from": "human", "value": "where are the men fixing the board?\n<image>"}, {"from": "gpt", "value": "one standing behind the board and one is next to the board"}]
00000002345.jpg
2345_evjvqa
[{"from": "human", "value": "what helps the men to climb up the store?\n<image>"}, {"from": "gpt", "value": "ladders"}]
00000002349.jpg
2349_evjvqa
[{"from": "human", "value": "where is the man sitting at?\n<image>"}, {"from": "gpt", "value": "he is sitting on the parking lot for disabled people"}]
00000002349.jpg
2349_evjvqa
[{"from": "human", "value": "where is the black car parking?\n<image>"}, {"from": "gpt", "value": "the black car is parking at the disabled people area"}]
00000002349.jpg
2349_evjvqa
[{"from": "human", "value": "how was the man sitting look like?\n<image>"}, {"from": "gpt", "value": "white t-shirt, black jeans and wearing black sneakers"}]
00000002356.jpg
2356_evjvqa
[{"from": "human", "value": "which side of the man with white hair is the woman standing?\n<image>"}, {"from": "gpt", "value": "the right side"}]
00000002359.jpg
2359_evjvqa
[{"from": "human", "value": "what color is the balloon?\n<image>"}, {"from": "gpt", "value": "the balloon is red"}]
00000002359.jpg
2359_evjvqa
[{"from": "human", "value": "what is the woman wearing on her shoulder?\n<image>"}, {"from": "gpt", "value": "a white handbag"}]
00000002359.jpg
2359_evjvqa
[{"from": "human", "value": "where is the trolley ?\n<image>"}, {"from": "gpt", "value": "the trolley is next to the tomatoes aisle"}]
00000002363.jpg
2363_evjvqa
[{"from": "human", "value": "where is the watch?\n<image>"}, {"from": "gpt", "value": "the watch is on the left hand of the person who is pushing the trolley"}]
00000002363.jpg
2363_evjvqa
[{"from": "human", "value": "what fruit on the trolley is yellow?\n<image>"}, {"from": "gpt", "value": "the banana"}]
00000002363.jpg
2363_evjvqa
[{"from": "human", "value": "what are the hands pushing?\n<image>"}, {"from": "gpt", "value": "the trolley"}]
00000002364.jpg
2364_evjvqa
[{"from": "human", "value": "what are the people gathering on the shore for?\n<image>"}, {"from": "gpt", "value": "to watch the boat racing"}]
00000002364.jpg
2364_evjvqa
[{"from": "human", "value": "what are the people holding the paddle at the same time for?\n<image>"}, {"from": "gpt", "value": "to race with other boats"}]
00000002364.jpg
2364_evjvqa
[{"from": "human", "value": "which boat is leading the race?\n<image>"}, {"from": "gpt", "value": "the boat of blue shirt members"}]
00000002370.jpg
2370_evjvqa
[{"from": "human", "value": "what products are sold at this area?\n<image>"}, {"from": "gpt", "value": "different kinds of bread"}]
00000002370.jpg
2370_evjvqa
[{"from": "human", "value": "what product can we find at this stall?\n<image>"}, {"from": "gpt", "value": "different kinds of bread"}]
00000002370.jpg
2370_evjvqa
[{"from": "human", "value": "who is carrying the kid in pink shirt on the trolley?\n<image>"}, {"from": "gpt", "value": "the man wearing black t-shirt"}]
00000002370.jpg
2370_evjvqa
[{"from": "human", "value": "what is the woman wearing black dress with white dots about to buy?\n<image>"}, {"from": "gpt", "value": "the woman is about to buy bread"}]
00000002371.jpg
2371_evjvqa
[{"from": "human", "value": "what special occasion are those decorations used for?\n<image>"}, {"from": "gpt", "value": "tet lunar new year"}]
00000002371.jpg
2371_evjvqa
[{"from": "human", "value": "what is the girl wearing jacket on her head is holding on her hand?\n<image>"}, {"from": "gpt", "value": "she is holding her purse"}]
00000002375.jpg
2375_evjvqa
[{"from": "human", "value": "how many kids are there on the swings?\n<image>"}, {"from": "gpt", "value": "there are two kids sitting on the swings"}]
00000002375.jpg
2375_evjvqa
[{"from": "human", "value": "where are those kids playing at?\n<image>"}, {"from": "gpt", "value": "in a playground of the park"}]
00000002375.jpg
2375_evjvqa
[{"from": "human", "value": "what is the kid wearing cap doing?\n<image>"}, {"from": "gpt", "value": "playing on the swing"}]
00000002380.jpg
2380_evjvqa
[{"from": "human", "value": "what is the girl wearing on her wrist?\n<image>"}, {"from": "gpt", "value": "there is a watch on her wrist"}]
00000002380.jpg
2380_evjvqa
[{"from": "human", "value": "what is the woman's hand holding?\n<image>"}, {"from": "gpt", "value": "she is holding the flowers"}]
00000002380.jpg
2380_evjvqa
[{"from": "human", "value": "who is standing close to the girl?\n<image>"}, {"from": "gpt", "value": "the girl is standing alone "}]
00000002383.jpg
2383_evjvqa
[{"from": "human", "value": "what is the woman doing with her camera?\n<image>"}, {"from": "gpt", "value": "she is taking pictures"}]
00000002383.jpg
2383_evjvqa
[{"from": "human", "value": "how was the costumes of the girl standing?\n<image>"}, {"from": "gpt", "value": "she wears white shirt, black skirt and hat"}]
00000002383.jpg
2383_evjvqa
[{"from": "human", "value": "what device is the girl holding on her hand?\n<image>"}, {"from": "gpt", "value": "she is holding a camera"}]
00000002384.jpg
2384_evjvqa
[{"from": "human", "value": "what store is the girl in red dress standing at?\n<image>"}, {"from": "gpt", "value": "she is standing in a bookstore"}]
00000002384.jpg
2384_evjvqa
[{"from": "human", "value": "what is the girl in red dress holding on her hand?\n<image>"}, {"from": "gpt", "value": "she is holding a book"}]
00000002384.jpg
2384_evjvqa
[{"from": "human", "value": "what are those coming to this bookstore for?\n<image>"}, {"from": "gpt", "value": "to buy books"}]
00000002385.jpg
2385_evjvqa
[{"from": "human", "value": "where are the life jacket placed on the boat?\n<image>"}, {"from": "gpt", "value": "it is placed at the back of the boat, behind the person in pink shirt"}]
00000002385.jpg
2385_evjvqa
[{"from": "human", "value": "what is the woman holding the paddles for?\n<image>"}, {"from": "gpt", "value": "she is holding the paddle to ride the boat"}]
00000002385.jpg
2385_evjvqa
[{"from": "human", "value": "what color is the shirt of the woman wearing conical hat?\n<image>"}, {"from": "gpt", "value": "the woman wearing conical hat is wearing a pink shirt"}]
00000002385.jpg
2385_evjvqa
[{"from": "human", "value": "where can we find the conical hat?\n<image>"}, {"from": "gpt", "value": "the conical hat is worn by the woman in pink shirt"}]
00000002386.jpg
2386_evjvqa
[{"from": "human", "value": "what is the staff in yellow t-shirt wearing on his hat?\n<image>"}, {"from": "gpt", "value": "he is wearing a yellow cap"}]
00000002386.jpg
2386_evjvqa
[{"from": "human", "value": "what are placing on the the table in front of the staff in yellow shirt?\n<image>"}, {"from": "gpt", "value": "breads and food"}]
00000002386.jpg
2386_evjvqa
[{"from": "human", "value": "how was the costumes of the man standing between two baskets of bread?\n<image>"}, {"from": "gpt", "value": "the costumes includes yellow t-shirt, black pants and yellow cap"}]
00000002389.jpg
2389_evjvqa
[{"from": "human", "value": "what is hung on the strings above?\n<image>"}, {"from": "gpt", "value": "the conical hats"}]
00000002389.jpg
2389_evjvqa
[{"from": "human", "value": "where can we find the voltage transformer?\n<image>"}, {"from": "gpt", "value": "the voltage transformer is on the left sidewalk of the road"}]
00000002389.jpg
2389_evjvqa
[{"from": "human", "value": "what is this girl doing in the middle of the road?\n<image>"}, {"from": "gpt", "value": "she is walking"}]
00000002390.jpg
2390_evjvqa
[{"from": "human", "value": "where is the bench in this room?\n<image>"}, {"from": "gpt", "value": "the bench is in the middle of the room"}]
00000002390.jpg
2390_evjvqa
[{"from": "human", "value": "how many paintinings are hanging on the walls?\n<image>"}, {"from": "gpt", "value": "there are six paintings hanging on the walls"}]
00000002390.jpg
2390_evjvqa
[{"from": "human", "value": "how many paintings are there on the black wall?\n<image>"}, {"from": "gpt", "value": "there is one painting on the black wall"}]
00000002392.jpg
2392_evjvqa
[{"from": "human", "value": "what can this aisle sell?\n<image>"}, {"from": "gpt", "value": "this aisle sells woman costumes"}]
00000002392.jpg
2392_evjvqa
[{"from": "human", "value": "where can we find the blue plastic chair?\n<image>"}, {"from": "gpt", "value": "behind the woman looking into the stalls"}]
00000002392.jpg
2392_evjvqa
[{"from": "human", "value": "what can we buy at this place?\n<image>"}, {"from": "gpt", "value": "women's clothes"}]
00000002395.jpg
2395_evjvqa
[{"from": "human", "value": "what are people sitting on the ground doing?\n<image>"}, {"from": "gpt", "value": "they are playing traditional musical instruments"}]
00000002395.jpg
2395_evjvqa
[{"from": "human", "value": "what are those people sitting on to play traditional musical instrument?\n<image>"}, {"from": "gpt", "value": "they are sitting on a mat"}]
00000002395.jpg
2395_evjvqa
[{"from": "human", "value": "what is the man wearing green shirt doing?\n<image>"}, {"from": "gpt", "value": "he is standing"}]
00000002396.jpg
2396_evjvqa
[{"from": "human", "value": "what are the costumes of thoses three people?\n<image>"}, {"from": "gpt", "value": "they are wearing red shirt with yellow star in the middle"}]
00000002396.jpg
2396_evjvqa
[{"from": "human", "value": "who is keeping the camera?\n<image>"}, {"from": "gpt", "value": "the camera is holding by the wan wearing hat"}]
00000004403.jpg
4403_evjvqa
[{"from": "human", "value": "where are these two boys sitting?\n<image>"}, {"from": "gpt", "value": "on a bench in a public place"}]
00000004403.jpg
4403_evjvqa
[{"from": "human", "value": "what are these two boys doing?\n<image>"}, {"from": "gpt", "value": "they are using phones"}]
00000004406.jpg
4406_evjvqa
[{"from": "human", "value": "what is the woman in the pink coat doing?\n<image>"}, {"from": "gpt", "value": "walking the motorbike"}]
00000004406.jpg
4406_evjvqa
[{"from": "human", "value": "what is the woman in the blue shirt doing?\n<image>"}, {"from": "gpt", "value": "she sales meat for the customer"}]
00000004406.jpg
4406_evjvqa
[{"from": "human", "value": "where is the woman in black overcoat standing?\n<image>"}, {"from": "gpt", "value": "at the meat stall in the supermarket"}]
00000004406.jpg
4406_evjvqa
[{"from": "human", "value": "what is the woman in the black coat going to buy?\n<image>"}, {"from": "gpt", "value": "the raw meat"}]
00000004406.jpg
4406_evjvqa
[{"from": "human", "value": "what do they sell at this stall?\n<image>"}, {"from": "gpt", "value": "the raw meat"}]
00000004407.jpg
4407_evjvqa
[{"from": "human", "value": "what does this man use the boat to carry?\n<image>"}, {"from": "gpt", "value": "the coconuts"}]
00000004407.jpg
4407_evjvqa
[{"from": "human", "value": "what means does this man carry coconuts?\n<image>"}, {"from": "gpt", "value": "the boat"}]
00000004407.jpg
4407_evjvqa
[{"from": "human", "value": "how are these coconuts transported?\n<image>"}, {"from": "gpt", "value": "by river"}]
00000004410.jpg
4410_evjvqa
[{"from": "human", "value": "what does the woman in the red coat do?\n<image>"}, {"from": "gpt", "value": "she sales the flowers"}]
00000004410.jpg
4410_evjvqa
[{"from": "human", "value": "what is the woman in the red coat doing?\n<image>"}, {"from": "gpt", "value": "she is giving the flowers to customers"}]
00000004410.jpg
4410_evjvqa
[{"from": "human", "value": "where is the bouquet stall openned by the woman?\n<image>"}, {"from": "gpt", "value": "on the side of the road"}]
README.md exists but content is empty. Use the Edit dataset card button to edit it.
Downloads last month
52
Edit dataset card