image
stringlengths 15
15
| id
stringlengths 8
12
| conversations
stringlengths 79
268
|
---|---|---|
00000002301.jpg | 2301_evjvqa | [{"from": "human", "value": "what color is the shirt of the girl wearing glasses?\n<image>"}, {"from": "gpt", "value": "the girl wearing glasses wears a red shirt"}] |
00000002301.jpg | 2301_evjvqa | [{"from": "human", "value": "what fruit is put on the table next to the girl in red t-shirt?\n<image>"}, {"from": "gpt", "value": "the girl in red t-shirt is standing next to the lychees table"}] |
00000002301.jpg | 2301_evjvqa | [{"from": "human", "value": "what are put on the table next to the woman in red shirt?\n<image>"}, {"from": "gpt", "value": "bunches of lychee"}] |
00000002302.jpg | 2302_evjvqa | [{"from": "human", "value": "what is this woman looking at?\n<image>"}, {"from": "gpt", "value": "she is looking at the milk"}] |
00000002302.jpg | 2302_evjvqa | [{"from": "human", "value": "what is the woman in white shirt doing?\n<image>"}, {"from": "gpt", "value": "the woman in white shirt is holding and looking at the bottle of milk"}] |
00000002302.jpg | 2302_evjvqa | [{"from": "human", "value": "what is the girl about to buy at this stall?\n<image>"}, {"from": "gpt", "value": "bottles of milk"}] |
00000002304.jpg | 2304_evjvqa | [{"from": "human", "value": "what is the girl in pink shirt wearing on her shoulder?\n<image>"}, {"from": "gpt", "value": "she is wearing a black purse"}] |
00000002304.jpg | 2304_evjvqa | [{"from": "human", "value": "where are those two peoples at together?\n<image>"}, {"from": "gpt", "value": "they are together in a supermarket"}] |
00000002304.jpg | 2304_evjvqa | [{"from": "human", "value": "what are the girl and the boy come to the market for?\n<image>"}, {"from": "gpt", "value": "shopping"}] |
00000002308.jpg | 2308_evjvqa | [{"from": "human", "value": "what is the girl in blue jacket coming to this store for?\n<image>"}, {"from": "gpt", "value": "she is coming for christmas decorations"}] |
00000002308.jpg | 2308_evjvqa | [{"from": "human", "value": "who is coming with the girl in blue jacket?\n<image>"}, {"from": "gpt", "value": "the girl in pink jacket"}] |
00000002308.jpg | 2308_evjvqa | [{"from": "human", "value": "how does the store owner look like?\n<image>"}, {"from": "gpt", "value": "the woman wearing blue t-shirt at the checkout"}] |
00000002308.jpg | 2308_evjvqa | [{"from": "human", "value": "what does the woman in blue t-shirt sell?\n<image>"}, {"from": "gpt", "value": "christmas decorations"}] |
00000002310.jpg | 2310_evjvqa | [{"from": "human", "value": "what is she doing at the supermarket?\n<image>"}, {"from": "gpt", "value": "scanning the qr code"}] |
00000002311.jpg | 2311_evjvqa | [{"from": "human", "value": "what would we want buy if we come to this store?\n<image>"}, {"from": "gpt", "value": "we would buy a motorbike if we come to this store"}] |
00000002314.jpg | 2314_evjvqa | [{"from": "human", "value": "what color is the board of the store?\n<image>"}, {"from": "gpt", "value": "the board of the store is painted orange and white"}] |
00000002318.jpg | 2318_evjvqa | [{"from": "human", "value": "how was the ambulance's license plate look like?\n<image>"}, {"from": "gpt", "value": "the ambulance plate is rectangular, blue background and white text"}] |
00000002319.jpg | 2319_evjvqa | [{"from": "human", "value": "what products are sold on the right aisle?\n<image>"}, {"from": "gpt", "value": "fresh vegestables"}] |
00000002319.jpg | 2319_evjvqa | [{"from": "human", "value": "what are the customers doing here?\n<image>"}, {"from": "gpt", "value": "they are shopping in the supermarket"}] |
00000002319.jpg | 2319_evjvqa | [{"from": "human", "value": "what color is the mask of the girl in white t-shirt?\n<image>"}, {"from": "gpt", "value": "the girl in white t-shirt wears a white mask"}] |
00000002328.jpg | 2328_evjvqa | [{"from": "human", "value": "what color is the masks of those three people?\n<image>"}, {"from": "gpt", "value": "they are wearing white masks"}] |
00000002328.jpg | 2328_evjvqa | [{"from": "human", "value": "how was the woman's shirt look like?\n<image>"}, {"from": "gpt", "value": "the woman wears a white shirt with black polka dots"}] |
00000002328.jpg | 2328_evjvqa | [{"from": "human", "value": "who is standing in this store?\n<image>"}, {"from": "gpt", "value": "there are three people, a woman and two men"}] |
00000002333.jpg | 2333_evjvqa | [{"from": "human", "value": "who is making a speech at this event?\n<image>"}, {"from": "gpt", "value": "the man with white hair"}] |
00000002333.jpg | 2333_evjvqa | [{"from": "human", "value": "what is the man with white hair standing for?\n<image>"}, {"from": "gpt", "value": "he is making a speech at the event?"}] |
00000002333.jpg | 2333_evjvqa | [{"from": "human", "value": "where is the man with white hair standing for the speech?\n<image>"}, {"from": "gpt", "value": "he is standing on the rostrum"}] |
00000002338.jpg | 2338_evjvqa | [{"from": "human", "value": "how were the cashier costumes look like?\n<image>"}, {"from": "gpt", "value": "white shirt and white hat"}] |
00000002338.jpg | 2338_evjvqa | [{"from": "human", "value": "what is the man in black t-shirt doing at the checkout?\n<image>"}, {"from": "gpt", "value": "he is packaging the goods into the plastic bag"}] |
00000002338.jpg | 2338_evjvqa | [{"from": "human", "value": "what color is the shirt of the man packaging the goods into the plastic bag?\n<image>"}, {"from": "gpt", "value": "the man packaging the goods into the plastic bag wears black t-shirt"}] |
00000002339.jpg | 2339_evjvqa | [{"from": "human", "value": "what is this store mainly sold?\n<image>"}, {"from": "gpt", "value": "this store mainly sells different kinds of dress"}] |
00000002340.jpg | 2340_evjvqa | [{"from": "human", "value": "which kid is wearing the bag?\n<image>"}, {"from": "gpt", "value": "the boy wearing black shirt is wearing the bag"}] |
00000002340.jpg | 2340_evjvqa | [{"from": "human", "value": "what are the kids doing at the book aisle?\n<image>"}, {"from": "gpt", "value": "they are looking for books"}] |
00000002345.jpg | 2345_evjvqa | [{"from": "human", "value": "what are those two man doing?\n<image>"}, {"from": "gpt", "value": "they are fixing the board for the store"}] |
00000002345.jpg | 2345_evjvqa | [{"from": "human", "value": "where are the men fixing the board?\n<image>"}, {"from": "gpt", "value": "one standing behind the board and one is next to the board"}] |
00000002345.jpg | 2345_evjvqa | [{"from": "human", "value": "what helps the men to climb up the store?\n<image>"}, {"from": "gpt", "value": "ladders"}] |
00000002349.jpg | 2349_evjvqa | [{"from": "human", "value": "where is the man sitting at?\n<image>"}, {"from": "gpt", "value": "he is sitting on the parking lot for disabled people"}] |
00000002349.jpg | 2349_evjvqa | [{"from": "human", "value": "where is the black car parking?\n<image>"}, {"from": "gpt", "value": "the black car is parking at the disabled people area"}] |
00000002349.jpg | 2349_evjvqa | [{"from": "human", "value": "how was the man sitting look like?\n<image>"}, {"from": "gpt", "value": "white t-shirt, black jeans and wearing black sneakers"}] |
00000002356.jpg | 2356_evjvqa | [{"from": "human", "value": "which side of the man with white hair is the woman standing?\n<image>"}, {"from": "gpt", "value": "the right side"}] |
00000002359.jpg | 2359_evjvqa | [{"from": "human", "value": "what color is the balloon?\n<image>"}, {"from": "gpt", "value": "the balloon is red"}] |
00000002359.jpg | 2359_evjvqa | [{"from": "human", "value": "what is the woman wearing on her shoulder?\n<image>"}, {"from": "gpt", "value": "a white handbag"}] |
00000002359.jpg | 2359_evjvqa | [{"from": "human", "value": "where is the trolley ?\n<image>"}, {"from": "gpt", "value": "the trolley is next to the tomatoes aisle"}] |
00000002363.jpg | 2363_evjvqa | [{"from": "human", "value": "where is the watch?\n<image>"}, {"from": "gpt", "value": "the watch is on the left hand of the person who is pushing the trolley"}] |
00000002363.jpg | 2363_evjvqa | [{"from": "human", "value": "what fruit on the trolley is yellow?\n<image>"}, {"from": "gpt", "value": "the banana"}] |
00000002363.jpg | 2363_evjvqa | [{"from": "human", "value": "what are the hands pushing?\n<image>"}, {"from": "gpt", "value": "the trolley"}] |
00000002364.jpg | 2364_evjvqa | [{"from": "human", "value": "what are the people gathering on the shore for?\n<image>"}, {"from": "gpt", "value": "to watch the boat racing"}] |
00000002364.jpg | 2364_evjvqa | [{"from": "human", "value": "what are the people holding the paddle at the same time for?\n<image>"}, {"from": "gpt", "value": "to race with other boats"}] |
00000002364.jpg | 2364_evjvqa | [{"from": "human", "value": "which boat is leading the race?\n<image>"}, {"from": "gpt", "value": "the boat of blue shirt members"}] |
00000002370.jpg | 2370_evjvqa | [{"from": "human", "value": "what products are sold at this area?\n<image>"}, {"from": "gpt", "value": "different kinds of bread"}] |
00000002370.jpg | 2370_evjvqa | [{"from": "human", "value": "what product can we find at this stall?\n<image>"}, {"from": "gpt", "value": "different kinds of bread"}] |
00000002370.jpg | 2370_evjvqa | [{"from": "human", "value": "who is carrying the kid in pink shirt on the trolley?\n<image>"}, {"from": "gpt", "value": "the man wearing black t-shirt"}] |
00000002370.jpg | 2370_evjvqa | [{"from": "human", "value": "what is the woman wearing black dress with white dots about to buy?\n<image>"}, {"from": "gpt", "value": "the woman is about to buy bread"}] |
00000002371.jpg | 2371_evjvqa | [{"from": "human", "value": "what special occasion are those decorations used for?\n<image>"}, {"from": "gpt", "value": "tet lunar new year"}] |
00000002371.jpg | 2371_evjvqa | [{"from": "human", "value": "what is the girl wearing jacket on her head is holding on her hand?\n<image>"}, {"from": "gpt", "value": "she is holding her purse"}] |
00000002375.jpg | 2375_evjvqa | [{"from": "human", "value": "how many kids are there on the swings?\n<image>"}, {"from": "gpt", "value": "there are two kids sitting on the swings"}] |
00000002375.jpg | 2375_evjvqa | [{"from": "human", "value": "where are those kids playing at?\n<image>"}, {"from": "gpt", "value": "in a playground of the park"}] |
00000002375.jpg | 2375_evjvqa | [{"from": "human", "value": "what is the kid wearing cap doing?\n<image>"}, {"from": "gpt", "value": "playing on the swing"}] |
00000002380.jpg | 2380_evjvqa | [{"from": "human", "value": "what is the girl wearing on her wrist?\n<image>"}, {"from": "gpt", "value": "there is a watch on her wrist"}] |
00000002380.jpg | 2380_evjvqa | [{"from": "human", "value": "what is the woman's hand holding?\n<image>"}, {"from": "gpt", "value": "she is holding the flowers"}] |
00000002380.jpg | 2380_evjvqa | [{"from": "human", "value": "who is standing close to the girl?\n<image>"}, {"from": "gpt", "value": "the girl is standing alone "}] |
00000002383.jpg | 2383_evjvqa | [{"from": "human", "value": "what is the woman doing with her camera?\n<image>"}, {"from": "gpt", "value": "she is taking pictures"}] |
00000002383.jpg | 2383_evjvqa | [{"from": "human", "value": "how was the costumes of the girl standing?\n<image>"}, {"from": "gpt", "value": "she wears white shirt, black skirt and hat"}] |
00000002383.jpg | 2383_evjvqa | [{"from": "human", "value": "what device is the girl holding on her hand?\n<image>"}, {"from": "gpt", "value": "she is holding a camera"}] |
00000002384.jpg | 2384_evjvqa | [{"from": "human", "value": "what store is the girl in red dress standing at?\n<image>"}, {"from": "gpt", "value": "she is standing in a bookstore"}] |
00000002384.jpg | 2384_evjvqa | [{"from": "human", "value": "what is the girl in red dress holding on her hand?\n<image>"}, {"from": "gpt", "value": "she is holding a book"}] |
00000002384.jpg | 2384_evjvqa | [{"from": "human", "value": "what are those coming to this bookstore for?\n<image>"}, {"from": "gpt", "value": "to buy books"}] |
00000002385.jpg | 2385_evjvqa | [{"from": "human", "value": "where are the life jacket placed on the boat?\n<image>"}, {"from": "gpt", "value": "it is placed at the back of the boat, behind the person in pink shirt"}] |
00000002385.jpg | 2385_evjvqa | [{"from": "human", "value": "what is the woman holding the paddles for?\n<image>"}, {"from": "gpt", "value": "she is holding the paddle to ride the boat"}] |
00000002385.jpg | 2385_evjvqa | [{"from": "human", "value": "what color is the shirt of the woman wearing conical hat?\n<image>"}, {"from": "gpt", "value": "the woman wearing conical hat is wearing a pink shirt"}] |
00000002385.jpg | 2385_evjvqa | [{"from": "human", "value": "where can we find the conical hat?\n<image>"}, {"from": "gpt", "value": "the conical hat is worn by the woman in pink shirt"}] |
00000002386.jpg | 2386_evjvqa | [{"from": "human", "value": "what is the staff in yellow t-shirt wearing on his hat?\n<image>"}, {"from": "gpt", "value": "he is wearing a yellow cap"}] |
00000002386.jpg | 2386_evjvqa | [{"from": "human", "value": "what are placing on the the table in front of the staff in yellow shirt?\n<image>"}, {"from": "gpt", "value": "breads and food"}] |
00000002386.jpg | 2386_evjvqa | [{"from": "human", "value": "how was the costumes of the man standing between two baskets of bread?\n<image>"}, {"from": "gpt", "value": "the costumes includes yellow t-shirt, black pants and yellow cap"}] |
00000002389.jpg | 2389_evjvqa | [{"from": "human", "value": "what is hung on the strings above?\n<image>"}, {"from": "gpt", "value": "the conical hats"}] |
00000002389.jpg | 2389_evjvqa | [{"from": "human", "value": "where can we find the voltage transformer?\n<image>"}, {"from": "gpt", "value": "the voltage transformer is on the left sidewalk of the road"}] |
00000002389.jpg | 2389_evjvqa | [{"from": "human", "value": "what is this girl doing in the middle of the road?\n<image>"}, {"from": "gpt", "value": "she is walking"}] |
00000002390.jpg | 2390_evjvqa | [{"from": "human", "value": "where is the bench in this room?\n<image>"}, {"from": "gpt", "value": "the bench is in the middle of the room"}] |
00000002390.jpg | 2390_evjvqa | [{"from": "human", "value": "how many paintinings are hanging on the walls?\n<image>"}, {"from": "gpt", "value": "there are six paintings hanging on the walls"}] |
00000002390.jpg | 2390_evjvqa | [{"from": "human", "value": "how many paintings are there on the black wall?\n<image>"}, {"from": "gpt", "value": "there is one painting on the black wall"}] |
00000002392.jpg | 2392_evjvqa | [{"from": "human", "value": "what can this aisle sell?\n<image>"}, {"from": "gpt", "value": "this aisle sells woman costumes"}] |
00000002392.jpg | 2392_evjvqa | [{"from": "human", "value": "where can we find the blue plastic chair?\n<image>"}, {"from": "gpt", "value": "behind the woman looking into the stalls"}] |
00000002392.jpg | 2392_evjvqa | [{"from": "human", "value": "what can we buy at this place?\n<image>"}, {"from": "gpt", "value": "women's clothes"}] |
00000002395.jpg | 2395_evjvqa | [{"from": "human", "value": "what are people sitting on the ground doing?\n<image>"}, {"from": "gpt", "value": "they are playing traditional musical instruments"}] |
00000002395.jpg | 2395_evjvqa | [{"from": "human", "value": "what are those people sitting on to play traditional musical instrument?\n<image>"}, {"from": "gpt", "value": "they are sitting on a mat"}] |
00000002395.jpg | 2395_evjvqa | [{"from": "human", "value": "what is the man wearing green shirt doing?\n<image>"}, {"from": "gpt", "value": "he is standing"}] |
00000002396.jpg | 2396_evjvqa | [{"from": "human", "value": "what are the costumes of thoses three people?\n<image>"}, {"from": "gpt", "value": "they are wearing red shirt with yellow star in the middle"}] |
00000002396.jpg | 2396_evjvqa | [{"from": "human", "value": "who is keeping the camera?\n<image>"}, {"from": "gpt", "value": "the camera is holding by the wan wearing hat"}] |
00000004403.jpg | 4403_evjvqa | [{"from": "human", "value": "where are these two boys sitting?\n<image>"}, {"from": "gpt", "value": "on a bench in a public place"}] |
00000004403.jpg | 4403_evjvqa | [{"from": "human", "value": "what are these two boys doing?\n<image>"}, {"from": "gpt", "value": "they are using phones"}] |
00000004406.jpg | 4406_evjvqa | [{"from": "human", "value": "what is the woman in the pink coat doing?\n<image>"}, {"from": "gpt", "value": "walking the motorbike"}] |
00000004406.jpg | 4406_evjvqa | [{"from": "human", "value": "what is the woman in the blue shirt doing?\n<image>"}, {"from": "gpt", "value": "she sales meat for the customer"}] |
00000004406.jpg | 4406_evjvqa | [{"from": "human", "value": "where is the woman in black overcoat standing?\n<image>"}, {"from": "gpt", "value": "at the meat stall in the supermarket"}] |
00000004406.jpg | 4406_evjvqa | [{"from": "human", "value": "what is the woman in the black coat going to buy?\n<image>"}, {"from": "gpt", "value": "the raw meat"}] |
00000004406.jpg | 4406_evjvqa | [{"from": "human", "value": "what do they sell at this stall?\n<image>"}, {"from": "gpt", "value": "the raw meat"}] |
00000004407.jpg | 4407_evjvqa | [{"from": "human", "value": "what does this man use the boat to carry?\n<image>"}, {"from": "gpt", "value": "the coconuts"}] |
00000004407.jpg | 4407_evjvqa | [{"from": "human", "value": "what means does this man carry coconuts?\n<image>"}, {"from": "gpt", "value": "the boat"}] |
00000004407.jpg | 4407_evjvqa | [{"from": "human", "value": "how are these coconuts transported?\n<image>"}, {"from": "gpt", "value": "by river"}] |
00000004410.jpg | 4410_evjvqa | [{"from": "human", "value": "what does the woman in the red coat do?\n<image>"}, {"from": "gpt", "value": "she sales the flowers"}] |
00000004410.jpg | 4410_evjvqa | [{"from": "human", "value": "what is the woman in the red coat doing?\n<image>"}, {"from": "gpt", "value": "she is giving the flowers to customers"}] |
00000004410.jpg | 4410_evjvqa | [{"from": "human", "value": "where is the bouquet stall openned by the woman?\n<image>"}, {"from": "gpt", "value": "on the side of the road"}] |
End of preview. Expand
in Dataset Viewer.
README.md exists but content is empty.
Use the Edit dataset card button to edit it.
- Downloads last month
- 52