image
				 
			stringlengths 15 
			15 
			 | id
				 
			stringlengths 8 
			12 
			 | conversations
				 
			stringlengths 79 
			268 
			 | 
|---|---|---|
	00000002301.jpg 
 | 
	2301_evjvqa 
 | 
	[{"from": "human", "value": "what color is the shirt of the girl wearing glasses?\n<image>"}, {"from": "gpt", "value": "the girl wearing glasses wears a red shirt"}] 
 | 
					
	00000002301.jpg 
 | 
	2301_evjvqa 
 | 
	[{"from": "human", "value": "what fruit is put on the table next to the girl in red t-shirt?\n<image>"}, {"from": "gpt", "value": "the girl in red t-shirt is standing next to the lychees table"}] 
 | 
					
	00000002301.jpg 
 | 
	2301_evjvqa 
 | 
	[{"from": "human", "value": "what are put on the table next to the woman in red shirt?\n<image>"}, {"from": "gpt", "value": "bunches of lychee"}] 
 | 
					
	00000002302.jpg 
 | 
	2302_evjvqa 
 | 
	[{"from": "human", "value": "what is this woman looking at?\n<image>"}, {"from": "gpt", "value": "she is looking at the milk"}] 
 | 
					
	00000002302.jpg 
 | 
	2302_evjvqa 
 | 
	[{"from": "human", "value": "what is the woman in white shirt doing?\n<image>"}, {"from": "gpt", "value": "the woman in white shirt is holding and looking at the bottle of milk"}] 
 | 
					
	00000002302.jpg 
 | 
	2302_evjvqa 
 | 
	[{"from": "human", "value": "what is the girl about to buy at this stall?\n<image>"}, {"from": "gpt", "value": "bottles of milk"}] 
 | 
					
	00000002304.jpg 
 | 
	2304_evjvqa 
 | 
	[{"from": "human", "value": "what is the girl in pink shirt wearing on her shoulder?\n<image>"}, {"from": "gpt", "value": "she is wearing a black purse"}] 
 | 
					
	00000002304.jpg 
 | 
	2304_evjvqa 
 | 
	[{"from": "human", "value": "where are those two peoples at together?\n<image>"}, {"from": "gpt", "value": "they are together in a supermarket"}] 
 | 
					
	00000002304.jpg 
 | 
	2304_evjvqa 
 | 
	[{"from": "human", "value": "what are the girl and the boy come to the market for?\n<image>"}, {"from": "gpt", "value": "shopping"}] 
 | 
					
	00000002308.jpg 
 | 
	2308_evjvqa 
 | 
	[{"from": "human", "value": "what is the girl in blue jacket coming to this store for?\n<image>"}, {"from": "gpt", "value": "she is coming for christmas decorations"}] 
 | 
					
	00000002308.jpg 
 | 
	2308_evjvqa 
 | 
	[{"from": "human", "value": "who is coming with the girl in blue jacket?\n<image>"}, {"from": "gpt", "value": "the girl in pink jacket"}] 
 | 
					
	00000002308.jpg 
 | 
	2308_evjvqa 
 | 
	[{"from": "human", "value": "how does the store owner look like?\n<image>"}, {"from": "gpt", "value": "the woman wearing blue t-shirt at the checkout"}] 
 | 
					
	00000002308.jpg 
 | 
	2308_evjvqa 
 | 
	[{"from": "human", "value": "what does the woman in blue t-shirt sell?\n<image>"}, {"from": "gpt", "value": "christmas decorations"}] 
 | 
					
	00000002310.jpg 
 | 
	2310_evjvqa 
 | 
	[{"from": "human", "value": "what is she doing at the supermarket?\n<image>"}, {"from": "gpt", "value": "scanning the qr code"}] 
 | 
					
	00000002311.jpg 
 | 
	2311_evjvqa 
 | 
	[{"from": "human", "value": "what would we want buy if we come to this store?\n<image>"}, {"from": "gpt", "value": "we would buy a motorbike if we come to this store"}] 
 | 
					
	00000002314.jpg 
 | 
	2314_evjvqa 
 | 
	[{"from": "human", "value": "what color is the board of the store?\n<image>"}, {"from": "gpt", "value": "the board of the store is painted orange and white"}] 
 | 
					
	00000002318.jpg 
 | 
	2318_evjvqa 
 | 
	[{"from": "human", "value": "how was the ambulance's license plate look like?\n<image>"}, {"from": "gpt", "value": "the ambulance plate is rectangular, blue background and white text"}] 
 | 
					
	00000002319.jpg 
 | 
	2319_evjvqa 
 | 
	[{"from": "human", "value": "what products are sold on the right aisle?\n<image>"}, {"from": "gpt", "value": "fresh vegestables"}] 
 | 
					
	00000002319.jpg 
 | 
	2319_evjvqa 
 | 
	[{"from": "human", "value": "what are the customers doing here?\n<image>"}, {"from": "gpt", "value": "they are shopping in the supermarket"}] 
 | 
					
	00000002319.jpg 
 | 
	2319_evjvqa 
 | 
	[{"from": "human", "value": "what color is the mask of the girl in white t-shirt?\n<image>"}, {"from": "gpt", "value": "the girl in white t-shirt wears a white mask"}] 
 | 
					
	00000002328.jpg 
 | 
	2328_evjvqa 
 | 
	[{"from": "human", "value": "what color is the masks of those three people?\n<image>"}, {"from": "gpt", "value": "they are wearing white masks"}] 
 | 
					
	00000002328.jpg 
 | 
	2328_evjvqa 
 | 
	[{"from": "human", "value": "how was the woman's shirt look like?\n<image>"}, {"from": "gpt", "value": "the woman wears a white shirt with black polka dots"}] 
 | 
					
	00000002328.jpg 
 | 
	2328_evjvqa 
 | 
	[{"from": "human", "value": "who is standing in this store?\n<image>"}, {"from": "gpt", "value": "there are three people, a woman and two men"}] 
 | 
					
	00000002333.jpg 
 | 
	2333_evjvqa 
 | 
	[{"from": "human", "value": "who is making a speech at this event?\n<image>"}, {"from": "gpt", "value": "the man with white hair"}] 
 | 
					
	00000002333.jpg 
 | 
	2333_evjvqa 
 | 
	[{"from": "human", "value": "what is the man with white hair standing for?\n<image>"}, {"from": "gpt", "value": "he is making a speech at the event?"}] 
 | 
					
	00000002333.jpg 
 | 
	2333_evjvqa 
 | 
	[{"from": "human", "value": "where is the man with white hair standing for the speech?\n<image>"}, {"from": "gpt", "value": "he is standing on the rostrum"}] 
 | 
					
	00000002338.jpg 
 | 
	2338_evjvqa 
 | 
	[{"from": "human", "value": "how were the cashier costumes look like?\n<image>"}, {"from": "gpt", "value": "white shirt and white hat"}] 
 | 
					
	00000002338.jpg 
 | 
	2338_evjvqa 
 | 
	[{"from": "human", "value": "what is the man in black t-shirt doing at the checkout?\n<image>"}, {"from": "gpt", "value": "he is packaging the goods into the plastic bag"}] 
 | 
					
	00000002338.jpg 
 | 
	2338_evjvqa 
 | 
	[{"from": "human", "value": "what color is the shirt of the man packaging the goods into the plastic bag?\n<image>"}, {"from": "gpt", "value": "the man packaging the goods into the plastic bag wears black t-shirt"}] 
 | 
					
	00000002339.jpg 
 | 
	2339_evjvqa 
 | 
	[{"from": "human", "value": "what is this store mainly sold?\n<image>"}, {"from": "gpt", "value": "this store mainly sells different kinds of dress"}] 
 | 
					
	00000002340.jpg 
 | 
	2340_evjvqa 
 | 
	[{"from": "human", "value": "which kid is wearing the bag?\n<image>"}, {"from": "gpt", "value": "the boy wearing black shirt is wearing the bag"}] 
 | 
					
	00000002340.jpg 
 | 
	2340_evjvqa 
 | 
	[{"from": "human", "value": "what are the kids doing at the book aisle?\n<image>"}, {"from": "gpt", "value": "they are looking for books"}] 
 | 
					
	00000002345.jpg 
 | 
	2345_evjvqa 
 | 
	[{"from": "human", "value": "what are those two man doing?\n<image>"}, {"from": "gpt", "value": "they are fixing the board for the store"}] 
 | 
					
	00000002345.jpg 
 | 
	2345_evjvqa 
 | 
	[{"from": "human", "value": "where are the men fixing the board?\n<image>"}, {"from": "gpt", "value": "one standing behind the board and one is next to the board"}] 
 | 
					
	00000002345.jpg 
 | 
	2345_evjvqa 
 | 
	[{"from": "human", "value": "what helps the men to climb up the store?\n<image>"}, {"from": "gpt", "value": "ladders"}] 
 | 
					
	00000002349.jpg 
 | 
	2349_evjvqa 
 | 
	[{"from": "human", "value": "where is the man sitting at?\n<image>"}, {"from": "gpt", "value": "he is sitting on the parking lot for disabled people"}] 
 | 
					
	00000002349.jpg 
 | 
	2349_evjvqa 
 | 
	[{"from": "human", "value": "where is the black car parking?\n<image>"}, {"from": "gpt", "value": "the black car is parking at the disabled people area"}] 
 | 
					
	00000002349.jpg 
 | 
	2349_evjvqa 
 | 
	[{"from": "human", "value": "how was the man sitting look like?\n<image>"}, {"from": "gpt", "value": "white t-shirt, black jeans and wearing black sneakers"}] 
 | 
					
	00000002356.jpg 
 | 
	2356_evjvqa 
 | 
	[{"from": "human", "value": "which side of the man with white hair is the woman standing?\n<image>"}, {"from": "gpt", "value": "the right side"}] 
 | 
					
	00000002359.jpg 
 | 
	2359_evjvqa 
 | 
	[{"from": "human", "value": "what color is the balloon?\n<image>"}, {"from": "gpt", "value": "the balloon is red"}] 
 | 
					
	00000002359.jpg 
 | 
	2359_evjvqa 
 | 
	[{"from": "human", "value": "what is the woman wearing on her shoulder?\n<image>"}, {"from": "gpt", "value": "a white handbag"}] 
 | 
					
	00000002359.jpg 
 | 
	2359_evjvqa 
 | 
	[{"from": "human", "value": "where is the trolley ?\n<image>"}, {"from": "gpt", "value": "the trolley is next to the tomatoes aisle"}] 
 | 
					
	00000002363.jpg 
 | 
	2363_evjvqa 
 | 
	[{"from": "human", "value": "where is the watch?\n<image>"}, {"from": "gpt", "value": "the watch is on the left hand of the person who is pushing the trolley"}] 
 | 
					
	00000002363.jpg 
 | 
	2363_evjvqa 
 | 
	[{"from": "human", "value": "what fruit on the trolley is yellow?\n<image>"}, {"from": "gpt", "value": "the banana"}] 
 | 
					
	00000002363.jpg 
 | 
	2363_evjvqa 
 | 
	[{"from": "human", "value": "what are the hands pushing?\n<image>"}, {"from": "gpt", "value": "the trolley"}] 
 | 
					
	00000002364.jpg 
 | 
	2364_evjvqa 
 | 
	[{"from": "human", "value": "what are the people gathering on the shore for?\n<image>"}, {"from": "gpt", "value": "to watch the boat racing"}] 
 | 
					
	00000002364.jpg 
 | 
	2364_evjvqa 
 | 
	[{"from": "human", "value": "what are the people holding the paddle at the same time for?\n<image>"}, {"from": "gpt", "value": "to race with other boats"}] 
 | 
					
	00000002364.jpg 
 | 
	2364_evjvqa 
 | 
	[{"from": "human", "value": "which boat is leading the race?\n<image>"}, {"from": "gpt", "value": "the boat of blue shirt members"}] 
 | 
					
	00000002370.jpg 
 | 
	2370_evjvqa 
 | 
	[{"from": "human", "value": "what products are sold at this area?\n<image>"}, {"from": "gpt", "value": "different kinds of bread"}] 
 | 
					
	00000002370.jpg 
 | 
	2370_evjvqa 
 | 
	[{"from": "human", "value": "what product can we find at this stall?\n<image>"}, {"from": "gpt", "value": "different kinds of bread"}] 
 | 
					
	00000002370.jpg 
 | 
	2370_evjvqa 
 | 
	[{"from": "human", "value": "who is carrying the kid in pink shirt on the trolley?\n<image>"}, {"from": "gpt", "value": "the man wearing black t-shirt"}] 
 | 
					
	00000002370.jpg 
 | 
	2370_evjvqa 
 | 
	[{"from": "human", "value": "what is the woman wearing black dress with white dots about to buy?\n<image>"}, {"from": "gpt", "value": "the woman is about to buy bread"}] 
 | 
					
	00000002371.jpg 
 | 
	2371_evjvqa 
 | 
	[{"from": "human", "value": "what special occasion are those decorations used for?\n<image>"}, {"from": "gpt", "value": "tet lunar new year"}] 
 | 
					
	00000002371.jpg 
 | 
	2371_evjvqa 
 | 
	[{"from": "human", "value": "what is the girl wearing jacket on her head is holding on her hand?\n<image>"}, {"from": "gpt", "value": "she is holding her purse"}] 
 | 
					
	00000002375.jpg 
 | 
	2375_evjvqa 
 | 
	[{"from": "human", "value": "how many kids are there on the swings?\n<image>"}, {"from": "gpt", "value": "there are two kids sitting on the swings"}] 
 | 
					
	00000002375.jpg 
 | 
	2375_evjvqa 
 | 
	[{"from": "human", "value": "where are those kids playing at?\n<image>"}, {"from": "gpt", "value": "in a playground of the park"}] 
 | 
					
	00000002375.jpg 
 | 
	2375_evjvqa 
 | 
	[{"from": "human", "value": "what is the kid wearing cap doing?\n<image>"}, {"from": "gpt", "value": "playing on the swing"}] 
 | 
					
	00000002380.jpg 
 | 
	2380_evjvqa 
 | 
	[{"from": "human", "value": "what is the girl wearing on her wrist?\n<image>"}, {"from": "gpt", "value": "there is a watch on her wrist"}] 
 | 
					
	00000002380.jpg 
 | 
	2380_evjvqa 
 | 
	[{"from": "human", "value": "what is the woman's hand holding?\n<image>"}, {"from": "gpt", "value": "she is holding the flowers"}] 
 | 
					
	00000002380.jpg 
 | 
	2380_evjvqa 
 | 
	[{"from": "human", "value": "who is standing close to the girl?\n<image>"}, {"from": "gpt", "value": "the girl is standing alone "}] 
 | 
					
	00000002383.jpg 
 | 
	2383_evjvqa 
 | 
	[{"from": "human", "value": "what is the woman doing with her camera?\n<image>"}, {"from": "gpt", "value": "she is taking pictures"}] 
 | 
					
	00000002383.jpg 
 | 
	2383_evjvqa 
 | 
	[{"from": "human", "value": "how was the costumes of the girl standing?\n<image>"}, {"from": "gpt", "value": "she wears white shirt, black skirt and hat"}] 
 | 
					
	00000002383.jpg 
 | 
	2383_evjvqa 
 | 
	[{"from": "human", "value": "what device is the girl holding on her hand?\n<image>"}, {"from": "gpt", "value": "she is holding a camera"}] 
 | 
					
	00000002384.jpg 
 | 
	2384_evjvqa 
 | 
	[{"from": "human", "value": "what store is the girl in red dress standing at?\n<image>"}, {"from": "gpt", "value": "she is standing in a bookstore"}] 
 | 
					
	00000002384.jpg 
 | 
	2384_evjvqa 
 | 
	[{"from": "human", "value": "what is the girl in red dress holding on her hand?\n<image>"}, {"from": "gpt", "value": "she is holding a book"}] 
 | 
					
	00000002384.jpg 
 | 
	2384_evjvqa 
 | 
	[{"from": "human", "value": "what are those coming to this bookstore for?\n<image>"}, {"from": "gpt", "value": "to buy books"}] 
 | 
					
	00000002385.jpg 
 | 
	2385_evjvqa 
 | 
	[{"from": "human", "value": "where are the life jacket placed on the boat?\n<image>"}, {"from": "gpt", "value": "it is placed at the back of the boat, behind the person in pink shirt"}] 
 | 
					
	00000002385.jpg 
 | 
	2385_evjvqa 
 | 
	[{"from": "human", "value": "what is the woman holding the paddles for?\n<image>"}, {"from": "gpt", "value": "she is holding the paddle to ride the boat"}] 
 | 
					
	00000002385.jpg 
 | 
	2385_evjvqa 
 | 
	[{"from": "human", "value": "what color is the shirt of the woman wearing conical hat?\n<image>"}, {"from": "gpt", "value": "the woman wearing conical hat is wearing a pink shirt"}] 
 | 
					
	00000002385.jpg 
 | 
	2385_evjvqa 
 | 
	[{"from": "human", "value": "where can we find the conical hat?\n<image>"}, {"from": "gpt", "value": "the conical hat is worn by the woman in pink shirt"}] 
 | 
					
	00000002386.jpg 
 | 
	2386_evjvqa 
 | 
	[{"from": "human", "value": "what is the staff in yellow t-shirt wearing on his hat?\n<image>"}, {"from": "gpt", "value": "he is wearing a yellow cap"}] 
 | 
					
	00000002386.jpg 
 | 
	2386_evjvqa 
 | 
	[{"from": "human", "value": "what are placing on the the table in front of the staff in yellow shirt?\n<image>"}, {"from": "gpt", "value": "breads and food"}] 
 | 
					
	00000002386.jpg 
 | 
	2386_evjvqa 
 | 
	[{"from": "human", "value": "how was the costumes of the man standing between two baskets of bread?\n<image>"}, {"from": "gpt", "value": "the costumes includes yellow t-shirt, black pants and yellow cap"}] 
 | 
					
	00000002389.jpg 
 | 
	2389_evjvqa 
 | 
	[{"from": "human", "value": "what is hung on the strings above?\n<image>"}, {"from": "gpt", "value": "the conical hats"}] 
 | 
					
	00000002389.jpg 
 | 
	2389_evjvqa 
 | 
	[{"from": "human", "value": "where can we find the voltage transformer?\n<image>"}, {"from": "gpt", "value": "the voltage transformer is on the left sidewalk of the road"}] 
 | 
					
	00000002389.jpg 
 | 
	2389_evjvqa 
 | 
	[{"from": "human", "value": "what is this girl doing in the middle of the road?\n<image>"}, {"from": "gpt", "value": "she is walking"}] 
 | 
					
	00000002390.jpg 
 | 
	2390_evjvqa 
 | 
	[{"from": "human", "value": "where is the bench in this room?\n<image>"}, {"from": "gpt", "value": "the bench is in the middle of the room"}] 
 | 
					
	00000002390.jpg 
 | 
	2390_evjvqa 
 | 
	[{"from": "human", "value": "how many paintinings are hanging on the walls?\n<image>"}, {"from": "gpt", "value": "there are six paintings hanging on the walls"}] 
 | 
					
	00000002390.jpg 
 | 
	2390_evjvqa 
 | 
	[{"from": "human", "value": "how many paintings are there on the black wall?\n<image>"}, {"from": "gpt", "value": "there is one painting on the black wall"}] 
 | 
					
	00000002392.jpg 
 | 
	2392_evjvqa 
 | 
	[{"from": "human", "value": "what can this aisle sell?\n<image>"}, {"from": "gpt", "value": "this aisle sells woman costumes"}] 
 | 
					
	00000002392.jpg 
 | 
	2392_evjvqa 
 | 
	[{"from": "human", "value": "where can we find the blue plastic chair?\n<image>"}, {"from": "gpt", "value": "behind the woman looking into the stalls"}] 
 | 
					
	00000002392.jpg 
 | 
	2392_evjvqa 
 | 
	[{"from": "human", "value": "what can we buy at this place?\n<image>"}, {"from": "gpt", "value": "women's clothes"}] 
 | 
					
	00000002395.jpg 
 | 
	2395_evjvqa 
 | 
	[{"from": "human", "value": "what are people sitting on the ground doing?\n<image>"}, {"from": "gpt", "value": "they are playing traditional musical instruments"}] 
 | 
					
	00000002395.jpg 
 | 
	2395_evjvqa 
 | 
	[{"from": "human", "value": "what are those people sitting on to play traditional musical instrument?\n<image>"}, {"from": "gpt", "value": "they are sitting on a mat"}] 
 | 
					
	00000002395.jpg 
 | 
	2395_evjvqa 
 | 
	[{"from": "human", "value": "what is the man wearing green shirt doing?\n<image>"}, {"from": "gpt", "value": "he is standing"}] 
 | 
					
	00000002396.jpg 
 | 
	2396_evjvqa 
 | 
	[{"from": "human", "value": "what are the costumes of thoses three people?\n<image>"}, {"from": "gpt", "value": "they are wearing red shirt with yellow star in the middle"}] 
 | 
					
	00000002396.jpg 
 | 
	2396_evjvqa 
 | 
	[{"from": "human", "value": "who is keeping the camera?\n<image>"}, {"from": "gpt", "value": "the camera is holding by the wan wearing hat"}] 
 | 
					
	00000004403.jpg 
 | 
	4403_evjvqa 
 | 
	[{"from": "human", "value": "where are these two boys sitting?\n<image>"}, {"from": "gpt", "value": "on a bench in a public place"}] 
 | 
					
	00000004403.jpg 
 | 
	4403_evjvqa 
 | 
	[{"from": "human", "value": "what are these two boys doing?\n<image>"}, {"from": "gpt", "value": "they are using phones"}] 
 | 
					
	00000004406.jpg 
 | 
	4406_evjvqa 
 | 
	[{"from": "human", "value": "what is the woman in the pink coat doing?\n<image>"}, {"from": "gpt", "value": "walking the motorbike"}] 
 | 
					
	00000004406.jpg 
 | 
	4406_evjvqa 
 | 
	[{"from": "human", "value": "what is the woman in the blue shirt doing?\n<image>"}, {"from": "gpt", "value": "she sales meat for the customer"}] 
 | 
					
	00000004406.jpg 
 | 
	4406_evjvqa 
 | 
	[{"from": "human", "value": "where is the woman in black overcoat standing?\n<image>"}, {"from": "gpt", "value": "at the meat stall in the supermarket"}] 
 | 
					
	00000004406.jpg 
 | 
	4406_evjvqa 
 | 
	[{"from": "human", "value": "what is the woman in the black coat going to buy?\n<image>"}, {"from": "gpt", "value": "the raw meat"}] 
 | 
					
	00000004406.jpg 
 | 
	4406_evjvqa 
 | 
	[{"from": "human", "value": "what do they sell at this stall?\n<image>"}, {"from": "gpt", "value": "the raw meat"}] 
 | 
					
	00000004407.jpg 
 | 
	4407_evjvqa 
 | 
	[{"from": "human", "value": "what does this man use the boat to carry?\n<image>"}, {"from": "gpt", "value": "the coconuts"}] 
 | 
					
	00000004407.jpg 
 | 
	4407_evjvqa 
 | 
	[{"from": "human", "value": "what means does this man carry coconuts?\n<image>"}, {"from": "gpt", "value": "the boat"}] 
 | 
					
	00000004407.jpg 
 | 
	4407_evjvqa 
 | 
	[{"from": "human", "value": "how are these coconuts transported?\n<image>"}, {"from": "gpt", "value": "by river"}] 
 | 
					
	00000004410.jpg 
 | 
	4410_evjvqa 
 | 
	[{"from": "human", "value": "what does the woman in the red coat do?\n<image>"}, {"from": "gpt", "value": "she sales the flowers"}] 
 | 
					
	00000004410.jpg 
 | 
	4410_evjvqa 
 | 
	[{"from": "human", "value": "what is the woman in the red coat doing?\n<image>"}, {"from": "gpt", "value": "she is giving the flowers to customers"}] 
 | 
					
	00000004410.jpg 
 | 
	4410_evjvqa 
 | 
	[{"from": "human", "value": "where is the bouquet stall openned by the woman?\n<image>"}, {"from": "gpt", "value": "on the side of the road"}] 
 | 
					
			Subsets and Splits
				
	
				
			
				
No community queries yet
The top public SQL queries from the community will appear here once available.