| image
				 stringlengths 15 15 | id
				 stringlengths 8 12 | conversations
				 stringlengths 79 268 | 
|---|---|---|
| 
	00000004412.jpg | 
	4412_evjvqa | 
	[{"from": "human", "value": "what holiday do these people go out for a walk?\n<image>"}, {"from": "gpt", "value": "the chirstmas"}] | 
| 
	00000004412.jpg | 
	4412_evjvqa | 
	[{"from": "human", "value": "what is the holiday decorated these statues?\n<image>"}, {"from": "gpt", "value": "the chirstmas"}] | 
| 
	00000004412.jpg | 
	4412_evjvqa | 
	[{"from": "human", "value": "where are these people going to visit?\n<image>"}, {"from": "gpt", "value": "at the christmas decoration area"}] | 
| 
	00000004414.jpg | 
	4414_evjvqa | 
	[{"from": "human", "value": "where is the road lead to?\n<image>"}, {"from": "gpt", "value": "the historical monument"}] | 
| 
	00000004414.jpg | 
	4414_evjvqa | 
	[{"from": "human", "value": "what do people put on both sides of the road?\n<image>"}, {"from": "gpt", "value": "planting the bushes"}] | 
| 
	00000004417.jpg | 
	4417_evjvqa | 
	[{"from": "human", "value": "where are these two rocks placed?\n<image>"}, {"from": "gpt", "value": "in front of the thatched house"}] | 
| 
	00000004417.jpg | 
	4417_evjvqa | 
	[{"from": "human", "value": "what do people put in front of the thatched house?\n<image>"}, {"from": "gpt", "value": "there are two rocks"}] | 
| 
	00000004419.jpg | 
	4419_evjvqa | 
	[{"from": "human", "value": "where is the woman going?\n<image>"}, {"from": "gpt", "value": "this woman is bringing rice to her relatives"}] | 
| 
	00000004419.jpg | 
	4419_evjvqa | 
	[{"from": "human", "value": "where is the woman going?\n<image>"}, {"from": "gpt", "value": "she is walking past two fences"}] | 
| 
	00000004420.jpg | 
	4420_evjvqa | 
	[{"from": "human", "value": "how many people are choosing clothes at the clothes shop on the left hand side?\n<image>"}, {"from": "gpt", "value": "there are two girls"}] | 
| 
	00000004420.jpg | 
	4420_evjvqa | 
	[{"from": "human", "value": "which way is the girl in the black t-shirt looking?\n<image>"}, {"from": "gpt", "value": "the clothes shop"}] | 
| 
	00000004421.jpg | 
	4421_evjvqa | 
	[{"from": "human", "value": "what do people put on the sidewalks on both sides of the road?\n<image>"}, {"from": "gpt", "value": "they plant the old trees"}] | 
| 
	00000004422.jpg | 
	4422_evjvqa | 
	[{"from": "human", "value": "what is the group of people standing?\n<image>"}, {"from": "gpt", "value": "at the fair area"}] | 
| 
	00000004422.jpg | 
	4422_evjvqa | 
	[{"from": "human", "value": "how many men are in this group of friends?\n<image>"}, {"from": "gpt", "value": "two"}] | 
| 
	00000004422.jpg | 
	4422_evjvqa | 
	[{"from": "human", "value": "how many girls are in this group of friends?\n<image>"}, {"from": "gpt", "value": "four"}] | 
| 
	00000004428.jpg | 
	4428_evjvqa | 
	[{"from": "human", "value": "where is this man taking the girl?\n<image>"}, {"from": "gpt", "value": "going to the supermarket to buy vegetable"}] | 
| 
	00000004428.jpg | 
	4428_evjvqa | 
	[{"from": "human", "value": "where does the girl go with the boy?\n<image>"}, {"from": "gpt", "value": "the supermarket"}] | 
| 
	00000004428.jpg | 
	4428_evjvqa | 
	[{"from": "human", "value": "what is the boy doing?\n<image>"}, {"from": "gpt", "value": "choosing vegetables"}] | 
| 
	00000004428.jpg | 
	4428_evjvqa | 
	[{"from": "human", "value": "what is the woman in the pink t-shirt doing?\n<image>"}, {"from": "gpt", "value": "choosing vegetable with the boy"}] | 
| 
	00000004428.jpg | 
	4428_evjvqa | 
	[{"from": "human", "value": "where are these men and women standing?\n<image>"}, {"from": "gpt", "value": "in the supermarket"}] | 
| 
	00000004433.jpg | 
	4433_evjvqa | 
	[{"from": "human", "value": "who is goinf to the supermarket with the man in blue t-shirt?\n<image>"}, {"from": "gpt", "value": "the man is wearing white t-shirt"}] | 
| 
	00000004433.jpg | 
	4433_evjvqa | 
	[{"from": "human", "value": "where is the boy in purple t-shirt with black mask standing?\n<image>"}, {"from": "gpt", "value": "at the vegetable stall in the supermarket"}] | 
| 
	00000004433.jpg | 
	4433_evjvqa | 
	[{"from": "human", "value": "what is the woman in yellow t-shirt doing?\n<image>"}, {"from": "gpt", "value": "she is choosing fruits"}] | 
| 
	00000004433.jpg | 
	4433_evjvqa | 
	[{"from": "human", "value": "what is the boy in white t-shirt going with the boy in blue t-shirt doing?\n<image>"}, {"from": "gpt", "value": "he is taking fruit and putting it in a bag"}] | 
| 
	00000004439.jpg | 
	4439_evjvqa | 
	[{"from": "human", "value": "what is the woman in yellow helmet doing?\n<image>"}, {"from": "gpt", "value": "in the restaurant"}] | 
| 
	00000004439.jpg | 
	4439_evjvqa | 
	[{"from": "human", "value": "what is the woman in white shirt doing?\n<image>"}, {"from": "gpt", "value": "in the restaurant"}] | 
| 
	00000004439.jpg | 
	4439_evjvqa | 
	[{"from": "human", "value": "who is the woman sitting at the table with?\n<image>"}, {"from": "gpt", "value": "the man is wearing white shirt"}] | 
| 
	00000004439.jpg | 
	4439_evjvqa | 
	[{"from": "human", "value": "whose car is parked in front of the restaurant?\n<image>"}, {"from": "gpt", "value": "it is of the woman in white shirt"}] | 
| 
	00000004439.jpg | 
	4439_evjvqa | 
	[{"from": "human", "value": "who is the man sitting at the table with?\n<image>"}, {"from": "gpt", "value": "the woman is wearing the white shirt"}] | 
| 
	00000004440.jpg | 
	4440_evjvqa | 
	[{"from": "human", "value": "where is the people standing?\n<image>"}, {"from": "gpt", "value": "these people are standing around the table of food"}] | 
| 
	00000004440.jpg | 
	4440_evjvqa | 
	[{"from": "human", "value": "what is on the table?\n<image>"}, {"from": "gpt", "value": "types of food"}] | 
| 
	00000004444.jpg | 
	4444_evjvqa | 
	[{"from": "human", "value": "what items the man in white shirt selling?\n<image>"}, {"from": "gpt", "value": "types of shoes"}] | 
| 
	00000004444.jpg | 
	4444_evjvqa | 
	[{"from": "human", "value": "where is the woman in ao dai visiting?\n<image>"}, {"from": "gpt", "value": "in the shopping fair"}] | 
| 
	00000004444.jpg | 
	4444_evjvqa | 
	[{"from": "human", "value": "where are the stalls sold?\n<image>"}, {"from": "gpt", "value": "in the shopping fair"}] | 
| 
	00000004445.jpg | 
	4445_evjvqa | 
	[{"from": "human", "value": "how many turns are there?\n<image>"}, {"from": "gpt", "value": "three"}] | 
| 
	00000004446.jpg | 
	4446_evjvqa | 
	[{"from": "human", "value": "what is the woman in yellow doing?\n<image>"}, {"from": "gpt", "value": "choosing the products"}] | 
| 
	00000004446.jpg | 
	4446_evjvqa | 
	[{"from": "human", "value": "where is the woman in the yellow shirt choosing goods?\n<image>"}, {"from": "gpt", "value": "in the supermarket"}] | 
| 
	00000004446.jpg | 
	4446_evjvqa | 
	[{"from": "human", "value": "which area is the woman in the yellow shirt choosing?\n<image>"}, {"from": "gpt", "value": "vegetable stall"}] | 
| 
	00000004446.jpg | 
	4446_evjvqa | 
	[{"from": "human", "value": "what is the woman in yellow intend to buy?\n<image>"}, {"from": "gpt", "value": "vegetable"}] | 
| 
	00000004448.jpg | 
	4448_evjvqa | 
	[{"from": "human", "value": "where is this view recorded?\n<image>"}, {"from": "gpt", "value": "the downwards"}] | 
| 
	00000004450.jpg | 
	4450_evjvqa | 
	[{"from": "human", "value": "what is the woman in conical hat doing?\n<image>"}, {"from": "gpt", "value": "she is rowing a canoe"}] | 
| 
	00000004450.jpg | 
	4450_evjvqa | 
	[{"from": "human", "value": "where is the woman in conical hat sitting?\n<image>"}, {"from": "gpt", "value": "on the boat"}] | 
| 
	00000004451.jpg | 
	4451_evjvqa | 
	[{"from": "human", "value": "what is this ship used to transport?\n<image>"}, {"from": "gpt", "value": "the containers"}] | 
| 
	00000004451.jpg | 
	4451_evjvqa | 
	[{"from": "human", "value": "where do people put the containers?\n<image>"}, {"from": "gpt", "value": "on the ship"}] | 
| 
	00000004452.jpg | 
	4452_evjvqa | 
	[{"from": "human", "value": "what is the woman in blue doing?\n<image>"}, {"from": "gpt", "value": "she is instructing the girl i red shirt to buy goods"}] | 
| 
	00000004452.jpg | 
	4452_evjvqa | 
	[{"from": "human", "value": "what is the girl in red doing?\n<image>"}, {"from": "gpt", "value": "choosing some fruits"}] | 
| 
	00000004452.jpg | 
	4452_evjvqa | 
	[{"from": "human", "value": "what item is the girl intending to buy?\n<image>"}, {"from": "gpt", "value": "lots of fruits"}] | 
| 
	00000004452.jpg | 
	4452_evjvqa | 
	[{"from": "human", "value": "what is the woman in the blue shirt working?\n<image>"}, {"from": "gpt", "value": "working as a fruit counter clerk in a supermarket"}] | 
| 
	00000004452.jpg | 
	4452_evjvqa | 
	[{"from": "human", "value": "where is the girl in red standing?\n<image>"}, {"from": "gpt", "value": "the fruit stall"}] | 
| 
	00000004452.jpg | 
	4452_evjvqa | 
	[{"from": "human", "value": "where is the woman in the blue shirt pointing?\n<image>"}, {"from": "gpt", "value": "the fruit stall"}] | 
| 
	00000004453.jpg | 
	4453_evjvqa | 
	[{"from": "human", "value": "which stall is the woman in black standing at?\n<image>"}, {"from": "gpt", "value": "the stuffed animal stall"}] | 
| 
	00000004457.jpg | 
	4457_evjvqa | 
	[{"from": "human", "value": "what do the stalls in this area sell?\n<image>"}, {"from": "gpt", "value": "selling all kinds of clothes and fabrics"}] | 
| 
	00000004457.jpg | 
	4457_evjvqa | 
	[{"from": "human", "value": "what do these people come here to buy?\n<image>"}, {"from": "gpt", "value": "choosing clothes"}] | 
| 
	00000004460.jpg | 
	4460_evjvqa | 
	[{"from": "human", "value": "what is this woman holding in her hand?\n<image>"}, {"from": "gpt", "value": "the gift "}] | 
| 
	00000004460.jpg | 
	4460_evjvqa | 
	[{"from": "human", "value": "where is the girl sitting?\n<image>"}, {"from": "gpt", "value": "in the room with a christmas tree"}] | 
| 
	00000004460.jpg | 
	4460_evjvqa | 
	[{"from": "human", "value": "what color shirt are you wearing?\n<image>"}, {"from": "gpt", "value": "red"}] | 
| 
	00000004461.jpg | 
	4461_evjvqa | 
	[{"from": "human", "value": "what are these trays used for?\n<image>"}, {"from": "gpt", "value": "to store all kinds of food"}] | 
| 
	00000004462.jpg | 
	4462_evjvqa | 
	[{"from": "human", "value": "where is the people visiting?\n<image>"}, {"from": "gpt", "value": "these people are sightseeing in a tunnel area"}] | 
| 
	00000004462.jpg | 
	4462_evjvqa | 
	[{"from": "human", "value": "what is the woman in white doing?\n<image>"}, {"from": "gpt", "value": "she is going down to the underground tunnel"}] | 
| 
	00000004464.jpg | 
	4464_evjvqa | 
	[{"from": "human", "value": "where is the people standing?\n<image>"}, {"from": "gpt", "value": "in the clothes shop"}] | 
| 
	00000004464.jpg | 
	4464_evjvqa | 
	[{"from": "human", "value": "how many men are standing in the clothes shop?\n<image>"}, {"from": "gpt", "value": "two"}] | 
| 
	00000004464.jpg | 
	4464_evjvqa | 
	[{"from": "human", "value": "how many women are standing in the clothes shop?\n<image>"}, {"from": "gpt", "value": "two"}] | 
| 
	00000004465.jpg | 
	4465_evjvqa | 
	[{"from": "human", "value": "where are these people standing?\n<image>"}, {"from": "gpt", "value": "on the beach"}] | 
| 
	00000004465.jpg | 
	4465_evjvqa | 
	[{"from": "human", "value": "what are two men using yellow plastic bags doing?\n<image>"}, {"from": "gpt", "value": "they are picking up trash on the beach"}] | 
| 
	00000004465.jpg | 
	4465_evjvqa | 
	[{"from": "human", "value": "what does the man in the white shirt use to contain garbage?\n<image>"}, {"from": "gpt", "value": "the yellow plastic bag"}] | 
| 
	00000004465.jpg | 
	4465_evjvqa | 
	[{"from": "human", "value": "what is the man in white shirt doing?\n<image>"}, {"from": "gpt", "value": "to pick up trash on the beach"}] | 
| 
	00000004466.jpg | 
	4466_evjvqa | 
	[{"from": "human", "value": "where is the girl in green standing?\n<image>"}, {"from": "gpt", "value": "in the grocery store"}] | 
| 
	00000004469.jpg | 
	4469_evjvqa | 
	[{"from": "human", "value": "where is the group of young people sitting to eat?\n<image>"}, {"from": "gpt", "value": "at a sidewalk restaurant"}] | 
| 
	00000004471.jpg | 
	4471_evjvqa | 
	[{"from": "human", "value": "what is the woman doing?\n<image>"}, {"from": "gpt", "value": "choosing beef"}] | 
| 
	00000004471.jpg | 
	4471_evjvqa | 
	[{"from": "human", "value": "what is the woman in black intend to buy?\n<image>"}, {"from": "gpt", "value": "lots of beef"}] | 
| 
	00000004471.jpg | 
	4471_evjvqa | 
	[{"from": "human", "value": "what do they sell at this stall?\n<image>"}, {"from": "gpt", "value": "types of beef"}] | 
| 
	00000004471.jpg | 
	4471_evjvqa | 
	[{"from": "human", "value": "what is the woman in black holding?\n<image>"}, {"from": "gpt", "value": "the woman is holding two blisters of raw meat"}] | 
| 
	00000004472.jpg | 
	4472_evjvqa | 
	[{"from": "human", "value": "where is the group of people standing?\n<image>"}, {"from": "gpt", "value": "on top of the tunnel"}] | 
| 
	00000004472.jpg | 
	4472_evjvqa | 
	[{"from": "human", "value": "where is the group of people intend visitting?\n<image>"}, {"from": "gpt", "value": "the historical monument"}] | 
| 
	00000004473.jpg | 
	4473_evjvqa | 
	[{"from": "human", "value": "where is the girl in black standing?\n<image>"}, {"from": "gpt", "value": "she is standing at the escalator in the supermarket"}] | 
| 
	00000004473.jpg | 
	4473_evjvqa | 
	[{"from": "human", "value": "who is going shopping with the girl in black?\n<image>"}, {"from": "gpt", "value": "the man"}] | 
| 
	00000004474.jpg | 
	4474_evjvqa | 
	[{"from": "human", "value": "where is the girl standing?\n<image>"}, {"from": "gpt", "value": "this girl is sitting near the christmas tree"}] | 
| 
	00000004477.jpg | 
	4477_evjvqa | 
	[{"from": "human", "value": "how many employees are standing at the reception?\n<image>"}, {"from": "gpt", "value": "there are two female staff"}] | 
| 
	00000004477.jpg | 
	4477_evjvqa | 
	[{"from": "human", "value": "how many people are standing at the freezer?\n<image>"}, {"from": "gpt", "value": "four"}] | 
| 
	00000004477.jpg | 
	4477_evjvqa | 
	[{"from": "human", "value": "what is the woman in the blue coat doing?\n<image>"}, {"from": "gpt", "value": "she is choosing the beef in the freezer"}] | 
| 
	00000004482.jpg | 
	4482_evjvqa | 
	[{"from": "human", "value": "what is the woman doing?\n<image>"}, {"from": "gpt", "value": "she is choosing some fresh food"}] | 
| 
	00000004482.jpg | 
	4482_evjvqa | 
	[{"from": "human", "value": "what item does the woman intend to buy?\n<image>"}, {"from": "gpt", "value": "some raw meat"}] | 
| 
	00000004482.jpg | 
	4482_evjvqa | 
	[{"from": "human", "value": "what does the stall next to the stall the woman is standing on sell for?\n<image>"}, {"from": "gpt", "value": "vegetables"}] | 
| 
	00000004485.jpg | 
	4485_evjvqa | 
	[{"from": "human", "value": "what color coat does the woman selling bouquet wear?\n<image>"}, {"from": "gpt", "value": "the pink overcoat"}] | 
| 
	00000004485.jpg | 
	4485_evjvqa | 
	[{"from": "human", "value": "what is the woman in pink jacket doing?\n<image>"}, {"from": "gpt", "value": "she is handing the bouquet of flowers to the woman wearing the helmet"}] | 
| 
	00000004485.jpg | 
	4485_evjvqa | 
	[{"from": "human", "value": "what item does the woman in black jacket buy?\n<image>"}, {"from": "gpt", "value": "the woman in black coat is buying a bouquet of flowers"}] | 
| 
	00000004485.jpg | 
	4485_evjvqa | 
	[{"from": "human", "value": "where is the woman in black standing?\n<image>"}, {"from": "gpt", "value": "at the bouquet stall"}] | 
| 
	00000004487.jpg | 
	4487_evjvqa | 
	[{"from": "human", "value": "what are the two teenagers doing?\n<image>"}, {"from": "gpt", "value": "they are taking a photos"}] | 
| 
	00000004487.jpg | 
	4487_evjvqa | 
	[{"from": "human", "value": "who is photopraphed by the boy?\n<image>"}, {"from": "gpt", "value": "the girl"}] | 
| 
	00000004487.jpg | 
	4487_evjvqa | 
	[{"from": "human", "value": "who is photopraphed by the girl?\n<image>"}, {"from": "gpt", "value": "the boy"}] | 
| 
	00000004487.jpg | 
	4487_evjvqa | 
	[{"from": "human", "value": "what is the boy holding the camera doing?\n<image>"}, {"from": "gpt", "value": "to take a photo for the girl"}] | 
| 
	00000004489.jpg | 
	4489_evjvqa | 
	[{"from": "human", "value": "which are the stalls selling?\n<image>"}, {"from": "gpt", "value": "clothes"}] | 
| 
	00000004492.jpg | 
	4492_evjvqa | 
	[{"from": "human", "value": "what is the boy doing?\n<image>"}, {"from": "gpt", "value": "he is selling the balloons"}] | 
| 
	00000004492.jpg | 
	4492_evjvqa | 
	[{"from": "human", "value": "where is the boy selling the balloons?\n<image>"}, {"from": "gpt", "value": "at the foot of the bridge"}] | 
| 
	00000004492.jpg | 
	4492_evjvqa | 
	[{"from": "human", "value": "what does this boy stand by the roadside selling?\n<image>"}, {"from": "gpt", "value": "selling the ballons"}] | 
| 
	00000004493.jpg | 
	4493_evjvqa | 
	[{"from": "human", "value": "how many men are standing in the shop?\n<image>"}, {"from": "gpt", "value": "two"}] | 
| 
	00000004493.jpg | 
	4493_evjvqa | 
	[{"from": "human", "value": "what are the two men doing?\n<image>"}, {"from": "gpt", "value": "they are choosing poducts"}] | 
| 
	00000004493.jpg | 
	4493_evjvqa | 
	[{"from": "human", "value": "where are the two men standing?\n<image>"}, {"from": "gpt", "value": "at the stall in the market"}] | 
| 
	00000004495.jpg | 
	4495_evjvqa | 
	[{"from": "human", "value": "where is the woman in the conical hat sitting?\n<image>"}, {"from": "gpt", "value": "on the sidewalk"}] | 
| 
	00000004495.jpg | 
	4495_evjvqa | 
	[{"from": "human", "value": "what is the woman in the conical hat doing?\n<image>"}, {"from": "gpt", "value": "she is selling on the sidewalk"}] | 
			Subsets and Splits
				
	
				
			
				
No community queries yet
The top public SQL queries from the community will appear here once available.
