task stringclasses 60 values | visual_input_component stringclasses 21 values | source stringclasses 64 values | options stringlengths 12 4.46k | question stringlengths 17 1.33k | context stringlengths 24 21.8k | input_image_path listlengths 1 62 | output stringclasses 10 values |
|---|---|---|---|---|---|---|---|
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.395, 0.317, 0.602, 0.821]
B: [0.504, 0.408, 0.513, 0.686]
C: [0.484, 0.439, 0.69, 0.943]
D: [0.313, 0.244, 0.52, 0.749] | Here is an object ([0.429, 0.154, 0.625, 0.786]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.395, 0.317, 0.602, 0.821]
B: [0.504, 0.408, 0.513, 0.686]
C: [0.484, 0.439, 0.69, 0.943]
D: [0.313, 0.244, 0.52, 0.749] | [
"./2D-spatial/single_object_tracking/single_object_tracking_134_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_134_1.jpg"
] | A |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.13, 0.0, 0.852, 1.0]
B: [0.071, 0.0, 0.793, 1.0]
C: [0.095, 0.306, 0.59, 0.322]
D: [0.98, 0.435, 0.996, 0.803] | Here is an object ([0.063, 0.0, 1.0, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.13, 0.0, 0.852, 1.0]
B: [0.071, 0.0, 0.793, 1.0]
C: [0.095, 0.306, 0.59, 0.322]
D: [0.98, 0.435, 0.996, 0.803] | [
"./2D-spatial/single_object_tracking/single_object_tracking_135_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_135_1.jpg"
] | B |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.277, 0.0, 0.519, 0.45]
B: [0.395, 0.013, 0.637, 0.463]
C: [0.497, 0.199, 0.843, 0.696]
D: [0.281, 0.114, 0.523, 0.564] | Here is an object ([0.264, 0.0, 0.491, 0.404]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.277, 0.0, 0.519, 0.45]
B: [0.395, 0.013, 0.637, 0.463]
C: [0.497, 0.199, 0.843, 0.696]
D: [0.281, 0.114, 0.523, 0.564] | [
"./2D-spatial/single_object_tracking/single_object_tracking_136_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_136_1.jpg"
] | A |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.901, 0.401, 0.985, 1.051]
B: [0.901, 0.401, 1.0, 1.0]
C: [0.504, 0.157, 0.877, 0.589]
D: [0.901, 0.206, 1.0, 0.804] | Here is an object ([0.934, 0.432, 1.0, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.901, 0.401, 0.985, 1.051]
B: [0.901, 0.401, 1.0, 1.0]
C: [0.504, 0.157, 0.877, 0.589]
D: [0.901, 0.206, 1.0, 0.804] | [
"./2D-spatial/single_object_tracking/single_object_tracking_137_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_137_1.jpg"
] | B |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.0, 0.267, 0.299, 0.561]
B: [0.0, 0.267, 0.309, 0.537]
C: [0.0, 0.267, 0.323, 0.568]
D: [0.0, 0.171, 0.323, 0.472] | Here is an object ([0.0, 0.246, 0.424, 0.611]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.0, 0.267, 0.299, 0.561]
B: [0.0, 0.267, 0.309, 0.537]
C: [0.0, 0.267, 0.323, 0.568]
D: [0.0, 0.171, 0.323, 0.472] | [
"./2D-spatial/single_object_tracking/single_object_tracking_138_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_138_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.0, 0.0, 0.606, 1.0]
B: [0.502, 0.601, 0.622, 0.924]
C: [0.287, 0.311, 0.747, 0.39]
D: [0.0, 0.0, 0.535, 1.157] | Here is an object ([0.0, 0.0, 0.923, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.0, 0.0, 0.606, 1.0]
B: [0.502, 0.601, 0.622, 0.924]
C: [0.287, 0.311, 0.747, 0.39]
D: [0.0, 0.0, 0.535, 1.157] | [
"./2D-spatial/single_object_tracking/single_object_tracking_139_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_139_1.jpg"
] | A |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.05, 0.728, 0.195, 0.956]
B: [0.193, 0.054, 0.217, 0.426]
C: [0.434, 0.371, 0.787, 1.0]
D: [0.519, 0.371, 0.872, 1.0] | Here is an object ([0.529, 0.507, 0.775, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.05, 0.728, 0.195, 0.956]
B: [0.193, 0.054, 0.217, 0.426]
C: [0.434, 0.371, 0.787, 1.0]
D: [0.519, 0.371, 0.872, 1.0] | [
"./2D-spatial/single_object_tracking/single_object_tracking_140_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_140_1.jpg"
] | D |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.409, 0.479, 0.546, 0.554]
B: [0.409, 0.479, 0.537, 0.55]
C: [0.429, 0.487, 0.557, 0.558]
D: [0.409, 0.479, 0.516, 0.56] | Here is an object ([0.455, 0.471, 0.564, 0.543]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.409, 0.479, 0.546, 0.554]
B: [0.409, 0.479, 0.537, 0.55]
C: [0.429, 0.487, 0.557, 0.558]
D: [0.409, 0.479, 0.516, 0.56] | [
"./2D-spatial/single_object_tracking/single_object_tracking_141_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_141_1.jpg"
] | B |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.514, 0.244, 0.854, 0.649]
B: [0.601, 0.221, 1.0, 0.662]
C: [0.514, 0.244, 0.913, 0.686]
D: [0.601, 0.308, 1.0, 0.75] | Here is an object ([0.589, 0.235, 0.943, 0.722]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.514, 0.244, 0.854, 0.649]
B: [0.601, 0.221, 1.0, 0.662]
C: [0.514, 0.244, 0.913, 0.686]
D: [0.601, 0.308, 1.0, 0.75] | [
"./2D-spatial/single_object_tracking/single_object_tracking_142_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_142_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.488, 0.207, 0.569, 0.358]
B: [0.469, 0.228, 0.549, 0.379]
C: [0.432, 0.458, 0.816, 0.517]
D: [0.019, 0.432, 0.448, 0.564] | Here is an object ([0.496, 0.242, 0.566, 0.381]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.488, 0.207, 0.569, 0.358]
B: [0.469, 0.228, 0.549, 0.379]
C: [0.432, 0.458, 0.816, 0.517]
D: [0.019, 0.432, 0.448, 0.564] | [
"./2D-spatial/single_object_tracking/single_object_tracking_143_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_143_1.jpg"
] | A |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.0, 0.161, 0.049, 0.542]
B: [0.699, 0.242, 0.79, 0.568]
C: [0.0, 0.099, 0.049, 0.479]
D: [0.0, 0.101, 0.049, 0.482] | Here is an object ([0.0, 0.094, 0.1, 0.554]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.0, 0.161, 0.049, 0.542]
B: [0.699, 0.242, 0.79, 0.568]
C: [0.0, 0.099, 0.049, 0.479]
D: [0.0, 0.101, 0.049, 0.482] | [
"./2D-spatial/single_object_tracking/single_object_tracking_144_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_144_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.377, 0.336, 1.024, 0.835]
B: [0.377, 0.336, 0.956, 0.956]
C: [0.377, 0.336, 1.061, 1.003]
D: [0.101, 0.381, 0.68, 1.0] | Here is an object ([0.433, 0.271, 0.981, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.377, 0.336, 1.024, 0.835]
B: [0.377, 0.336, 0.956, 0.956]
C: [0.377, 0.336, 1.061, 1.003]
D: [0.101, 0.381, 0.68, 1.0] | [
"./2D-spatial/single_object_tracking/single_object_tracking_145_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_145_1.jpg"
] | B |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.364, 0.487, 0.55, 0.693]
B: [0.364, 0.487, 0.529, 0.668]
C: [0.273, 0.447, 0.459, 0.653]
D: [0.378, 0.558, 0.564, 0.764] | Here is an object ([0.342, 0.415, 0.542, 0.607]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.364, 0.487, 0.55, 0.693]
B: [0.364, 0.487, 0.529, 0.668]
C: [0.273, 0.447, 0.459, 0.653]
D: [0.378, 0.558, 0.564, 0.764] | [
"./2D-spatial/single_object_tracking/single_object_tracking_146_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_146_1.jpg"
] | A |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.116, 0.26, 0.833, 0.936]
B: [0.116, 0.26, 0.734, 1.0]
C: [0.0, 0.26, 0.619, 1.0]
D: [0.116, 0.626, 0.322, 0.66] | Here is an object ([0.113, 0.256, 0.725, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.116, 0.26, 0.833, 0.936]
B: [0.116, 0.26, 0.734, 1.0]
C: [0.0, 0.26, 0.619, 1.0]
D: [0.116, 0.626, 0.322, 0.66] | [
"./2D-spatial/single_object_tracking/single_object_tracking_147_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_147_1.jpg"
] | B |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.409, 0.407, 0.471, 0.524]
B: [0.402, 0.449, 0.465, 0.565]
C: [0.404, 0.357, 0.466, 0.474]
D: [0.137, 0.357, 0.261, 0.697] | Here is an object ([0.479, 0.539, 0.527, 0.662]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.409, 0.407, 0.471, 0.524]
B: [0.402, 0.449, 0.465, 0.565]
C: [0.404, 0.357, 0.466, 0.474]
D: [0.137, 0.357, 0.261, 0.697] | [
"./2D-spatial/single_object_tracking/single_object_tracking_148_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_148_1.jpg"
] | A |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.453, 0.503, 0.507, 0.681]
B: [0.128, 0.867, 0.552, 0.899]
C: [0.276, 0.35, 0.747, 0.397]
D: [0.453, 0.503, 0.503, 0.706] | Here is an object ([0.487, 0.506, 0.544, 0.672]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.453, 0.503, 0.507, 0.681]
B: [0.128, 0.867, 0.552, 0.899]
C: [0.276, 0.35, 0.747, 0.397]
D: [0.453, 0.503, 0.503, 0.706] | [
"./2D-spatial/single_object_tracking/single_object_tracking_149_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_149_1.jpg"
] | A |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.22, 0.242, 0.862, 0.635]
B: [0.161, 0.114, 0.634, 0.354]
C: [0.562, 0.422, 0.925, 0.835]
D: [0.359, 0.388, 1.0, 0.781] | Here is an object ([0.209, 0.215, 0.863, 0.618]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.22, 0.242, 0.862, 0.635]
B: [0.161, 0.114, 0.634, 0.354]
C: [0.562, 0.422, 0.925, 0.835]
D: [0.359, 0.388, 1.0, 0.781] | [
"./2D-spatial/single_object_tracking/single_object_tracking_150_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_150_1.jpg"
] | A |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.285, 0.511, 1.0, 0.756]
B: [0.606, 0.539, 0.62, 0.972]
C: [0.22, 0.585, 0.935, 0.829]
D: [0.285, 0.511, 1.085, 0.719] | Here is an object ([0.435, 0.412, 1.0, 0.749]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.285, 0.511, 1.0, 0.756]
B: [0.606, 0.539, 0.62, 0.972]
C: [0.22, 0.585, 0.935, 0.829]
D: [0.285, 0.511, 1.085, 0.719] | [
"./2D-spatial/single_object_tracking/single_object_tracking_151_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_151_1.jpg"
] | A |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.209, 0.343, 0.797, 0.886]
B: [0.028, 0.369, 0.616, 0.912]
C: [0.0, 0.146, 0.588, 0.689]
D: [0.337, 0.056, 0.549, 0.196] | Here is an object ([0.021, 0.375, 0.605, 0.915]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.209, 0.343, 0.797, 0.886]
B: [0.028, 0.369, 0.616, 0.912]
C: [0.0, 0.146, 0.588, 0.689]
D: [0.337, 0.056, 0.549, 0.196] | [
"./2D-spatial/single_object_tracking/single_object_tracking_152_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_152_1.jpg"
] | B |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.36, 0.068, 0.727, 0.486]
B: [0.048, 0.221, 0.545, 0.911]
C: [0.116, 0.31, 0.613, 1.0]
D: [0.116, 0.31, 0.68, 1.039] | Here is an object ([0.116, 0.312, 0.606, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.36, 0.068, 0.727, 0.486]
B: [0.048, 0.221, 0.545, 0.911]
C: [0.116, 0.31, 0.613, 1.0]
D: [0.116, 0.31, 0.68, 1.039] | [
"./2D-spatial/single_object_tracking/single_object_tracking_153_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_153_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.251, 0.228, 1.0, 1.0]
B: [0.0, 0.0, 0.749, 0.772]
C: [0.0, 0.228, 0.749, 1.0]
D: [0.0, 0.113, 0.749, 0.885] | Here is an object ([0.0, 0.119, 0.75, 0.885]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 960 and the height is 720. | Select from the following choices.
A: [0.251, 0.228, 1.0, 1.0]
B: [0.0, 0.0, 0.749, 0.772]
C: [0.0, 0.228, 0.749, 1.0]
D: [0.0, 0.113, 0.749, 0.885] | [
"./2D-spatial/single_object_tracking/single_object_tracking_154_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_154_1.jpg"
] | D |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.084, 0.149, 0.391, 0.438]
B: [0.0, 0.071, 0.836, 1.133]
C: [0.0, 0.071, 0.905, 1.0]
D: [0.095, 0.0, 1.0, 0.929] | Here is an object ([0.0, 0.001, 0.894, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.084, 0.149, 0.391, 0.438]
B: [0.0, 0.071, 0.836, 1.133]
C: [0.0, 0.071, 0.905, 1.0]
D: [0.095, 0.0, 1.0, 0.929] | [
"./2D-spatial/single_object_tracking/single_object_tracking_155_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_155_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.066, 0.203, 0.189, 0.603]
B: [0.204, 0.146, 0.611, 1.0]
C: [0.03, 0.146, 0.437, 1.0]
D: [0.03, 0.146, 0.445, 0.939] | Here is an object ([0.034, 0.21, 0.511, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.066, 0.203, 0.189, 0.603]
B: [0.204, 0.146, 0.611, 1.0]
C: [0.03, 0.146, 0.437, 1.0]
D: [0.03, 0.146, 0.445, 0.939] | [
"./2D-spatial/single_object_tracking/single_object_tracking_156_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_156_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.3, 0.251, 0.613, 1.0]
B: [0.39, 0.0, 0.712, 0.697]
C: [0.708, 0.621, 0.739, 0.844]
D: [0.39, 0.0, 0.703, 0.749] | Here is an object ([0.242, 0.0, 0.613, 0.656]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.3, 0.251, 0.613, 1.0]
B: [0.39, 0.0, 0.712, 0.697]
C: [0.708, 0.621, 0.739, 0.844]
D: [0.39, 0.0, 0.703, 0.749] | [
"./2D-spatial/single_object_tracking/single_object_tracking_157_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_157_1.jpg"
] | D |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.408, 0.212, 0.655, 0.739]
B: [0.17, 0.383, 0.197, 0.639]
C: [0.408, 0.212, 0.661, 0.754]
D: [0.408, 0.212, 0.665, 0.856] | Here is an object ([0.403, 0.207, 0.651, 0.767]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.408, 0.212, 0.655, 0.739]
B: [0.17, 0.383, 0.197, 0.639]
C: [0.408, 0.212, 0.661, 0.754]
D: [0.408, 0.212, 0.665, 0.856] | [
"./2D-spatial/single_object_tracking/single_object_tracking_158_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_158_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.594, 0.279, 0.91, 0.968]
B: [0.486, 0.013, 0.765, 0.59]
C: [0.446, 0.122, 0.805, 0.543]
D: [0.594, 0.279, 0.872, 0.857] | Here is an object ([0.596, 0.289, 0.867, 0.853]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1270 and the height is 720. | Select from the following choices.
A: [0.594, 0.279, 0.91, 0.968]
B: [0.486, 0.013, 0.765, 0.59]
C: [0.446, 0.122, 0.805, 0.543]
D: [0.594, 0.279, 0.872, 0.857] | [
"./2D-spatial/single_object_tracking/single_object_tracking_159_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_159_1.jpg"
] | D |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.002, 0.087, 0.223, 0.472]
B: [0.194, 0.114, 0.683, 0.775]
C: [0.069, 0.221, 0.233, 0.621]
D: [0.179, 0.339, 0.668, 1.0] | Here is an object ([0.228, 0.0, 0.719, 0.607]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.002, 0.087, 0.223, 0.472]
B: [0.194, 0.114, 0.683, 0.775]
C: [0.069, 0.221, 0.233, 0.621]
D: [0.179, 0.339, 0.668, 1.0] | [
"./2D-spatial/single_object_tracking/single_object_tracking_160_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_160_1.jpg"
] | B |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.459, 0.085, 0.476, 0.549]
B: [0.248, 0.667, 0.747, 0.828]
C: [0.512, 0.371, 0.652, 0.542]
D: [0.512, 0.371, 0.626, 0.522] | Here is an object ([0.509, 0.357, 0.635, 0.535]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.459, 0.085, 0.476, 0.549]
B: [0.248, 0.667, 0.747, 0.828]
C: [0.512, 0.371, 0.652, 0.542]
D: [0.512, 0.371, 0.626, 0.522] | [
"./2D-spatial/single_object_tracking/single_object_tracking_161_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_161_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.32, 0.046, 0.584, 0.879]
B: [0.177, 0.0, 0.491, 0.917]
C: [0.494, 0.643, 0.716, 0.814]
D: [0.32, 0.046, 0.634, 0.963] | Here is an object ([0.324, 0.046, 0.635, 0.968]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.32, 0.046, 0.584, 0.879]
B: [0.177, 0.0, 0.491, 0.917]
C: [0.494, 0.643, 0.716, 0.814]
D: [0.32, 0.046, 0.634, 0.963] | [
"./2D-spatial/single_object_tracking/single_object_tracking_162_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_162_1.jpg"
] | D |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.43, 0.485, 0.783, 0.656]
B: [0.502, 0.41, 0.579, 0.64]
C: [0.463, 0.338, 0.54, 0.568]
D: [0.488, 0.294, 0.566, 0.525] | Here is an object ([0.476, 0.335, 0.562, 0.568]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.43, 0.485, 0.783, 0.656]
B: [0.502, 0.41, 0.579, 0.64]
C: [0.463, 0.338, 0.54, 0.568]
D: [0.488, 0.294, 0.566, 0.525] | [
"./2D-spatial/single_object_tracking/single_object_tracking_163_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_163_1.jpg"
] | D |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.461, 0.357, 0.537, 0.714]
B: [0.461, 0.357, 0.526, 0.771]
C: [0.095, 0.572, 0.489, 0.808]
D: [0.465, 0.401, 0.541, 0.758] | Here is an object ([0.466, 0.358, 0.545, 0.706]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.461, 0.357, 0.537, 0.714]
B: [0.461, 0.357, 0.526, 0.771]
C: [0.095, 0.572, 0.489, 0.808]
D: [0.465, 0.401, 0.541, 0.758] | [
"./2D-spatial/single_object_tracking/single_object_tracking_164_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_164_1.jpg"
] | A |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.442, 0.604, 0.598, 0.832]
B: [0.49, 0.487, 0.658, 0.771]
C: [0.442, 0.604, 0.61, 0.887]
D: [0.409, 0.69, 0.577, 0.974] | Here is an object ([0.455, 0.621, 0.626, 0.886]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.442, 0.604, 0.598, 0.832]
B: [0.49, 0.487, 0.658, 0.771]
C: [0.442, 0.604, 0.61, 0.887]
D: [0.409, 0.69, 0.577, 0.974] | [
"./2D-spatial/single_object_tracking/single_object_tracking_165_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_165_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.416, 0.133, 0.897, 0.514]
B: [0.416, 0.133, 0.995, 0.537]
C: [0.433, 0.497, 0.685, 0.806]
D: [0.421, 0.0, 1.0, 0.404] | Here is an object ([0.436, 0.083, 0.995, 0.561]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 406 and the height is 720. | Select from the following choices.
A: [0.416, 0.133, 0.897, 0.514]
B: [0.416, 0.133, 0.995, 0.537]
C: [0.433, 0.497, 0.685, 0.806]
D: [0.421, 0.0, 1.0, 0.404] | [
"./2D-spatial/single_object_tracking/single_object_tracking_166_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_166_1.jpg"
] | B |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.228, 0.108, 0.396, 0.479]
B: [0.171, 0.0, 0.923, 0.742]
C: [0.171, 0.093, 1.0, 0.824]
D: [0.171, 0.0, 1.0, 0.731] | Here is an object ([0.165, 0.0, 1.0, 0.726]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.228, 0.108, 0.396, 0.479]
B: [0.171, 0.0, 0.923, 0.742]
C: [0.171, 0.093, 1.0, 0.824]
D: [0.171, 0.0, 1.0, 0.731] | [
"./2D-spatial/single_object_tracking/single_object_tracking_167_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_167_1.jpg"
] | D |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.074, 0.186, 0.488, 1.0]
B: [0.058, 0.151, 0.472, 0.965]
C: [0.159, 0.186, 0.639, 0.935]
D: [0.159, 0.186, 0.573, 1.0] | Here is an object ([0.179, 0.022, 0.554, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.074, 0.186, 0.488, 1.0]
B: [0.058, 0.151, 0.472, 0.965]
C: [0.159, 0.186, 0.639, 0.935]
D: [0.159, 0.186, 0.573, 1.0] | [
"./2D-spatial/single_object_tracking/single_object_tracking_168_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_168_1.jpg"
] | D |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.0, 0.287, 0.342, 0.665]
B: [0.078, 0.428, 0.42, 0.806]
C: [0.34, 0.412, 0.643, 0.438]
D: [0.0, 0.287, 0.341, 0.682] | Here is an object ([0.0, 0.297, 0.397, 0.665]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.0, 0.287, 0.342, 0.665]
B: [0.078, 0.428, 0.42, 0.806]
C: [0.34, 0.412, 0.643, 0.438]
D: [0.0, 0.287, 0.341, 0.682] | [
"./2D-spatial/single_object_tracking/single_object_tracking_169_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_169_1.jpg"
] | A |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.464, 0.276, 0.727, 1.0]
B: [0.464, 0.276, 0.745, 0.993]
C: [0.517, 0.276, 0.78, 1.0]
D: [0.464, 0.276, 0.692, 0.875] | Here is an object ([0.455, 0.276, 0.688, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.464, 0.276, 0.727, 1.0]
B: [0.464, 0.276, 0.745, 0.993]
C: [0.517, 0.276, 0.78, 1.0]
D: [0.464, 0.276, 0.692, 0.875] | [
"./2D-spatial/single_object_tracking/single_object_tracking_170_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_170_1.jpg"
] | A |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.534, 0.237, 0.687, 0.515]
B: [0.534, 0.237, 0.662, 0.522]
C: [0.534, 0.237, 0.641, 0.497]
D: [0.499, 0.261, 0.628, 0.546] | Here is an object ([0.58, 0.235, 0.755, 0.518]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.534, 0.237, 0.687, 0.515]
B: [0.534, 0.237, 0.662, 0.522]
C: [0.534, 0.237, 0.641, 0.497]
D: [0.499, 0.261, 0.628, 0.546] | [
"./2D-spatial/single_object_tracking/single_object_tracking_171_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_171_1.jpg"
] | B |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.081, 0.196, 0.637, 1.131]
B: [0.081, 0.196, 0.748, 1.113]
C: [0.081, 0.196, 0.658, 0.994]
D: [0.611, 0.761, 0.737, 0.843] | Here is an object ([0.136, 0.15, 0.672, 0.881]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.081, 0.196, 0.637, 1.131]
B: [0.081, 0.196, 0.748, 1.113]
C: [0.081, 0.196, 0.658, 0.994]
D: [0.611, 0.761, 0.737, 0.843] | [
"./2D-spatial/single_object_tracking/single_object_tracking_172_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_172_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.23, 0.069, 0.931, 1.0]
B: [0.23, 0.069, 0.792, 1.121]
C: [0.218, 0.069, 0.919, 1.0]
D: [0.457, 0.265, 0.69, 0.581] | Here is an object ([0.231, 0.124, 0.86, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.23, 0.069, 0.931, 1.0]
B: [0.23, 0.069, 0.792, 1.121]
C: [0.218, 0.069, 0.919, 1.0]
D: [0.457, 0.265, 0.69, 0.581] | [
"./2D-spatial/single_object_tracking/single_object_tracking_173_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_173_1.jpg"
] | A |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.159, 0.225, 0.294, 0.533]
B: [0.218, 0.453, 0.636, 0.631]
C: [0.292, 0.406, 0.459, 0.643]
D: [0.292, 0.406, 0.456, 0.7] | Here is an object ([0.29, 0.426, 0.471, 0.7]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.159, 0.225, 0.294, 0.533]
B: [0.218, 0.453, 0.636, 0.631]
C: [0.292, 0.406, 0.459, 0.643]
D: [0.292, 0.406, 0.456, 0.7] | [
"./2D-spatial/single_object_tracking/single_object_tracking_174_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_174_1.jpg"
] | D |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.143, 0.454, 0.249, 0.654]
B: [0.077, 0.669, 0.136, 0.985]
C: [0.145, 0.525, 0.252, 0.725]
D: [0.143, 0.454, 0.266, 0.657] | Here is an object ([0.12, 0.461, 0.237, 0.653]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.143, 0.454, 0.249, 0.654]
B: [0.077, 0.669, 0.136, 0.985]
C: [0.145, 0.525, 0.252, 0.725]
D: [0.143, 0.454, 0.266, 0.657] | [
"./2D-spatial/single_object_tracking/single_object_tracking_175_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_175_1.jpg"
] | A |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.602, 0.0, 0.946, 0.739]
B: [0.468, 0.376, 0.48, 0.842]
C: [0.44, 0.261, 0.783, 1.0]
D: [0.393, 0.261, 0.736, 1.0] | Here is an object ([0.446, 0.211, 0.622, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 960 and the height is 720. | Select from the following choices.
A: [0.602, 0.0, 0.946, 0.739]
B: [0.468, 0.376, 0.48, 0.842]
C: [0.44, 0.261, 0.783, 1.0]
D: [0.393, 0.261, 0.736, 1.0] | [
"./2D-spatial/single_object_tracking/single_object_tracking_176_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_176_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.0, 0.001, 0.722, 1.126]
B: [0.0, 0.001, 0.598, 1.193]
C: [0.0, 0.001, 0.724, 0.999]
D: [0.0, 0.001, 0.738, 1.196] | Here is an object ([0.0, 0.0, 0.755, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.0, 0.001, 0.722, 1.126]
B: [0.0, 0.001, 0.598, 1.193]
C: [0.0, 0.001, 0.724, 0.999]
D: [0.0, 0.001, 0.738, 1.196] | [
"./2D-spatial/single_object_tracking/single_object_tracking_177_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_177_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.183, 0.761, 0.235, 0.919]
B: [0.683, 0.257, 0.857, 0.718]
C: [0.351, 0.0, 1.0, 1.0]
D: [0.351, 0.0, 0.877, 0.803] | Here is an object ([0.313, 0.0, 1.0, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.183, 0.761, 0.235, 0.919]
B: [0.683, 0.257, 0.857, 0.718]
C: [0.351, 0.0, 1.0, 1.0]
D: [0.351, 0.0, 0.877, 0.803] | [
"./2D-spatial/single_object_tracking/single_object_tracking_178_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_178_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.334, 0.014, 1.103, 1.108]
B: [0.2, 0.281, 0.454, 0.629]
C: [0.334, 0.014, 0.926, 0.993]
D: [0.334, 0.014, 1.0, 1.0] | Here is an object ([0.235, 0.001, 1.0, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.334, 0.014, 1.103, 1.108]
B: [0.2, 0.281, 0.454, 0.629]
C: [0.334, 0.014, 0.926, 0.993]
D: [0.334, 0.014, 1.0, 1.0] | [
"./2D-spatial/single_object_tracking/single_object_tracking_179_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_179_1.jpg"
] | D |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.356, 0.011, 0.83, 0.357]
B: [0.183, 0.207, 0.581, 1.011]
C: [0.183, 0.207, 0.68, 0.996]
D: [0.183, 0.207, 0.616, 1.11] | Here is an object ([0.211, 0.165, 0.67, 0.982]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.356, 0.011, 0.83, 0.357]
B: [0.183, 0.207, 0.581, 1.011]
C: [0.183, 0.207, 0.68, 0.996]
D: [0.183, 0.207, 0.616, 1.11] | [
"./2D-spatial/single_object_tracking/single_object_tracking_180_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_180_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.479, 0.108, 0.884, 0.665]
B: [0.552, 0.097, 0.956, 0.654]
C: [0.317, 0.204, 0.722, 0.761]
D: [0.479, 0.108, 0.859, 0.699] | Here is an object ([0.457, 0.218, 0.777, 0.725]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.479, 0.108, 0.884, 0.665]
B: [0.552, 0.097, 0.956, 0.654]
C: [0.317, 0.204, 0.722, 0.761]
D: [0.479, 0.108, 0.859, 0.699] | [
"./2D-spatial/single_object_tracking/single_object_tracking_181_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_181_1.jpg"
] | A |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.603, 0.522, 0.715, 0.79]
B: [0.531, 0.461, 0.641, 0.671]
C: [0.523, 0.396, 0.632, 0.606]
D: [0.537, 0.519, 0.702, 0.668] | Here is an object ([0.584, 0.392, 0.634, 0.551]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.603, 0.522, 0.715, 0.79]
B: [0.531, 0.461, 0.641, 0.671]
C: [0.523, 0.396, 0.632, 0.606]
D: [0.537, 0.519, 0.702, 0.668] | [
"./2D-spatial/single_object_tracking/single_object_tracking_182_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_182_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.0, 0.013, 0.871, 1.0]
B: [0.129, 0.013, 1.047, 0.982]
C: [0.696, 0.489, 0.793, 0.943]
D: [0.129, 0.013, 1.0, 1.0] | Here is an object ([0.059, 0.0, 1.0, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.0, 0.013, 0.871, 1.0]
B: [0.129, 0.013, 1.047, 0.982]
C: [0.696, 0.489, 0.793, 0.943]
D: [0.129, 0.013, 1.0, 1.0] | [
"./2D-spatial/single_object_tracking/single_object_tracking_183_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_183_1.jpg"
] | D |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.42, 0.268, 0.625, 0.971]
B: [0.42, 0.268, 0.636, 0.778]
C: [0.42, 0.268, 0.66, 0.865]
D: [0.42, 0.268, 0.639, 0.919] | Here is an object ([0.411, 0.272, 0.654, 0.865]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 960 and the height is 720. | Select from the following choices.
A: [0.42, 0.268, 0.625, 0.971]
B: [0.42, 0.268, 0.636, 0.778]
C: [0.42, 0.268, 0.66, 0.865]
D: [0.42, 0.268, 0.639, 0.919] | [
"./2D-spatial/single_object_tracking/single_object_tracking_184_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_184_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.174, 0.0, 0.691, 0.558]
B: [0.483, 0.21, 1.0, 0.768]
C: [0.382, 0.046, 0.899, 0.604]
D: [0.432, 0.364, 0.76, 0.779] | Here is an object ([0.384, 0.018, 0.968, 0.479]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.174, 0.0, 0.691, 0.558]
B: [0.483, 0.21, 1.0, 0.768]
C: [0.382, 0.046, 0.899, 0.604]
D: [0.432, 0.364, 0.76, 0.779] | [
"./2D-spatial/single_object_tracking/single_object_tracking_185_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_185_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.595, 0.536, 0.865, 0.782]
B: [0.595, 0.536, 0.829, 0.744]
C: [0.074, 0.478, 0.275, 0.861]
D: [0.059, 0.325, 0.287, 0.339] | Here is an object ([0.487, 0.554, 0.705, 0.758]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.595, 0.536, 0.865, 0.782]
B: [0.595, 0.536, 0.829, 0.744]
C: [0.074, 0.478, 0.275, 0.861]
D: [0.059, 0.325, 0.287, 0.339] | [
"./2D-spatial/single_object_tracking/single_object_tracking_186_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_186_1.jpg"
] | B |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.384, 0.524, 0.459, 0.842]
B: [0.0, 0.0, 0.77, 0.999]
C: [0.126, 0.49, 0.423, 0.603]
D: [0.0, 0.0, 0.685, 0.894] | Here is an object ([0.0, 0.0, 0.784, 0.999]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.384, 0.524, 0.459, 0.842]
B: [0.0, 0.0, 0.77, 0.999]
C: [0.126, 0.49, 0.423, 0.603]
D: [0.0, 0.0, 0.685, 0.894] | [
"./2D-spatial/single_object_tracking/single_object_tracking_187_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_187_1.jpg"
] | B |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.0, 0.046, 0.939, 0.84]
B: [0.057, 0.29, 0.25, 0.646]
C: [0.578, 0.11, 0.852, 0.163]
D: [0.0, 0.046, 0.89, 0.84] | Here is an object ([0.0, 0.001, 0.961, 0.874]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.0, 0.046, 0.939, 0.84]
B: [0.057, 0.29, 0.25, 0.646]
C: [0.578, 0.11, 0.852, 0.163]
D: [0.0, 0.046, 0.89, 0.84] | [
"./2D-spatial/single_object_tracking/single_object_tracking_188_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_188_1.jpg"
] | A |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.334, 0.31, 0.566, 0.938]
B: [0.275, 0.312, 0.504, 1.0]
C: [0.334, 0.31, 0.563, 0.997]
D: [0.591, 0.644, 0.888, 0.765] | Here is an object ([0.262, 0.143, 0.509, 0.997]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.334, 0.31, 0.566, 0.938]
B: [0.275, 0.312, 0.504, 1.0]
C: [0.334, 0.31, 0.563, 0.997]
D: [0.591, 0.644, 0.888, 0.765] | [
"./2D-spatial/single_object_tracking/single_object_tracking_189_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_189_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.0, 0.565, 0.12, 0.9]
B: [0.0, 0.565, 0.126, 0.917]
C: [0.055, 0.589, 0.181, 0.94]
D: [0.825, 0.094, 0.94, 0.535] | Here is an object ([0.0, 0.05, 1.0, 0.86]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.0, 0.565, 0.12, 0.9]
B: [0.0, 0.565, 0.126, 0.917]
C: [0.055, 0.589, 0.181, 0.94]
D: [0.825, 0.094, 0.94, 0.535] | [
"./2D-spatial/single_object_tracking/single_object_tracking_190_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_190_1.jpg"
] | B |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.644, 0.44, 0.805, 0.861]
B: [0.587, 0.544, 0.748, 0.965]
C: [0.644, 0.44, 0.811, 0.821]
D: [0.644, 0.44, 0.801, 0.908] | Here is an object ([0.572, 0.41, 0.747, 0.842]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.644, 0.44, 0.805, 0.861]
B: [0.587, 0.544, 0.748, 0.965]
C: [0.644, 0.44, 0.811, 0.821]
D: [0.644, 0.44, 0.801, 0.908] | [
"./2D-spatial/single_object_tracking/single_object_tracking_191_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_191_1.jpg"
] | A |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.6, 0.292, 0.677, 0.412]
B: [0.747, 0.479, 0.991, 1.056]
C: [0.747, 0.479, 1.0, 1.0]
D: [0.042, 0.16, 0.117, 0.547] | Here is an object ([0.755, 0.472, 1.0, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.6, 0.292, 0.677, 0.412]
B: [0.747, 0.479, 0.991, 1.056]
C: [0.747, 0.479, 1.0, 1.0]
D: [0.042, 0.16, 0.117, 0.547] | [
"./2D-spatial/single_object_tracking/single_object_tracking_192_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_192_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.268, 0.356, 0.665, 1.0]
B: [0.384, 0.329, 0.781, 0.974]
C: [0.5, 0.258, 0.897, 0.903]
D: [0.466, 0.153, 0.863, 0.797] | Here is an object ([0.386, 0.329, 0.791, 0.968]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.268, 0.356, 0.665, 1.0]
B: [0.384, 0.329, 0.781, 0.974]
C: [0.5, 0.258, 0.897, 0.903]
D: [0.466, 0.153, 0.863, 0.797] | [
"./2D-spatial/single_object_tracking/single_object_tracking_193_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_193_1.jpg"
] | B |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.243, 0.956, 0.284, 0.975]
B: [0.382, 0.301, 0.875, 0.646]
C: [0.382, 0.301, 1.019, 0.606]
D: [0.382, 0.301, 0.919, 0.646] | Here is an object ([0.411, 0.268, 0.728, 0.903]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 960 and the height is 720. | Select from the following choices.
A: [0.243, 0.956, 0.284, 0.975]
B: [0.382, 0.301, 0.875, 0.646]
C: [0.382, 0.301, 1.019, 0.606]
D: [0.382, 0.301, 0.919, 0.646] | [
"./2D-spatial/single_object_tracking/single_object_tracking_194_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_194_1.jpg"
] | D |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.388, 0.347, 0.992, 0.839]
B: [0.388, 0.347, 0.977, 0.91]
C: [0.477, 0.579, 0.912, 0.9]
D: [0.388, 0.347, 1.089, 0.938] | Here is an object ([0.386, 0.367, 0.98, 0.921]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.388, 0.347, 0.992, 0.839]
B: [0.388, 0.347, 0.977, 0.91]
C: [0.477, 0.579, 0.912, 0.9]
D: [0.388, 0.347, 1.089, 0.938] | [
"./2D-spatial/single_object_tracking/single_object_tracking_195_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_195_1.jpg"
] | B |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.228, 0.114, 0.254, 0.601]
B: [0.327, 0.138, 1.0, 1.0]
C: [0.327, 0.138, 1.021, 0.939]
D: [0.327, 0.0, 1.0, 0.863] | Here is an object ([0.332, 0.122, 1.0, 1.0]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.228, 0.114, 0.254, 0.601]
B: [0.327, 0.138, 1.0, 1.0]
C: [0.327, 0.138, 1.021, 0.939]
D: [0.327, 0.0, 1.0, 0.863] | [
"./2D-spatial/single_object_tracking/single_object_tracking_196_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_196_1.jpg"
] | B |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.426, 0.447, 0.531, 0.756]
B: [0.426, 0.447, 0.534, 0.776]
C: [0.426, 0.447, 0.53, 0.783]
D: [0.867, 0.138, 0.923, 0.214] | Here is an object ([0.431, 0.433, 0.585, 0.769]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.426, 0.447, 0.531, 0.756]
B: [0.426, 0.447, 0.534, 0.776]
C: [0.426, 0.447, 0.53, 0.783]
D: [0.867, 0.138, 0.923, 0.214] | [
"./2D-spatial/single_object_tracking/single_object_tracking_197_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_197_1.jpg"
] | C |
single_object_tracking | ['synthetic_image'] | youtubevis2019_sot | A: [0.341, 0.492, 0.753, 0.8]
B: [0.152, 0.436, 0.563, 0.744]
C: [0.168, 0.04, 0.502, 0.061]
D: [0.593, 0.619, 0.761, 0.656] | Here is an object ([0.366, 0.504, 0.786, 0.806]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.341, 0.492, 0.753, 0.8]
B: [0.152, 0.436, 0.563, 0.744]
C: [0.168, 0.04, 0.502, 0.061]
D: [0.593, 0.619, 0.761, 0.656] | [
"./2D-spatial/single_object_tracking/single_object_tracking_198_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_198_1.jpg"
] | A |
single_object_tracking | ['synthetic_image'] | ovis_sot | A: [0.049, 0.143, 0.895, 0.719]
B: [0.049, 0.143, 0.806, 0.606]
C: [0.049, 0.143, 0.788, 0.667]
D: [0.246, 0.05, 0.512, 0.212] | Here is an object ([0.056, 0.144, 0.791, 0.665]) in the Image 1. Please give the coordinations of this object in the Image 2. The bounding box coordinates are in the format [x1, y1, x2, y2], where [x1, y1] are the top-left coordinates and [x2, y2] are the bottom-right coordinates of the target object's bounding box. Note that the width of the input RGB image is 1280 and the height is 720. | Select from the following choices.
A: [0.049, 0.143, 0.895, 0.719]
B: [0.049, 0.143, 0.806, 0.606]
C: [0.049, 0.143, 0.788, 0.667]
D: [0.246, 0.05, 0.512, 0.212] | [
"./2D-spatial/single_object_tracking/single_object_tracking_199_0.jpg",
"./2D-spatial/single_object_tracking/single_object_tracking_199_1.jpg"
] | C |
Multiview_Action_Recognition | natural image | PKUMMD | A: read a book
B: drink water
C: ride a bike
D: play guitar | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: read a book
B: drink water
C: ride a bike
D: play guitar | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_0_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_0_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_0_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_0_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_0_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_0_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_0_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_0_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_0_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_0_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_0_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_0_11.png"
] | B |
Multiview_Action_Recognition | natural image | PKUMMD | A: running
B: sitting down
C: lying down
D: standing up | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: running
B: sitting down
C: lying down
D: standing up | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_1_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_1_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_1_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_1_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_1_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_1_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_1_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_1_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_1_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_1_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_1_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_1_11.png"
] | D |
Multiview_Action_Recognition | natural image | PKUMMD | A: lying down
B: standing up
C: sitting down
D: jumping | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: lying down
B: standing up
C: sitting down
D: jumping | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_2_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_2_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_2_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_2_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_2_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_2_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_2_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_2_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_2_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_2_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_2_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_2_11.png"
] | B |
Multiview_Action_Recognition | natural image | PKUMMD | A: ride bicycle
B: play guitar
C: write letter
D: eat meal | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: ride bicycle
B: play guitar
C: write letter
D: eat meal | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_3_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_3_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_3_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_3_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_3_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_3_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_3_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_3_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_3_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_3_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_3_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_3_11.png"
] | D |
Multiview_Action_Recognition | natural image | PKUMMD | A: sit
B: jump
C: pickup
D: run | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: sit
B: jump
C: pickup
D: run | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_4_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_4_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_4_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_4_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_4_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_4_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_4_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_4_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_4_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_4_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_4_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_4_11.png"
] | C |
Multiview_Action_Recognition | natural image | PKUMMD | A: make a phone call
B: play a guitar
C: ride a bicycle
D: drink a coffee | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: make a phone call
B: play a guitar
C: ride a bicycle
D: drink a coffee | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_5_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_5_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_5_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_5_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_5_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_5_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_5_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_5_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_5_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_5_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_5_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_5_11.png"
] | A |
Multiview_Action_Recognition | natural image | PKUMMD | A: read book
B: play piano
C: jog
D: eat meal | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: read book
B: play piano
C: jog
D: eat meal | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_6_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_6_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_6_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_6_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_6_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_6_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_6_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_6_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_6_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_6_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_6_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_6_11.png"
] | D |
Multiview_Action_Recognition | natural image | PKUMMD | A: pickup
B: sit
C: run
D: jump | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: pickup
B: sit
C: run
D: jump | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_7_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_7_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_7_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_7_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_7_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_7_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_7_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_7_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_7_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_7_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_7_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_7_11.png"
] | A |
Multiview_Action_Recognition | natural image | PKUMMD | A: running
B: sleeping
C: dancing
D: reading | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: running
B: sleeping
C: dancing
D: reading | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_8_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_8_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_8_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_8_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_8_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_8_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_8_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_8_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_8_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_8_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_8_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_8_11.png"
] | D |
Multiview_Action_Recognition | natural image | PKUMMD | A: ride a bicycle
B: make a phone call
C: cook a meal
D: play a piano | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: ride a bicycle
B: make a phone call
C: cook a meal
D: play a piano | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_9_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_9_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_9_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_9_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_9_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_9_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_9_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_9_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_9_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_9_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_9_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_9_11.png"
] | B |
Multiview_Action_Recognition | natural image | PKUMMD | A: read a book
B: tie shoelaces
C: check time (from watch)
D: wave hand | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: read a book
B: tie shoelaces
C: check time (from watch)
D: wave hand | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_10_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_10_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_10_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_10_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_10_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_10_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_10_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_10_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_10_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_10_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_10_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_10_11.png"
] | C |
Multiview_Action_Recognition | natural image | PKUMMD | A: raise hand (greeting)
B: touch chest (stomachache
C: tie shoelaces (preparing to run)
D: clap hands (applause) | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: raise hand (greeting)
B: touch chest (stomachache
C: tie shoelaces (preparing to run)
D: clap hands (applause) | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_11_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_11_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_11_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_11_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_11_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_11_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_11_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_11_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_11_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_11_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_11_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_11_11.png"
] | B |
Multiview_Action_Recognition | natural image | PKUMMD | A: pickup
B: sit
C: jump
D: run | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: pickup
B: sit
C: jump
D: run | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_12_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_12_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_12_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_12_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_12_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_12_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_12_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_12_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_12_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_12_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_12_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_12_11.png"
] | A |
Multiview_Action_Recognition | natural image | PKUMMD | A: reading a book
B: cooking a meal
C: writing a letter
D: brushing teeth | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: reading a book
B: cooking a meal
C: writing a letter
D: brushing teeth | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_13_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_13_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_13_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_13_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_13_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_13_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_13_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_13_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_13_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_13_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_13_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_13_11.png"
] | D |
Multiview_Action_Recognition | natural image | PKUMMD | A: sit down
B: jump
C: wave hand
D: touch chest (stomachache | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: sit down
B: jump
C: wave hand
D: touch chest (stomachache | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_14_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_14_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_14_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_14_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_14_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_14_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_14_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_14_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_14_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_14_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_14_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_14_11.png"
] | D |
Multiview_Action_Recognition | natural image | PKUMMD | A: standing up
B: jumping
C: running
D: sitting down | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: standing up
B: jumping
C: running
D: sitting down | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_15_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_15_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_15_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_15_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_15_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_15_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_15_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_15_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_15_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_15_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_15_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_15_11.png"
] | A |
Multiview_Action_Recognition | natural image | PKUMMD | A: pickup
B: run
C: sit down
D: jump | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: pickup
B: run
C: sit down
D: jump | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_16_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_16_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_16_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_16_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_16_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_16_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_16_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_16_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_16_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_16_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_16_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_16_11.png"
] | A |
Multiview_Action_Recognition | natural image | PKUMMD | A: jogging
B: brushing teeth
C: eating
D: reading a book | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: jogging
B: brushing teeth
C: eating
D: reading a book | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_17_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_17_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_17_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_17_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_17_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_17_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_17_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_17_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_17_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_17_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_17_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_17_11.png"
] | B |
Multiview_Action_Recognition | natural image | PKUMMD | A: eat sandwich
B: read book
C: ride bicycle
D: wear jacket | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: eat sandwich
B: read book
C: ride bicycle
D: wear jacket | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_18_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_18_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_18_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_18_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_18_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_18_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_18_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_18_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_18_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_18_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_18_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_18_11.png"
] | D |
Multiview_Action_Recognition | natural image | PKUMMD | A: take off a hat
B: tie shoelaces
C: put on a hat
D: put on gloves | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: take off a hat
B: tie shoelaces
C: put on a hat
D: put on gloves | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_19_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_19_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_19_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_19_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_19_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_19_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_19_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_19_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_19_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_19_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_19_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_19_11.png"
] | C |
Multiview_Action_Recognition | natural image | PKUMMD | A: ride bicycle
B: wear jacket
C: read book
D: cook dinner | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: ride bicycle
B: wear jacket
C: read book
D: cook dinner | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_20_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_20_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_20_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_20_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_20_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_20_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_20_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_20_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_20_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_20_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_20_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_20_11.png"
] | B |
Multiview_Action_Recognition | natural image | PKUMMD | A: ride a bicycle
B: tie a shoelace
C: drink water
D: read a book | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: ride a bicycle
B: tie a shoelace
C: drink water
D: read a book | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_21_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_21_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_21_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_21_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_21_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_21_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_21_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_21_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_21_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_21_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_21_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_21_11.png"
] | C |
Multiview_Action_Recognition | natural image | PKUMMD | A: sitting down
B: jumping
C: standing up
D: lying down | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: sitting down
B: jumping
C: standing up
D: lying down | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_22_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_22_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_22_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_22_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_22_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_22_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_22_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_22_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_22_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_22_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_22_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_22_11.png"
] | C |
Multiview_Action_Recognition | natural image | PKUMMD | A: drink water
B: read a book
C: tie shoes
D: climb stairs | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: drink water
B: read a book
C: tie shoes
D: climb stairs | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_23_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_23_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_23_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_23_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_23_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_23_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_23_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_23_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_23_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_23_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_23_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_23_11.png"
] | A |
Multiview_Action_Recognition | natural image | PKUMMD | A: play a guitar
B: drink water
C: ride a bike
D: write a note | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: play a guitar
B: drink water
C: ride a bike
D: write a note | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_24_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_24_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_24_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_24_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_24_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_24_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_24_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_24_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_24_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_24_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_24_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_24_11.png"
] | B |
Multiview_Action_Recognition | natural image | PKUMMD | A: take off a hat
B: put on a hat
C: pick up a book
D: tie shoelaces | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: take off a hat
B: put on a hat
C: pick up a book
D: tie shoelaces | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_25_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_25_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_25_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_25_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_25_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_25_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_25_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_25_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_25_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_25_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_25_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_25_11.png"
] | B |
Multiview_Action_Recognition | natural image | PKUMMD | A: ride bike
B: read book
C: play guitar
D: eat meal | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: ride bike
B: read book
C: play guitar
D: eat meal | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_26_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_26_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_26_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_26_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_26_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_26_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_26_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_26_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_26_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_26_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_26_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_26_11.png"
] | D |
Multiview_Action_Recognition | natural image | PKUMMD | A: touch chest (stomachache
B: throw a ball
C: jump up
D: tie shoelaces | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: touch chest (stomachache
B: throw a ball
C: jump up
D: tie shoelaces | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_27_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_27_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_27_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_27_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_27_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_27_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_27_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_27_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_27_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_27_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_27_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_27_11.png"
] | A |
Multiview_Action_Recognition | natural image | PKUMMD | A: wave
B: sit down
C: jump
D: pickup | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: wave
B: sit down
C: jump
D: pickup | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_28_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_28_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_28_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_28_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_28_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_28_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_28_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_28_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_28_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_28_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_28_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_28_11.png"
] | D |
Multiview_Action_Recognition | natural image | PKUMMD | A: eat sandwich
B: ride bicycle
C: wear jacket
D: play guitar | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: eat sandwich
B: ride bicycle
C: wear jacket
D: play guitar | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_29_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_29_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_29_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_29_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_29_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_29_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_29_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_29_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_29_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_29_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_29_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_29_11.png"
] | C |
Multiview_Action_Recognition | natural image | PKUMMD | A: dancing
B: reading
C: sleeping
D: cooking | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: dancing
B: reading
C: sleeping
D: cooking | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_30_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_30_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_30_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_30_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_30_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_30_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_30_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_30_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_30_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_30_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_30_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_30_11.png"
] | B |
Multiview_Action_Recognition | natural image | PKUMMD | A: brushing teeth
B: washing face
C: brushing hair
D: combing hair | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: brushing teeth
B: washing face
C: brushing hair
D: combing hair | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_31_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_31_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_31_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_31_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_31_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_31_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_31_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_31_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_31_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_31_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_31_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_31_11.png"
] | A |
Multiview_Action_Recognition | natural image | PKUMMD | A: tie shoelaces
B: check time (from watch)
C: drink water
D: read a book | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: tie shoelaces
B: check time (from watch)
C: drink water
D: read a book | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_32_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_32_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_32_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_32_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_32_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_32_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_32_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_32_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_32_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_32_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_32_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_32_11.png"
] | B |
Multiview_Action_Recognition | natural image | PKUMMD | A: jump up
B: touch chest (stomachache
C: wave hand
D: sit down | Given the set of images from three different views (i.e., left, middle and right views), please identify the action that this person performs. | Your task is recognize human actions or activities in a scene using information from multiple views.
Select from the following choices.
A: jump up
B: touch chest (stomachache
C: wave hand
D: sit down | [
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_33_0.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_33_1.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_33_2.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_33_3.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_33_4.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_33_5.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_33_6.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_33_7.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_33_8.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_33_9.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_33_10.png",
"./3D-spatial/Multiview_Action_Recognition/Multiview_Action_Recognition_33_11.png"
] | B |