Dataset Viewer
Auto-converted to Parquet
images
images list
data_source
string
reward_model
dict
extra_info
dict
prompt
list
ability
string
env_name
string
bbox_format
string
source_dataset
string
question
string
image
string
answer
string
target_instances
list
conversations
list
search
int64
sealvqa_spatial
{ "ground_truth": "No, the down more head is not on the right side of the horse with 2 straps on face." }
{ "answer": "No, the down more head is not on the right side of the horse with 2 straps on face.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes are in absolute pixel coordinates with format [x, y, width, height], where (x, y) is the top-left corner, width is horizontal extent, and height is vertical extent.", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2014/train2014/COCO_train2014_000000127622.jpg", "index": "sealvqa_spatial_41905", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the down more head on the right side of the horse with 2 straps on face?" }, { "from": "gpt", "value": "No, the down more head is not on the right side of the horse with 2 straps on face." } ], "question": "the object; the object.\nIs the down more head on the right side of the horse with 2 straps on face?", "question_id": "41905", "search": 1, "target_instances": [ { "bbox": [ 227.65, 209.3, 144.8, 345.37 ], "instance_id": "", "name": "horse with 2 straps on face" }, { "bbox": [ 67.15, 234.54, 127.45, 357.69 ], "instance_id": "", "name": "down more head" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a specific region of an image by cropping it based on a bounding box (bbox) and an optional object label.\",\"parameters\":{\"type\":\"object\",\"properties\":{\"bbox_2d\":{\"type\":\"array\",\"items\":{\"type\":\"number\"},\"minItems\":4,\"maxItems\":4,\"description\":\"The bounding box of the region to zoom in, as [x1, y1, x2, y2], where (x1, y1) is the top-left corner and (x2, y2) is the bottom-right corner.\"},\"label\":{\"type\":\"string\",\"description\":\"The name or label of the object in the specified bounding box (optional).\"}},\"required\":[\"bbox\"]}}}\n</tools>\n\n# How to call a tool\nReturn a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call>\n\n**Example**: \n<tool_call> \n{\"name\": \"image_zoom_in_tool\", \"arguments\": {\"bbox_2d\": [10, 20, 100, 200], \"label\": \"the apple on the desk\"}} \n</tool_call>", "role": "system" }, { "content": "<image>\nthe object; the object.\nIs the down more head on the right side of the horse with 2 straps on face?\nThink first, call **image_zoom_in_tool** if needed, then answer. Format strictly as: <think>...</think> <tool_call>...</tool_call> (if tools needed) <answer>...</answer> ", "role": "user" } ]
sealvqa_spatial_spatial
visual_toolbox_v2
absolute_xywh
sealvqa_spatial
null
null
null
null
null
null
sealvqa_gqa
{ "ground_truth": "The ball is to the right of the man that is wearing socks." }
{ "answer": "The ball is to the right of the man that is wearing socks.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes are in absolute pixel coordinates with format [x, y, width, height], where (x, y) is the top-left corner, width is horizontal extent, and height is vertical extent.", "conversation_turns": 2, "data_split": "train", "dataset_type": "gqa", "has_targets": true, "image_path": "gqa/images/2361897.jpg", "index": "sealvqa_gqa_41905", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the ball to the left or to the right of the man that is wearing socks?" }, { "from": "gpt", "value": "The ball is to the right of the man that is wearing socks." } ], "question": "Is the ball to the left or to the right of the man that is wearing socks?", "question_id": "41905", "search": 1, "target_instances": [ { "bbox": [ 299, 166, 27, 26 ], "instance_id": "2224233", "name": "ball" }, { "bbox": [ 122, 188, 11, 17 ], "instance_id": "1793964", "name": "socks" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a specific region of an image by cropping it based on a bounding box (bbox) and an optional object label.\",\"parameters\":{\"type\":\"object\",\"properties\":{\"bbox_2d\":{\"type\":\"array\",\"items\":{\"type\":\"number\"},\"minItems\":4,\"maxItems\":4,\"description\":\"The bounding box of the region to zoom in, as [x1, y1, x2, y2], where (x1, y1) is the top-left corner and (x2, y2) is the bottom-right corner.\"},\"label\":{\"type\":\"string\",\"description\":\"The name or label of the object in the specified bounding box (optional).\"}},\"required\":[\"bbox\"]}}}\n</tools>\n\n# How to call a tool\nReturn a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call>\n\n**Example**: \n<tool_call> \n{\"name\": \"image_zoom_in_tool\", \"arguments\": {\"bbox_2d\": [10, 20, 100, 200], \"label\": \"the apple on the desk\"}} \n</tool_call>", "role": "system" }, { "content": "<image>\nIs the ball to the left or to the right of the man that is wearing socks?\nThink first, call **image_zoom_in_tool** if needed, then answer. Format strictly as: <think>...</think> <tool_call>...</tool_call> (if tools needed) <answer>...</answer> ", "role": "user" } ]
sealvqa_gqa_spatial
visual_toolbox_v2
absolute_xywh
sealvqa_gqa
Is the ball to the left or to the right of the man that is wearing socks?
gqa/images/2361897.jpg
The ball is to the right of the man that is wearing socks.
[ { "bbox": [ 299, 166, 27, 26 ], "instance_id": "2224233", "name": "ball" }, { "bbox": [ 122, 188, 11, 17 ], "instance_id": "1793964", "name": "socks" } ]
[ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the ball to the left or to the right of the man that is wearing socks?" }, { "from": "gpt", "value": "The ball is to the right of the man that is wearing socks." } ]
1
README.md exists but content is empty.
Downloads last month
92