Dataset Viewer
Auto-converted to Parquet Duplicate
id
string
total_frames
int32
annotations
string
trajectory_image
image
video_path
string
00005613
144
{"0": [{"type": "point", "x": 362, "y": 123, "x1": 0, "y1": 0, "x2": 0, "y2": 0}], "1": [{"type": "point", "x": 365, "y": 118, "x1": 0, "y1": 0, "x2": 0, "y2": 0}], "2": [{"type": "point", "x": 362, "y": 116, "x1": 0, "y1": 0, "x2": 0, "y2": 0}], "3": [{"type": "point", "x": 362, "y": 114, "x1": 0, "y1": 0, "x2": 0, "y...
RoboManip-Traj-Demo/00005613/00005613.mp4
00007356
240
{"0": [{"type": "point", "x": 434, "y": 268, "x1": 0, "y1": 0, "x2": 0, "y2": 0}], "1": [{"type": "point", "x": 437, "y": 268, "x1": 0, "y1": 0, "x2": 0, "y2": 0}], "2": [{"type": "point", "x": 438, "y": 267, "x1": 0, "y1": 0, "x2": 0, "y2": 0}], "3": [{"type": "point", "x": 439, "y": 272, "x1": 0, "y1": 0, "x2": 0, "y...
RoboManip-Traj-Demo/00007356/00007356.mp4
00011929
111
{"111": [{"type": "point", "x": 647, "y": 388, "x1": 0, "y1": 0, "x2": 0, "y2": 0}], "112": [{"type": "point", "x": 642, "y": 387, "x1": 0, "y1": 0, "x2": 0, "y2": 0}], "113": [{"type": "point", "x": 640, "y": 391, "x1": 0, "y1": 0, "x2": 0, "y2": 0}], "114": [{"type": "point", "x": 638, "y": 392, "x1": 0, "y1": 0, "x2...
RoboManip-Traj-Demo/00011929/00011929.mp4
00021065
141
{"141": [{"type": "point", "x": 631, "y": 381, "x1": 0, "y1": 0, "x2": 0, "y2": 0}], "142": [{"type": "point", "x": 630, "y": 384, "x1": 0, "y1": 0, "x2": 0, "y2": 0}], "143": [{"type": "point", "x": 633, "y": 385, "x1": 0, "y1": 0, "x2": 0, "y2": 0}], "144": [{"type": "point", "x": 631, "y": 386, "x1": 0, "y1": 0, "x2...
RoboManip-Traj-Demo/00021065/00021065.mp4
00025458
495
{"0": [{"type": "point", "x": 584, "y": 483, "x1": 0, "y1": 0, "x2": 0, "y2": 0}], "1": [{"type": "point", "x": 583, "y": 487, "x1": 0, "y1": 0, "x2": 0, "y2": 0}], "2": [{"type": "point", "x": 585, "y": 481, "x1": 0, "y1": 0, "x2": 0, "y2": 0}], "3": [{"type": "point", "x": 590, "y": 482, "x1": 0, "y1": 0, "x2": 0, "y...
RoboManip-Traj-Demo/00025458/00025458.mp4
00030962
110
{"0": [{"type": "point", "x": 736, "y": 213, "x1": 0, "y1": 0, "x2": 0, "y2": 0}], "2": [{"type": "point", "x": 734, "y": 212, "x1": 0, "y1": 0, "x2": 0, "y2": 0}], "4": [{"type": "point", "x": 736, "y": 212, "x1": 0, "y1": 0, "x2": 0, "y2": 0}], "6": [{"type": "point", "x": 736, "y": 213, "x1": 0, "y1": 0, "x2": 0, "y...
RoboManip-Traj-Demo/00030962/00030962.mp4
00034975
141
{"103": [{"type": "point", "x": 600, "y": 502}], "1": [{"type": "point", "x": 836, "y": 468}, {"type": "point", "x": 894, "y": 456}, {"type": "point", "x": 964, "y": 452}, {"type": "point", "x": 1019, "y": 450}, {"type": "point", "x": 1099, "y": 450}, {"type": "point", "x": 1156, "y": 448}], "14": [{"type": "point", "x...
RoboManip-Traj-Demo/00034975/00034975.mp4
00035684
148
{"20": [{"type": "rect", "x1": 647, "y1": 547, "x2": 726, "y2": 610}, {"type": "point", "x": 802, "y": 519}, {"type": "point", "x": 790, "y": 529}, {"type": "point", "x": 785, "y": 531}, {"type": "point", "x": 781, "y": 534}, {"type": "point", "x": 774, "y": 536}, {"type": "point", "x": 766, "y": 540}, {"type": "point"...
RoboManip-Traj-Demo/00035684/00035684.mp4
00037182
330
{"16": [{"type": "point", "x": 518, "y": 200}, {"type": "point", "x": 519, "y": 213}, {"type": "point", "x": 519, "y": 225}, {"type": "point", "x": 519, "y": 238}, {"type": "point", "x": 518, "y": 244}, {"type": "point", "x": 518, "y": 264}, {"type": "point", "x": 518, "y": 264}, {"type": "point", "x": 517, "y": 296}, ...
RoboManip-Traj-Demo/00037182/00037182.mp4
00042098
145
"{\"3\": [{\"type\": \"point\", \"x\": 965, \"y\": 187}, {\"type\": \"point\", \"x\": 902, \"y\": 21(...TRUNCATED)
RoboManip-Traj-Demo/00042098/00042098.mp4
End of preview. Expand in Data Studio

Codatta Robotic Manipulation Trajectory (Sample)

Overview

This dataset contains high-quality annotated trajectories of robotic gripper manipulations. Produced by Codatta, it focuses on third-person views of robotic arms performing pick-and-place or manipulation tasks. The dataset is designed to train models for fine-grained control, trajectory prediction, and object interaction tasks.

The scope specifically includes third-person views (fixed camera recording the robot) while explicitly excluding first-person views (Eye-in-Hand) to ensure consistent coordinate mapping.

Dataset Contents

Each sample in this dataset includes the raw video, a visualization of the trajectory, and a rigorous JSON annotation of keyframes and coordinate points.

Data Fields

  • id (string): Unique identifier for the trajectory sequence.
  • total_frames (int32): Total number of frames in the video sequence.
  • video_path (string): Path to the source MP4 video file recording the manipulation action.
  • trajectory_image (image): A JPEG preview showing the overlaid trajectory path or keyframe visualization.
  • annotations (string): A JSON-formatted string containing the detailed coordinate data. It contains lists of keyframes, timestamps, and 5-point coordinates for the gripper.

Annotation Standards

The data follows a strict protocol to ensure precision:

1. Keyframe Selection Annotations are sparse, focusing on specific Keyframes defined by the following events:

  • Start Frame: The gripper first appears in the screen.
  • End Frame: The gripper leaves the screen.
  • Velocity Change: Frames where the speed direction suddenly changes (marking the minimum speed point).
  • State Change: Frames where the gripper opens or closes.
  • Contact: The precise moment the gripper touches the object.

2. The 5-Point Annotation Method For every annotated keyframe, the gripper is labeled with 5 specific coordinate points to capture its pose and state accurately:

Point ID Description Location Detail
Point 1 & 2 Fingertips Center of the bottom edge of the gripper tips.
Point 3 & 4 Gripper Ends The rearmost points of the closing area (indicating the finger direction).
Point 5 Tiger's Mouth The center of the crossbeam (base of the gripper).

3. Quality Control

  • Accuracy: All datasets passed a rigorous quality assurance process with a minimum 95% accuracy rate.
  • Occlusion Handling: Sequences where the gripper is fully occluded or only shows a side profile without clear features are discarded.

Key Statistics

  • Total Examples: 50 annotated examples (Sample Dataset).
  • Language: English (en).
  • Splits: Train split available.
  • Download Size: ~38.7 MB.
  • Dataset Size: ~39.0 MB.

Usage

This dataset is suitable for research and development in the field of Embodied AI and Computer Vision. It is specifically curated to support the following downstream tasks and application scenarios:

  • Trajectory Prediction: The high-precision coordinate data allows for training models to predict the future path of a gripper based on initial visual contexts.
  • Keyframe Extraction & Event Detection: By leveraging the labeled event types (e.g., "Contact", "Velocity Change"), models can be trained to automatically identify critical moments in long-horizon manipulation tasks.
  • Fine-Grained Robotic Control: The 5-point annotation system provides detailed pose information, enabling Imitation Learning (IL) from human-demonstrated or teleoperated data for precise pick-and-place operations.
  • Object Interaction Analysis: The dataset helps in understanding gripper-object relationships, specifically modeling the transition states when the gripper opens, closes, or makes contact with an object.

Usage Example

from datasets import load_dataset
import json

# Load the dataset
ds = load_dataset("Codatta/robotic-manipulation-trajectory", split="train")

# Access a sample
sample = ds[0]

# View the image
print(f"Trajectory ID: {sample['id']}")
sample['trajectory_image'].show()

# Parse annotations
annotations = json.loads(sample['annotations'])
print(f"Keyframes count: {len(annotations)}")

License and Open-Source Details

  • License: This dataset is released under the OpenRAIL license.
Downloads last month
16