Data Format

Dataset Structure

Every RoboX dataset is organized by campaign. Each campaign dataset contains clips, annotations, and metadata.

robox-egograsp/
├── manifest.json               # Dataset version, schema version, stats
├── metadata.parquet             # Per-clip metadata (environment, duration, quality score)
├── clips/
│   ├── clip_001.mp4            # Egocentric video
│   ├── clip_002.mp4
│   └── ...
├── annotations/
│   ├── clip_001.json           # All annotation layers for this clip
│   ├── clip_002.json
│   └── ...
└── README.md                   # Dataset card with license, citation, changelog

Video Format

Property
Value

Container

MP4 (H.264)

Resolution

1080p–4K (varies by device)

Frame rate

30 fps (standard), 60 fps (where supported)

Color space

sRGB

Audio

Stripped during anonymization

Typical duration

10–120 seconds per clip

All videos are anonymized before inclusion. Faces, text, license plates, and device identifiers have been removed or blurred on-device.

Annotation Schema

Each clip's annotation file contains multiple layers. Not every campaign includes all layers — see the layer availability table below.

Per-Clip Annotation Structure

Annotation Layer Details

Temporal Segmentation

Action start/end boundaries with labeled phases. Each segment includes a label from the campaign's action vocabulary.

Object Bounding Boxes

Per-frame object detection with bounding box coordinates and category labels.

Bounding box format: [x_min, y_min, x_max, y_max] in pixel coordinates.

Hand Pose

21-keypoint hand skeleton per frame, following the MediaPipe Hand Landmarks topology.

Keypoint coordinates are normalized to [0, 1] relative to frame dimensions.

Gaze Direction

Estimated gaze vector per frame.

Coordinates represent the estimated fixation point as normalized [0, 1] within the frame.

Spatial Layout

Scene structure and surface annotations per clip (not per-frame).

Interaction Classification

Action-object pair labels per temporal segment.

Layer Availability by Campaign

Layer
EgoGrasp
EgoScene
EgoNav
EgoDaily
EgoSocial

Temporal segmentation

Yes

Yes

Yes

TBD

Object bounding

Yes

Yes

Yes

TBD

Hand pose

Yes

Yes

TBD

Gaze direction

Yes

Yes

Yes

TBD

Spatial layout

Yes

Yes

TBD

Interaction classification

Yes

Yes

TBD

Metadata Schema (Parquet)

The metadata.parquet file contains one row per clip with the following columns:

Column
Type
Description

clip_id

string

Unique clip identifier

campaign

string

Campaign name

duration_sec

float

Clip duration in seconds

fps

int

Frame rate

resolution

string

Video resolution (e.g. "1920x1080")

environment

string

Environment category (kitchen, office, warehouse, etc.)

device_model

string

Anonymized device model identifier

quality_score

float

Overall quality score (0–1)

annotation_layers

list[string]

Available annotation layers for this clip

object_categories

list[string]

Object categories detected in the clip

schema_version

string

Annotation schema version

created_at

timestamp

Clip creation date

Versioning

Datasets are versioned by release date using the format YYYY.MM.patch (e.g., 2026.03.1). The manifest.json file in each dataset root tracks the current version and schema version.

When annotation pipelines are updated, new layers may be applied retroactively to existing clips. The schema version in each annotation file indicates which pipeline produced it. See the Changelog for release history.

Last updated