Use Cases

RoboX data supports a range of robotics and AI applications.


The Challenge

Autonomous systems need to navigate real environments, not just drive down empty roads, but handle the full complexity of real-world spaces: crowds, obstacles, changing conditions, ambiguous paths.

Traditional navigation data comes from vehicle-mounted sensors (useful for driving) or building floor plans (static, quickly outdated). Neither captures how humans actually navigate on foot through complex environments.

How RoboX Helps

Egocentric data shows navigation from the human perspective:

  • Path selection: Which route do people choose when multiple options exist?

  • Obstacle handling: How do people navigate around obstacles, other pedestrians, unexpected barriers?

  • Environmental adaptation: How does navigation change in rain, darkness, crowds?

  • Attention patterns: What do people look at when navigating? What signals do they use?

Applications

Application
Data Used
Outcome

Indoor robot navigation

Wi-Fi/BLE RSSI, IMU, visual

Robots that navigate malls, airports, hospitals without GPS

Last-mile delivery

Visual, GPS, IMU

Delivery robots that handle sidewalks, crosswalks, building entrances

Warehouse automation

Depth, IMU, visual

Robots that navigate dynamic warehouse environments

Assistive navigation

Visual, audio, position

Systems that help visually impaired users navigate


Perception

The Challenge

Robots need to perceive and understand their environment: identify objects, assess surfaces, recognize situations, predict what will happen next.

Most perception training data comes from static images or video taken from fixed positions. This doesn't match the dynamic, first-person view a robot experiences during operation.

How RoboX Helps

Egocentric perception data captures:

  • Objects as they appear during approach (changing scale, angle, occlusion)

  • Surface characteristics from walking/driving perspective

  • Dynamic scene changes (people moving, doors opening, lighting shifts)

  • Contextual relationships (what objects appear together, how spaces are organized)

Applications

Application
Data Used
Outcome

Road surface assessment

Camera, IMU

Detect potholes, cracks, hazards for vehicles and pedestrians

Obstacle detection

Camera, depth

Identify and classify obstacles in robot's path

Scene understanding

Camera, audio

Recognize environment type and appropriate behavior

Object recognition

Camera, depth

Identify objects in context of normal use


Drone Operations

The Challenge

Drones operate in 3D space with unique constraints: altitude management, landing zone identification, noise considerations, airspace awareness.

Most drone training data comes from aerial footage, which doesn't help with ground-level decisions like landing zone assessment or understanding human environments the drone must interact with.

How RoboX Helps

Ground-level data from humans provides:

  • Landing zone characteristics (surface type, obstacles, space availability)

  • Acoustic environment mapping (noise-sensitive areas to avoid)

  • Altitude context (building heights, floor levels, terrain variation)

  • Human activity patterns (where people congregate, movement flows)

Applications

Application
Data Used
Outcome

Acoustic routing

Microphone

Route planning that minimizes noise disturbance

Landing zone selection

Camera, depth, barometer

Identify safe, appropriate landing locations

Altitude management

Barometer, GPS

Accurate floor-level detection in urban environments

Delivery planning

Position, visual

Understand delivery destinations from ground level


Humanoid Robotics

The Challenge

Humanoid robots aim to operate in human environments and perform human-like tasks. This requires understanding not just what humans do, but how they do it: the subtle movement patterns, balance adjustments, and behavioral rhythms of human activity.

Lab demonstrations provide some training data, but lab behavior differs from natural behavior. Teleoperation provides robot-perspective data but with unnatural control inputs.

How RoboX Helps

Egocentric data from natural human activity captures:

  • Movement patterns: How people walk, turn, stop, start, navigate obstacles

  • Manipulation context: How objects are approached, grasped, used

  • Behavioral rhythms: Pacing, pauses, attention shifts during activities

  • Environmental interaction: How people use doors, elevators, furniture, tools

Applications

Application
Data Used
Outcome

Locomotion training

IMU, visual

Natural walking, turning, stair navigation

Imitation learning

Visual, IMU, audio

Robots that learn tasks from human demonstration

Social navigation

Visual, position

Appropriate behavior around people

Activity execution

Visual, IMU

Task completion with natural movement patterns


Spatial Computing

The Challenge

AR/VR systems need to understand 3D space, track user position, and blend digital content with physical environments. This requires detailed spatial understanding that works across diverse real-world settings.

How RoboX Helps

Egocentric spatial data provides:

  • Room-scale 3D geometry from diverse environments

  • Real-world lighting conditions and variations

  • Surface characteristics and materials

  • Spatial relationships and typical room layouts

Applications

Application
Data Used
Outcome

AR anchoring

Depth, visual

Stable placement of virtual objects in real scenes

Inside-out tracking

Visual, IMU

Position tracking without external sensors

Scene reconstruction

Depth, visual

3D models of real environments

Occlusion handling

Depth, visual

Virtual objects correctly hidden by real objects


Autonomous Vehicles

The Challenge

Self-driving vehicles primarily use vehicle-mounted sensors, but they must interact with pedestrians, cyclists, and other non-vehicle road users. Understanding human behavior from the human perspective improves prediction and safety.

How RoboX Helps

Pedestrian-perspective data shows:

  • How pedestrians approach crossings and intersections

  • Cyclist behavior and navigation patterns

  • Road surface conditions as experienced by non-vehicle users

  • Interaction patterns between pedestrians and traffic

Applications

Application
Data Used
Outcome

Pedestrian prediction

Visual, position, IMU

Anticipate pedestrian movement and intent

Road condition assessment

Camera, IMU

Understand surface conditions affecting all road users

Intersection behavior

Visual, position

Model how people navigate complex intersections

Vulnerable road user safety

Visual, audio

Better detection and response to pedestrians/cyclists


Research & Academia

The Challenge

Academic research requires large, diverse, well-documented datasets. Collecting such data independently is expensive and time-consuming, often limiting research scope.

How RoboX Helps

RoboX provides:

  • Scale: Data from thousands of collectors across dozens of countries

  • Diversity: Varied environments, conditions, and behaviors

  • Documentation: Clear methodology, sensor specifications, known limitations

  • Accessibility: API access and downloadable datasets

  • Freshness: Continuously updated, reflecting current conditions

Research Areas Supported

  • Computer vision and perception

  • Robot learning and imitation

  • Navigation and SLAM

  • Human activity recognition

  • Spatial computing and 3D reconstruction

  • Sensor fusion and multi-modal learning

Last updated