- 103,354, 20s 10Hz segments (over 20 million frames), mined for interesting interactions
- 574 hours of data
- Sensor data
- 4 short-range lidars
- 1 mid-range lidar
- Object data
- 10.8M objects with tracking IDs
- Labels for 3 object classes - Vehicles, Pedestrians, Cyclists
- 3D bounding boxes for each object
- Mined for interesting behaviors and scenarios for behavior prediction research, such as unprotected turns, merges, lane changes, and intersections
- 3D bounding boxes are generated by a model trained on the Perception Dataset and detailed in our paper: Offboard 3D Object Detection from Point Cloud Sequences
Map data
- 3D map data for each segment
- Locations include: San Francisco, Phoenix, Mountain View, Los Angeles, Detroit, and Seattle
- Added entrances to driveways (the map already Includes lane centers, lane boundaries, road boundaries, crosswalks, speed bumps and stop signs)
- Adjusted some road edge boundary height estimates