The dataset viewer is not available for this split.
Error code: RowsPostProcessingError
Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
Dataset Card for EgoCart
EgoCart is a large-scale benchmark dataset for egocentric, image-based indoor localization in a retail store. RGB images and depth maps were captured by cameras mounted on shopping carts moving through a real supermarket. Each frame is annotated with a 3-DOF camera pose (position + orientation) and a store-zone class label.
This card covers the egocart_videos FiftyOne dataset: 9 MP4 videos — one per recording sequence — with frame-level pose and zone annotations.
This is a FiftyOne dataset with 9 samples.
Installation
If you haven't already, install FiftyOne:
pip install -U fiftyone
Usage
import fiftyone as fo
from fiftyone.utils.huggingface import load_from_hub
# Load the dataset
# Note: other available arguments include 'max_samples', etc
dataset = load_from_hub("Voxel51/egocart_videos")
# Launch the App
session = fo.launch_app(dataset)
Navigating the dataset
import fiftyone as fo
dataset = fo.load_dataset("egocart_videos")
# Split views
train = dataset.match_tags("train") # sequences 0, 1, 2, 4, 6, 8 (6 videos)
test = dataset.match_tags("test") # sequences 3, 5, 7 (3 videos)
# Access a specific sequence
seq4 = dataset.match(fo.ViewField("sequence_id") == "4").first()
# Frame-level access — FiftyOne frames are 1-indexed
first_frame = seq4.frames[1]
middle_frame = seq4.frames[len(seq4.frames) // 2]
print(first_frame.location_x, first_frame.location_y)
print(first_frame.zone_id)
print(first_frame.heading_deg)
# Iterate frames in a video
for frame_number, frame in seq4.frames.items():
print(frame_number, frame.zone_id, frame.location_x, frame.location_y)
# Convert frame-level fields to a flat DataFrame
frame_df = dataset.to_frames().to_pandas()
Filtering on frame-level fields
FiftyOne lets you filter videos by properties of their frames using match_frames:
# Videos that contain at least one frame in zone 15
has_zone_15 = dataset.match_frames(fo.ViewField("zone_id") == 15)
# Videos where the cart was in the right half of the store
has_right_half = dataset.match_frames(fo.ViewField("location_x") > 0)
# Videos where the cart was facing roughly East (heading ≈ 0°)
has_east = dataset.match_frames(
(fo.ViewField("heading_deg") > -30) & (fo.ViewField("heading_deg") < 30)
)
# Build a frame-level view to analyse pose across all sequences
frames_view = dataset.to_frames()
print(frames_view.count("frames")) # 19,531 total frames
Dataset Details
Dataset Description
EgoCart was collected to study the problem of localising shopping carts in a large retail store from egocentric images. It supports research into indoor localisation (image retrieval, pose regression), egocentric video understanding, and analysis of customer movement patterns.
RGB-D cameras (1280 × 720) mounted on shopping carts captured footage at 50 fps during real shopping sessions in a single Italian retail store. Nine independent recording sequences were made; six form the training split and three the test split. The store floor plan spans roughly 40 m × 17 m.
- Original dataset creators: E. Spera, A. Furnari, S. Battiato, G. M. Farinella — University of Catania, Italy
- FiftyOne curation: Harpreet Sahota
- License: Research use only (see original dataset page)
Dataset Sources
- Original dataset page: http://iplab.dmi.unict.it/EgocentricShoppingCartLocalization/
- Paper: Spera et al., EgoCart: A Benchmark Dataset for Large-Scale Indoor Image-Based Localization in Retail Stores, IEEE TCSVT 2021. https://ieeexplore.ieee.org/document/8835071
Uses
Direct Use
- Indoor localisation / place recognition: Use frame-level pose annotations as ground truth for image retrieval and nearest-neighbour localisation methods.
- Pose regression from video: Train networks to predict
(location_x, location_y, heading_deg)from RGB frames, exploiting temporal context across the video. - Store zone classification: Predict
zone_id(1–16) from each video frame as a discrete localization proxy. - Trajectory analysis: Examine how shopping carts move through the store over time; the video format makes it easy to study motion continuity, revisitation patterns, and zone transitions.
Out-of-Scope Use
- The dataset captures a single retail store at a single point in time. Models trained here will not generalise to different stores or layouts without re-collection.
- The camera faces forward along aisles; there are no identifiable people in the footage, making it unsuitable for pedestrian detection or customer re-identification.
- Videos encode only camera heading (yaw). Camera pitch and roll are not captured, so the dataset is unsuitable for full 6-DOF pose estimation.
Dataset Structure
FiftyOne video dataset overview
egocart_videos is a FiftyOne video dataset (media_type = "video") containing 9 samples. In FiftyOne's data model, each sample corresponds to one video file; annotations that change over time are stored as frame-level fields accessible via sample.frames[frame_number] (1-indexed).
egocart_videos
│
├── Sample 0 filepath = egocart_seq_0.mp4 tags = ["train"] sequence_id = "0"
│ ├── Frame 1 location_x=-17.90 location_y=4.77 heading_deg=-2.49 zone_id=16
│ ├── Frame 2 location_x=-17.93 location_y=4.76 heading_deg=-4.57 zone_id=16
│ └── ... (1,838 frames total)
│
├── Sample 1 filepath = egocart_seq_1.mp4 tags = ["train"] sequence_id = "1"
│ └── ... (1,740 frames)
│
...
│
└── Sample 8 filepath = egocart_seq_8.mp4 tags = ["train"] sequence_id = "8"
└── ... (3,101 frames)
Sample-level fields
| Field | Type | Description |
|---|---|---|
filepath |
StringField |
Absolute path to the MP4 file |
tags |
ListField(StringField) |
["train"] or ["test"] |
sequence_id |
StringField |
Recording run identifier ("0"–"8") |
metadata |
VideoMetadata |
Auto-populated: duration, fps, resolution, frame count |
Frame-level fields
Each frame corresponds to one original RGB image from the dataset. Frames are 1-indexed in FiftyOne.
| Field | Type | Description |
|---|---|---|
frame_number |
FrameNumberField |
1-indexed position within the video |
location_x |
FloatField |
Cart X position in the store (metres, ≈ −20 to +20) |
location_y |
FloatField |
Cart Y position in the store (metres, ≈ −9 to +8) |
orientation_u |
FloatField |
Camera heading unit-vector, X component |
orientation_v |
FloatField |
Camera heading unit-vector, Y component |
heading_deg |
FloatField |
atan2(v, u) in degrees — see coordinate system below |
zone_id |
IntField |
Store location class (1–16) |
Train / test split
| Split | Sequences | Frames | Duration at 15 fps |
|---|---|---|---|
| Train | 0, 1, 2, 4, 6, 8 | 13,360 | ≈ 15 min total |
| Test | 3, 5, 7 | 6,171 | ≈ 7 min total |
Sequences were recorded independently on different occasions. The split is at the sequence level, ensuring no temporal overlap between train and test.
Video encoding
Videos are encoded with H.264 / yuv420p at 15 fps (original capture rate: 50 fps, so playback is ≈ 3× slower than real time). This makes the cart's motion easy to inspect frame-by-frame in the FiftyOne App. Frame-level annotations are aligned to the 15 fps video — frame n in the video corresponds to row n in the sorted annotation file for that sequence.
Store coordinate system
The coordinate origin is a fixed reference point inside the store. Axes are in metres:
| Axis | Range | Direction |
|---|---|---|
| X | −20.2 to +19.8 m | Left → Right |
| Y | −9.5 to +8.0 m | Bottom → Top |
(orientation_u, orientation_v) is a 2D unit vector in the store's XY floor plane (u² + v² ≈ 1.0). The derived field heading_deg = atan2(v, u) gives a compass-style angle: 0° = East (+X), 90° = North (+Y), ±180° = West (−X), −90° = South (−Y).
The store aisles run primarily North–South, so most frames show heading values near ±90°. The cart reverses direction between adjacent aisle lanes (serpentine route), which is clearly visible as alternating heading clusters in each sequence.
Store zone distribution
Zones 1–9 are sequential single-aisle sections along the central spine from left to right. Zones 10–16 are larger structural zones (perimeter walkways, wide sections).
| Zone | Description | Train frames | Test frames |
|---|---|---|---|
| 1 | Left aisle zone 1 | 311 | 173 |
| 2 | Left aisle zone 2 | 244 | 131 |
| 3 | Left aisle zone 3 | 241 | 177 |
| 4 | Left aisle zone 4 | 278 | 117 |
| 5 | Left aisle zone 5 | 665 | 174 |
| 6 | Center aisle zone 6 | 403 | 217 |
| 7 | Center aisle zone 7 | 394 | 201 |
| 8 | Center aisle zone 8 | 266 | 190 |
| 9 | Right aisle zone 9 | 366 | 235 |
| 10 | Right half aisles | 1,722 | 751 |
| 11 | Back wall aisle (full width) | 1,610 | 658 |
| 12 | Right wall | 1,355 | 409 |
| 13 | Bottom right | 897 | 456 |
| 14 | Bottom left | 1,138 | 788 |
| 15 | Central bottom aisle (full width) | 2,668 | 1,130 |
| 16 | Left wall | 802 | 364 |
Zone distribution is imbalanced: zone 15 is the most frequent because the cart must traverse the full-width bottom aisle when moving between the two halves of the store. Zones 1–4 are the rarest.
Dataset Creation
Curation Rationale
Indoor localisation from images is a practically important but data-scarce problem. At the time of publication, existing datasets were too small, collected in corridor environments, or lacked ground-truth poses. EgoCart provides a large, real-world retail benchmark with accurate 3-DOF pose annotations to enable reproducible benchmarking of image retrieval and pose regression methods.
The video format presented here groups the original frame sequences into their natural temporal structure, making it straightforward to study motion context, temporal consistency of predictions, and zone transition behaviour.
Source Data
Data Collection and Processing
RGB-D cameras were mounted on shopping carts and driven through a real retail store in Italy. Cameras captured at 50 fps — confirmed by the 20 ms (20-unit) increments in the numeric portion of each source filename. Ground-truth 3-DOF poses (x, y, u, v) were obtained through a separate localisation system. Zone labels (1–16) were assigned based on which manually defined store region contains each frame's (x, y) position.
Who are the source data producers?
The dataset was produced by researchers at the Image Processing Laboratory (IPLab), Department of Mathematics and Computer Science, University of Catania, Italy.
Annotations
Annotation process
Camera poses (x, y, u, v) were obtained via an automated localisation pipeline. Zone labels correspond to manually defined rectangular floor regions; each frame receives the label of the zone containing its (x, y) position.
Who are the annotators?
Researchers at the University of Catania (Spera, Furnari, Battiato, Farinella).
Citation
BibTeX:
@article{spera2021egocart,
author = {Spera, E. and Furnari, A. and Battiato, S. and Farinella, G. M.},
title = {{EgoCart}: A Benchmark Dataset for Large-Scale Indoor Image-Based Localization in Retail Stores},
journal = {IEEE Transactions on Circuits and Systems for Video Technology},
volume = {31},
number = {4},
pages = {1253--1267},
year = {2021},
month = apr,
doi = {10.1109/TCSVT.2019.2941040}
}
APA:
Spera, E., Furnari, A., Battiato, S., & Farinella, G. M. (2021). EgoCart: A benchmark dataset for large-scale indoor image-based localization in retail stores. IEEE Transactions on Circuits and Systems for Video Technology, 31(4), 1253–1267. https://doi.org/10.1109/TCSVT.2019.2941040
- Downloads last month
- 2,234
