File size: 8,270 Bytes
ff57d41
fc21d7a
ff57d41
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4b87265
ff57d41
4b87265
ff57d41
4b87265
12f3ebf
ff57d41
4b87265
ff57d41
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3846287
 
 
ff57d41
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c38c22c
ff57d41
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6b7fa3e
ff57d41
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
---
license: other
license_name: scared-c-research-only
license_link: LICENSE
tags:
- medical
- surgical
- endoscopy
- depth-estimation
- rgb-d
- robotics
- computer-vision
- structure-from-motion
- da-vinci
pretty_name: SCARED-C
size_categories:
- 10K<n<100K
source_datasets:
- extended|scared
task_categories:
- depth-estimation
- image-to-image
configs:
- config_name: default
  data_files:
  - split: train
    path: "dataset_*/keyframe_*/data/frame_data.tar.gz"
  - split: validation
    path: "dataset_*/keyframe_*/data/frame_data.tar.gz"
---

# Dataset Card for SCARED-C

SCARED-C is a corrected version of the [SCARED](https://endovissub2019-scared.grand-challenge.org/) endoscopic depth estimation dataset. By replacing the original kinematics-based camera poses with poses re-estimated through Structure-from-Motion (COLMAP) followed by a metric scale recovery step, SCARED-C expands the number of reliable RGB-D pairs in SCARED from 35 keyframes to **17,135 frames** — a roughly 490× increase in reliably labeled real-tissue surgical data.

![SCARED vs SCARED-C](./scared_corrected.pdf)
*Left: original SCARED depth maps misaligned with RGB due to robot kinematics errors. Right: corrected SCARED-C depth maps after COLMAP-based pose re-estimation.*

## Dataset Details

### Dataset Description

The original SCARED dataset provides ground-truth depth maps for ex-vivo porcine abdominal scenes captured using a structured light sensor mounted on a da Vinci endoscope. Because the sensor can only capture depth from a static viewpoint, the dataset was extended by moving the endoscope arm and projecting the keyframe depth map into neighboring video frames using the robot's forward kinematics. In practice, the da Vinci system is cable-driven, and the resulting kinematics errors cause severe misalignment between the projected depth maps and their corresponding RGB images, rendering most non-keyframe data unsuitable for training.

SCARED-C addresses this limitation by:

1. Running [COLMAP](https://colmap.github.io/) (Structure-from-Motion) on the left camera frames of each video sequence to re-estimate camera poses directly from image data, jointly with the keyframe image.
2. Recovering metric scale via a simple algorithm that aligns the COLMAP sparse reconstruction to the structured-light keyframe depth map (using the median ratio of metric-to-unscaled depth).
3. Reprojecting the keyframe depth map through the corrected metric poses to produce reliable RGB-D pairs across the full video sequence.

After excluding frames that COLMAP fails to co-register and the entirety of datasets 4 and 5 (which have known calibration issues in the original release), the resulting dataset contains 17,135 RGB-D pairs across 12 keyframe sequences spanning datasets 1, 2, 3, 6, and 7.

- **Language(s):** N/A (image data)
- **License:** Inherits the license of the original SCARED dataset. Intended for non-commercial academic research use only. Users should verify they have access rights to the original SCARED data before using SCARED-C.

### Dataset Sources

- **Repository (data):** https://huggingface.co/datasets/juseonghan/SCARED-C
- **Repository (code):** https://github.com/juseonghan/SCARED-C
- **Paper:** SCARED-C: Corrected Camera Poses for Endoscopic Depth Estimation (Han et al.)
- **Original dataset:** [Allan et al., 2021](https://arxiv.org/abs/2101.01133) — Stereo Correspondence and Reconstruction of Endoscopic Data Challenge

## Uses

### Direct Use

SCARED-C is intended for research in surgical computer vision, particularly:

- Training and evaluating monocular and stereo **depth estimation** models on real surgical tissue.
- Benchmarking **Structure-from-Motion** and **SLAM** methods in endoscopic settings.
- Evaluating zero-shot generalization of foundation models (e.g., FoundationStereo, Depth Anything) to surgical imagery.

### Out-of-Scope Use

- **Clinical use of any kind.** This dataset consists of ex-vivo porcine abdominal scenes and is not suitable for training models intended for direct clinical deployment without further validation, regulatory approval, and in-vivo data.
- **Commercial use** is restricted by the underlying SCARED license terms.

## Dataset Structure

The format of SCARED-C is **identical to the original SCARED dataset**, with all `dataset_x/keyframe_y/data/frame_data.tar.gz` files replaced with corrected poses. Three additional files are provided per sequence for convenience:

```
dataset_x/
  keyframe_y/
    data/
      frame_data.tar.gz       # Corrected per-frame poses + reprojected depth maps
      rgb_frames.tar.gz       # Pre-extracted RGB frames corresponding to frame_log.json
      rgb.mp4                 # Original video, included for completeness
    frame_log.json            # List of frames successfully co-registered by COLMAP
    intrinsics_colmap.yaml    # COLMAP-refined intrinsics (readable via cv2.FileStorage)
    endoscope_calibration.yaml  # Original stereo calibration (unchanged)
```

**`frame_log.json`** lists all frames that COLMAP successfully co-registered with the keyframe. The vast majority of original frames are retained; the most notable exception is `dataset_2/keyframe_1`, which drops from 88 to 11 frames due to limited visual overlap.

**`intrinsics_colmap.yaml`** contains the intrinsics refined by COLMAP during bundle adjustment. We did not observe a meaningful difference in downstream performance versus the original intrinsics, but they are included for completeness. They can be loaded with `cv2.FileStorage()` in the same way as `endoscope_calibration.yaml`; both files are needed if you require stereo calibration.

**`rgb_frames.tar.gz`** is provided as a convenience — these are the registered frames extracted from `rgb.mp4`, in correspondence with `frame_log.json`.


Users are also recommended to use the [SCARED-toolkit](https://github.com/dimitrisPs/scared_toolkit) for data processing and extraction. 

### Splits

We provide a suggested ~70/30 train/validation split to enable consistent benchmarking across future work. Sequence codes follow the format `{dataset}_{keyframe}`:

| Split | Sequences | Frames |
|---|---|---|
| **Train** | 1_1, 1_3, 2_2, 2_4, 3_1, 3_2, 3_3, 6_1, 6_2, 6_3, 7_1, 7_4 | 12,330 |
| **Validation** | 1_2, 1_4, 1_5, 2_1, 2_3, 2_5, 3_4, 3_5, 6_4, 6_5, 7_2, 7_3, 7_5 | 4,805 |
| **Total** | 25 sequences | **17,135** |

The 25 keyframes (with structured-light ground-truth depth) from datasets 1, 2, 3, 6, and 7 are recommended as a held-out test set, since they are not part of any video sequence and therefore never appear in training.

## Dataset Creation

### Source Data

#### Data Collection and Processing

The source RGB images and structured-light keyframe depth maps come directly from the original SCARED dataset. Our processing pipeline is specified in the associated preprint. 

Datasets 4 and 5 from the original SCARED release have known calibration issues and are excluded.

#### Who are the source data producers?

The original SCARED data was collected by Intuitive Surgical and released through the EndoVis sub-challenge at MICCAI 2019. The scenes are ex-vivo porcine abdomens captured with a structured-light sensor mounted on a da Vinci endoscope.

### Annotations

#### Annotation process

No human annotation. Depth labels are derived from the structured-light keyframe (sensor-measured) and reprojected through COLMAP-estimated camera poses.

#### Personal and Sensitive Information

The dataset contains no human subjects. All scenes are ex-vivo porcine tissue.

## Validation

Please see the associated preprint for validation experiments.

## Citation

**BibTeX:**

```bibtex
@article{han2026scaredc,
  title   = {SCARED-C: Corrected Camera Poses for Endoscopic Depth Estimation},
  author  = {Han, John J. and Schmidt, Adam and Allan, Max and Wu, Jie Ying and Mohareri, Omid},
  journal = {arXiv preprint},
  year    = {2026}
}
```

## More Information

- Scale-recovery and processing code: https://github.com/juseonghan/SCARED-C
- Original SCARED challenge: https://endovissub2019-scared.grand-challenge.org/

## Dataset Card Authors

John J. Han

## Dataset Card Contact

Please open an issue on the [GitHub repository](https://github.com/juseonghan/SCARED-C/issues).