Datasets:
Add files using upload-large-folder tool
Browse files- .gitattributes +2 -0
- Perception/fine_grained_recognition/assets.tar.zst +3 -0
- Perception/fine_grained_recognition/fine_grained_recognition.json +0 -0
- Perception/object_counting/assets.tar.zst +3 -0
- Perception/object_counting/object_counting.json +0 -0
- Perception/object_localization/assets.tar.zst +3 -0
- Perception/object_localization/object_localization.json +0 -0
- Perception/object_presence/assets.tar.zst +3 -0
- Perception/object_presence/object_presence.json +0 -0
- README.md +190 -0
- Relational reasoning/aggregate_distribution/aggregate_distribution.json +0 -0
- Relational reasoning/aggregate_distribution/assets.tar.zst +3 -0
- Relational reasoning/directional/assets.tar.zst +3 -0
- Relational reasoning/directional/directional.json +0 -0
- Relational reasoning/projective_ordering/assets.tar.zst +3 -0
- Relational reasoning/projective_ordering/projective_ordering.json +0 -0
- Relational reasoning/proximity/assets.tar.zst +3 -0
- Relational reasoning/proximity/proximity.json +3 -0
- Relational reasoning/topological/assets.tar.zst +3 -0
- Relational reasoning/topological/topological.json +0 -0
- Temporal reasoning/category_turnover/assets.tar.zst +3 -0
- Temporal reasoning/category_turnover/category_turnover.json +0 -0
- Temporal reasoning/net_change/assets.tar.zst +3 -0
- Temporal reasoning/net_change/net_change.json +0 -0
- Temporal reasoning/semantic_transition/assets.tar.zst +3 -0
- Temporal reasoning/semantic_transition/semantic_transition.json +0 -0
- croissant.json +476 -0
- dataset_manifest.json +135 -0
- metadata.jsonl +3 -0
.gitattributes
CHANGED
|
@@ -58,3 +58,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 58 |
# Video files - compressed
|
| 59 |
*.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 60 |
*.webm filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
| 58 |
# Video files - compressed
|
| 59 |
*.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 60 |
*.webm filter=lfs diff=lfs merge=lfs -text
|
| 61 |
+
metadata.jsonl filter=lfs diff=lfs merge=lfs -text
|
| 62 |
+
Relational[[:space:]]reasoning/proximity/proximity.json filter=lfs diff=lfs merge=lfs -text
|
Perception/fine_grained_recognition/assets.tar.zst
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b21bae4c289a75058c839250150e571cd6d424f3d7a54000c473028c8bec19f4
|
| 3 |
+
size 1417644607
|
Perception/fine_grained_recognition/fine_grained_recognition.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
Perception/object_counting/assets.tar.zst
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b35dc7d7f8e9df24b781df00c1fda222c51d46be0a4e32ad0d63c92114c4ff8d
|
| 3 |
+
size 2131892267
|
Perception/object_counting/object_counting.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
Perception/object_localization/assets.tar.zst
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4661374ef80a36d25cc0268328f5fcc72415a018d2241f1e2a6f20ec5583fbea
|
| 3 |
+
size 2495263000
|
Perception/object_localization/object_localization.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
Perception/object_presence/assets.tar.zst
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a43b3ee1d2afcd3cc81e6cd830140cec4eeedcaee138381644138e1db336ecf6
|
| 3 |
+
size 1554646842
|
Perception/object_presence/object_presence.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
README.md
ADDED
|
@@ -0,0 +1,190 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: cc-by-nc-4.0
|
| 3 |
+
language:
|
| 4 |
+
- en
|
| 5 |
+
pretty_name: RSFaith-Bench
|
| 6 |
+
size_categories:
|
| 7 |
+
- 10K<n<100K
|
| 8 |
+
task_categories:
|
| 9 |
+
- visual-question-answering
|
| 10 |
+
tags:
|
| 11 |
+
- remote-sensing
|
| 12 |
+
- vision-language
|
| 13 |
+
- benchmark
|
| 14 |
+
- scene-graph
|
| 15 |
+
- multiple-choice
|
| 16 |
+
- geospatial
|
| 17 |
+
---
|
| 18 |
+
|
| 19 |
+
# RSFaith-Bench
|
| 20 |
+
|
| 21 |
+
RSFaith-Bench is a remote-sensing vision-language benchmark designed to
|
| 22 |
+
evaluate grounded visual reasoning beyond surface-level object recognition.
|
| 23 |
+
The benchmark covers perception, relational reasoning, and temporal reasoning
|
| 24 |
+
over remote-sensing imagery. Each example is formulated as a multiple-choice
|
| 25 |
+
question and is paired with a compact scene graph, supporting evidence, and an
|
| 26 |
+
executable reasoning program.
|
| 27 |
+
|
| 28 |
+
The release contains 13,511 question-answer records, 16,288 referenced images,
|
| 29 |
+
and 12,876 compact scene graphs.
|
| 30 |
+
|
| 31 |
+
## Using the Dataset
|
| 32 |
+
|
| 33 |
+
The annotation files can be loaded directly as JSON. The root-level
|
| 34 |
+
`metadata.jsonl` provides a flat index over all records:
|
| 35 |
+
|
| 36 |
+
```python
|
| 37 |
+
from datasets import load_dataset
|
| 38 |
+
|
| 39 |
+
dataset = load_dataset(
|
| 40 |
+
"json",
|
| 41 |
+
data_files={"benchmark": "metadata.jsonl"},
|
| 42 |
+
split="benchmark",
|
| 43 |
+
)
|
| 44 |
+
print(dataset[0])
|
| 45 |
+
```
|
| 46 |
+
|
| 47 |
+
Images and scene graphs are stored in per-subcategory archives. To restore the
|
| 48 |
+
file layout referenced by the JSON records, extract the archives in place:
|
| 49 |
+
|
| 50 |
+
```bash
|
| 51 |
+
huggingface-cli download <namespace>/RSFaith-Bench \
|
| 52 |
+
--repo-type dataset \
|
| 53 |
+
--local-dir RSFaith-Bench
|
| 54 |
+
|
| 55 |
+
find RSFaith-Bench -name assets.tar.zst -print0 |
|
| 56 |
+
while IFS= read -r -d '' archive; do
|
| 57 |
+
(cd "$(dirname "$archive")" && tar -I zstd -xf assets.tar.zst)
|
| 58 |
+
done
|
| 59 |
+
```
|
| 60 |
+
|
| 61 |
+
After extraction, the `images` and `scene_graph` fields in each subcategory
|
| 62 |
+
JSON file resolve relative to that subcategory directory. The `image_t1`,
|
| 63 |
+
`image_t2`, and `scene_graph` fields in `metadata.jsonl` resolve relative to
|
| 64 |
+
the repository root.
|
| 65 |
+
|
| 66 |
+
## File Organization
|
| 67 |
+
|
| 68 |
+
The dataset is organized by reasoning level and subcategory. Each subcategory
|
| 69 |
+
directory contains:
|
| 70 |
+
|
| 71 |
+
- `<subcategory>.json`: question-answer records for the subcategory.
|
| 72 |
+
- `assets.tar.zst`: compressed `images/` and `scene_graphs/` directories.
|
| 73 |
+
|
| 74 |
+
```
|
| 75 |
+
RSFaith-Bench
|
| 76 |
+
├── README.md
|
| 77 |
+
├── metadata.jsonl
|
| 78 |
+
├── dataset_manifest.json
|
| 79 |
+
├── croissant.json
|
| 80 |
+
├── Perception
|
| 81 |
+
│ ├── object_presence
|
| 82 |
+
│ │ ├── object_presence.json
|
| 83 |
+
│ │ └── assets.tar.zst
|
| 84 |
+
│ ├── object_counting
|
| 85 |
+
│ │ ├── object_counting.json
|
| 86 |
+
│ │ └── assets.tar.zst
|
| 87 |
+
│ ├── fine_grained_recognition
|
| 88 |
+
│ │ ├── fine_grained_recognition.json
|
| 89 |
+
│ │ └── assets.tar.zst
|
| 90 |
+
│ └── object_localization
|
| 91 |
+
│ ├── object_localization.json
|
| 92 |
+
│ └── assets.tar.zst
|
| 93 |
+
├── Relational reasoning
|
| 94 |
+
│ ├── directional
|
| 95 |
+
│ │ ├── directional.json
|
| 96 |
+
│ │ └── assets.tar.zst
|
| 97 |
+
│ ├── topological
|
| 98 |
+
│ │ ├── topological.json
|
| 99 |
+
│ │ └── assets.tar.zst
|
| 100 |
+
│ ├── proximity
|
| 101 |
+
│ │ ├── proximity.json
|
| 102 |
+
│ │ └── assets.tar.zst
|
| 103 |
+
│ ├── projective_ordering
|
| 104 |
+
│ │ ├── projective_ordering.json
|
| 105 |
+
│ │ └── assets.tar.zst
|
| 106 |
+
│ └── aggregate_distribution
|
| 107 |
+
│ ├── aggregate_distribution.json
|
| 108 |
+
│ └── assets.tar.zst
|
| 109 |
+
└── Temporal reasoning
|
| 110 |
+
├── category_turnover
|
| 111 |
+
│ ├── category_turnover.json
|
| 112 |
+
│ └── assets.tar.zst
|
| 113 |
+
├── net_change
|
| 114 |
+
│ ├── net_change.json
|
| 115 |
+
│ └── assets.tar.zst
|
| 116 |
+
└── semantic_transition
|
| 117 |
+
├── semantic_transition.json
|
| 118 |
+
└── assets.tar.zst
|
| 119 |
+
```
|
| 120 |
+
|
| 121 |
+
## Data Fields
|
| 122 |
+
|
| 123 |
+
Each question-answer record contains the following fields:
|
| 124 |
+
|
| 125 |
+
- `question_id`: anonymized question identifier.
|
| 126 |
+
- `scene_id`: anonymized scene identifier.
|
| 127 |
+
- `level`: high-level reasoning category.
|
| 128 |
+
- `subcategory`: fine-grained reasoning category.
|
| 129 |
+
- `question`: natural-language question.
|
| 130 |
+
- `answer`: correct answer.
|
| 131 |
+
- `answer_type`: answer representation.
|
| 132 |
+
- `choices`: multiple-choice options.
|
| 133 |
+
- `images`: relative image paths.
|
| 134 |
+
- `scene_graph`: relative scene graph path.
|
| 135 |
+
- `support`: grounded support evidence.
|
| 136 |
+
- `program`: executable reasoning specification.
|
| 137 |
+
|
| 138 |
+
## Dataset Statistics
|
| 139 |
+
|
| 140 |
+
| Level | Subcategory | Records |
|
| 141 |
+
| --- | --- | ---: |
|
| 142 |
+
| Perception | object_presence | 969 |
|
| 143 |
+
| Perception | object_counting | 1,085 |
|
| 144 |
+
| Perception | fine_grained_recognition | 1,228 |
|
| 145 |
+
| Perception | object_localization | 1,298 |
|
| 146 |
+
| Relational reasoning | directional | 1,166 |
|
| 147 |
+
| Relational reasoning | topological | 921 |
|
| 148 |
+
| Relational reasoning | proximity | 987 |
|
| 149 |
+
| Relational reasoning | projective_ordering | 944 |
|
| 150 |
+
| Relational reasoning | aggregate_distribution | 866 |
|
| 151 |
+
| Temporal reasoning | category_turnover | 1,515 |
|
| 152 |
+
| Temporal reasoning | net_change | 1,273 |
|
| 153 |
+
| Temporal reasoning | semantic_transition | 1,259 |
|
| 154 |
+
|
| 155 |
+
## Dataset Construction
|
| 156 |
+
|
| 157 |
+
RSFaith-Bench is constructed from remote-sensing scenes represented as grounded scene graphs. The scene graphs encode objects, spatial relations, temporal
|
| 158 |
+
changes, and compact global inventories when applicable. Question-answer pairs
|
| 159 |
+
are generated from programmatic templates and then curated to balance reasoning
|
| 160 |
+
categories, answer distributions, and scene coverage. The released records
|
| 161 |
+
retain the reasoning support and program specification so that each answer can
|
| 162 |
+
be traced back to the corresponding scene graph.
|
| 163 |
+
|
| 164 |
+
## Licensing
|
| 165 |
+
|
| 166 |
+
The dataset is released under the [Creative Commons Attribution Non Commercial 4.0](https://creativecommons.org/licenses/by-nc/4.0/deed.en), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
|
| 167 |
+
|
| 168 |
+
## Citation Information
|
| 169 |
+
|
| 170 |
+
If you use RSFaith-Bench in your research, please cite the accompanying paper:
|
| 171 |
+
|
| 172 |
+
```bibtex
|
| 173 |
+
@misc{rsfaithbench2026,
|
| 174 |
+
title = {RSFaith-Bench: When Correct Answers Come with Unfaithful Evidence in Remote Sensing MLLMs},
|
| 175 |
+
author = {Anonymous},
|
| 176 |
+
year = {2026}
|
| 177 |
+
}
|
| 178 |
+
```
|
| 179 |
+
|
| 180 |
+
## Acknowledgement
|
| 181 |
+
|
| 182 |
+
RSFaith-Bench is built from remote-sensing data sources including
|
| 183 |
+
[DIOR](https://gcheng-nwpu.github.io/),
|
| 184 |
+
[DOTA](https://captain-whu.github.io/DOTA/dataset.html),
|
| 185 |
+
[FAIR1M](https://www.gaofen-challenge.com/benchmark),
|
| 186 |
+
[SECOND](https://github.com/GeoZcx/A-deeply-supervised-image-fusion-network-for-change-detection-in-remote-sensing-images/tree/master/dataset),
|
| 187 |
+
[xBD](https://xview2.org/dataset), and
|
| 188 |
+
[ReCon1M](https://arxiv.org/abs/2406.06028). We thank the creators and
|
| 189 |
+
maintainers of these datasets for making their resources available to the
|
| 190 |
+
research community.
|
Relational reasoning/aggregate_distribution/aggregate_distribution.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
Relational reasoning/aggregate_distribution/assets.tar.zst
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c78328a90c8555656ae64a9a6437d1b24a23793a2fadef5ff6a9b1ddc3c4b3f0
|
| 3 |
+
size 5081368574
|
Relational reasoning/directional/assets.tar.zst
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ed8f2a58ea90c6759f66d7efc0dce539156c041c5a3726e87904e57776c53e12
|
| 3 |
+
size 1833785421
|
Relational reasoning/directional/directional.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
Relational reasoning/projective_ordering/assets.tar.zst
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3cc1f0cb5a7f41846ebe02023337da02ed8b2d059fa80d63c142e9abe0c20c12
|
| 3 |
+
size 1473705066
|
Relational reasoning/projective_ordering/projective_ordering.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
Relational reasoning/proximity/assets.tar.zst
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:47314a336e421cab483087d3520a3f12a91a99600f3ab53a883cd75b9f61d471
|
| 3 |
+
size 1490183220
|
Relational reasoning/proximity/proximity.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5cca3b26e1802115b7dc7643941c7a84d14d1ca1f28cc8e66fd6ef60867c759f
|
| 3 |
+
size 10594022
|
Relational reasoning/topological/assets.tar.zst
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fdf95a91ed5c5da86a4602d391adb17c7599c96bd045b49891caa0f59278b7bb
|
| 3 |
+
size 1238362287
|
Relational reasoning/topological/topological.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
Temporal reasoning/category_turnover/assets.tar.zst
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ce0d134db9d80e6146ae27ba88565db0d6f52bbce109aaf958d502e89c55b63d
|
| 3 |
+
size 1103133575
|
Temporal reasoning/category_turnover/category_turnover.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
Temporal reasoning/net_change/assets.tar.zst
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fb16fdc4375f6287e90b8f4cc39ff4df72d672e15b0c8d7be47e5f5e3f531800
|
| 3 |
+
size 2299874948
|
Temporal reasoning/net_change/net_change.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
Temporal reasoning/semantic_transition/assets.tar.zst
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f8c7d7d4af9fdac7f0500a8dead9824c07eeac1c1acf2f05126aa59ca3af45ca
|
| 3 |
+
size 689231000
|
Temporal reasoning/semantic_transition/semantic_transition.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
croissant.json
ADDED
|
@@ -0,0 +1,476 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"@context": {
|
| 3 |
+
"@language": "en",
|
| 4 |
+
"@vocab": "https://schema.org/",
|
| 5 |
+
"cr": "http://mlcommons.org/croissant/",
|
| 6 |
+
"sc": "https://schema.org/",
|
| 7 |
+
"recordSet": "cr:recordSet",
|
| 8 |
+
"field": "cr:field",
|
| 9 |
+
"dataType": "cr:dataType",
|
| 10 |
+
"source": "cr:source",
|
| 11 |
+
"fileObject": "cr:fileObject",
|
| 12 |
+
"extract": "cr:extract",
|
| 13 |
+
"column": "cr:column"
|
| 14 |
+
},
|
| 15 |
+
"@type": "sc:Dataset",
|
| 16 |
+
"name": "RSFaith-Bench",
|
| 17 |
+
"alternateName": "RSFaith-Bench",
|
| 18 |
+
"description": "RSFaith-Bench is a remote-sensing vision-language benchmark with multiple-choice question-answer records, compact scene graphs, supporting evidence, and executable reasoning programs.",
|
| 19 |
+
"version": "1.0.0",
|
| 20 |
+
"license": "https://creativecommons.org/licenses/by-nc/4.0/deed.en",
|
| 21 |
+
"inLanguage": "en",
|
| 22 |
+
"keywords": [
|
| 23 |
+
"remote sensing",
|
| 24 |
+
"visual question answering",
|
| 25 |
+
"vision-language benchmark",
|
| 26 |
+
"scene graph",
|
| 27 |
+
"multiple choice",
|
| 28 |
+
"geospatial"
|
| 29 |
+
],
|
| 30 |
+
"distribution": [
|
| 31 |
+
{
|
| 32 |
+
"@type": "cr:FileObject",
|
| 33 |
+
"@id": "metadata_jsonl",
|
| 34 |
+
"name": "metadata.jsonl",
|
| 35 |
+
"contentUrl": "metadata.jsonl",
|
| 36 |
+
"encodingFormat": "application/jsonl",
|
| 37 |
+
"contentSize": 42232446,
|
| 38 |
+
"description": "Flat index over all RSFaith-Bench records."
|
| 39 |
+
},
|
| 40 |
+
{
|
| 41 |
+
"@type": "cr:FileObject",
|
| 42 |
+
"@id": "dataset_manifest_json",
|
| 43 |
+
"name": "dataset_manifest.json",
|
| 44 |
+
"contentUrl": "dataset_manifest.json",
|
| 45 |
+
"encodingFormat": "application/json",
|
| 46 |
+
"contentSize": 4475,
|
| 47 |
+
"description": "Dataset counts and archive manifest."
|
| 48 |
+
},
|
| 49 |
+
{
|
| 50 |
+
"@type": "cr:FileObject",
|
| 51 |
+
"@id": "Perception__fine_grained_recognition__records_json",
|
| 52 |
+
"name": "fine_grained_recognition.json",
|
| 53 |
+
"contentUrl": "Perception/fine_grained_recognition/fine_grained_recognition.json",
|
| 54 |
+
"encodingFormat": "application/json",
|
| 55 |
+
"contentSize": 3110650,
|
| 56 |
+
"description": "Question-answer records for Perception/fine_grained_recognition."
|
| 57 |
+
},
|
| 58 |
+
{
|
| 59 |
+
"@type": "cr:FileObject",
|
| 60 |
+
"@id": "Perception__fine_grained_recognition__assets_archive",
|
| 61 |
+
"name": "assets.tar.zst",
|
| 62 |
+
"contentUrl": "Perception/fine_grained_recognition/assets.tar.zst",
|
| 63 |
+
"encodingFormat": "application/zstd",
|
| 64 |
+
"contentSize": 1417644607,
|
| 65 |
+
"description": "Zstandard-compressed tar archive containing images/ and scene_graphs/ for Perception/fine_grained_recognition."
|
| 66 |
+
},
|
| 67 |
+
{
|
| 68 |
+
"@type": "cr:FileObject",
|
| 69 |
+
"@id": "Perception__object_counting__records_json",
|
| 70 |
+
"name": "object_counting.json",
|
| 71 |
+
"contentUrl": "Perception/object_counting/object_counting.json",
|
| 72 |
+
"encodingFormat": "application/json",
|
| 73 |
+
"contentSize": 2898226,
|
| 74 |
+
"description": "Question-answer records for Perception/object_counting."
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"@type": "cr:FileObject",
|
| 78 |
+
"@id": "Perception__object_counting__assets_archive",
|
| 79 |
+
"name": "assets.tar.zst",
|
| 80 |
+
"contentUrl": "Perception/object_counting/assets.tar.zst",
|
| 81 |
+
"encodingFormat": "application/zstd",
|
| 82 |
+
"contentSize": 2131892267,
|
| 83 |
+
"description": "Zstandard-compressed tar archive containing images/ and scene_graphs/ for Perception/object_counting."
|
| 84 |
+
},
|
| 85 |
+
{
|
| 86 |
+
"@type": "cr:FileObject",
|
| 87 |
+
"@id": "Perception__object_localization__records_json",
|
| 88 |
+
"name": "object_localization.json",
|
| 89 |
+
"contentUrl": "Perception/object_localization/object_localization.json",
|
| 90 |
+
"encodingFormat": "application/json",
|
| 91 |
+
"contentSize": 2803453,
|
| 92 |
+
"description": "Question-answer records for Perception/object_localization."
|
| 93 |
+
},
|
| 94 |
+
{
|
| 95 |
+
"@type": "cr:FileObject",
|
| 96 |
+
"@id": "Perception__object_localization__assets_archive",
|
| 97 |
+
"name": "assets.tar.zst",
|
| 98 |
+
"contentUrl": "Perception/object_localization/assets.tar.zst",
|
| 99 |
+
"encodingFormat": "application/zstd",
|
| 100 |
+
"contentSize": 2495263000,
|
| 101 |
+
"description": "Zstandard-compressed tar archive containing images/ and scene_graphs/ for Perception/object_localization."
|
| 102 |
+
},
|
| 103 |
+
{
|
| 104 |
+
"@type": "cr:FileObject",
|
| 105 |
+
"@id": "Perception__object_presence__records_json",
|
| 106 |
+
"name": "object_presence.json",
|
| 107 |
+
"contentUrl": "Perception/object_presence/object_presence.json",
|
| 108 |
+
"encodingFormat": "application/json",
|
| 109 |
+
"contentSize": 1491906,
|
| 110 |
+
"description": "Question-answer records for Perception/object_presence."
|
| 111 |
+
},
|
| 112 |
+
{
|
| 113 |
+
"@type": "cr:FileObject",
|
| 114 |
+
"@id": "Perception__object_presence__assets_archive",
|
| 115 |
+
"name": "assets.tar.zst",
|
| 116 |
+
"contentUrl": "Perception/object_presence/assets.tar.zst",
|
| 117 |
+
"encodingFormat": "application/zstd",
|
| 118 |
+
"contentSize": 1554646842,
|
| 119 |
+
"description": "Zstandard-compressed tar archive containing images/ and scene_graphs/ for Perception/object_presence."
|
| 120 |
+
},
|
| 121 |
+
{
|
| 122 |
+
"@type": "cr:FileObject",
|
| 123 |
+
"@id": "Relational_reasoning__aggregate_distribution__records_json",
|
| 124 |
+
"name": "aggregate_distribution.json",
|
| 125 |
+
"contentUrl": "Relational reasoning/aggregate_distribution/aggregate_distribution.json",
|
| 126 |
+
"encodingFormat": "application/json",
|
| 127 |
+
"contentSize": 7018837,
|
| 128 |
+
"description": "Question-answer records for Relational reasoning/aggregate_distribution."
|
| 129 |
+
},
|
| 130 |
+
{
|
| 131 |
+
"@type": "cr:FileObject",
|
| 132 |
+
"@id": "Relational_reasoning__aggregate_distribution__assets_archive",
|
| 133 |
+
"name": "assets.tar.zst",
|
| 134 |
+
"contentUrl": "Relational reasoning/aggregate_distribution/assets.tar.zst",
|
| 135 |
+
"encodingFormat": "application/zstd",
|
| 136 |
+
"contentSize": 5081368574,
|
| 137 |
+
"description": "Zstandard-compressed tar archive containing images/ and scene_graphs/ for Relational reasoning/aggregate_distribution."
|
| 138 |
+
},
|
| 139 |
+
{
|
| 140 |
+
"@type": "cr:FileObject",
|
| 141 |
+
"@id": "Relational_reasoning__directional__records_json",
|
| 142 |
+
"name": "directional.json",
|
| 143 |
+
"contentUrl": "Relational reasoning/directional/directional.json",
|
| 144 |
+
"encodingFormat": "application/json",
|
| 145 |
+
"contentSize": 4697127,
|
| 146 |
+
"description": "Question-answer records for Relational reasoning/directional."
|
| 147 |
+
},
|
| 148 |
+
{
|
| 149 |
+
"@type": "cr:FileObject",
|
| 150 |
+
"@id": "Relational_reasoning__directional__assets_archive",
|
| 151 |
+
"name": "assets.tar.zst",
|
| 152 |
+
"contentUrl": "Relational reasoning/directional/assets.tar.zst",
|
| 153 |
+
"encodingFormat": "application/zstd",
|
| 154 |
+
"contentSize": 1833785421,
|
| 155 |
+
"description": "Zstandard-compressed tar archive containing images/ and scene_graphs/ for Relational reasoning/directional."
|
| 156 |
+
},
|
| 157 |
+
{
|
| 158 |
+
"@type": "cr:FileObject",
|
| 159 |
+
"@id": "Relational_reasoning__projective_ordering__records_json",
|
| 160 |
+
"name": "projective_ordering.json",
|
| 161 |
+
"contentUrl": "Relational reasoning/projective_ordering/projective_ordering.json",
|
| 162 |
+
"encodingFormat": "application/json",
|
| 163 |
+
"contentSize": 5354614,
|
| 164 |
+
"description": "Question-answer records for Relational reasoning/projective_ordering."
|
| 165 |
+
},
|
| 166 |
+
{
|
| 167 |
+
"@type": "cr:FileObject",
|
| 168 |
+
"@id": "Relational_reasoning__projective_ordering__assets_archive",
|
| 169 |
+
"name": "assets.tar.zst",
|
| 170 |
+
"contentUrl": "Relational reasoning/projective_ordering/assets.tar.zst",
|
| 171 |
+
"encodingFormat": "application/zstd",
|
| 172 |
+
"contentSize": 1473705066,
|
| 173 |
+
"description": "Zstandard-compressed tar archive containing images/ and scene_graphs/ for Relational reasoning/projective_ordering."
|
| 174 |
+
},
|
| 175 |
+
{
|
| 176 |
+
"@type": "cr:FileObject",
|
| 177 |
+
"@id": "Relational_reasoning__proximity__records_json",
|
| 178 |
+
"name": "proximity.json",
|
| 179 |
+
"contentUrl": "Relational reasoning/proximity/proximity.json",
|
| 180 |
+
"encodingFormat": "application/json",
|
| 181 |
+
"contentSize": 10594022,
|
| 182 |
+
"description": "Question-answer records for Relational reasoning/proximity."
|
| 183 |
+
},
|
| 184 |
+
{
|
| 185 |
+
"@type": "cr:FileObject",
|
| 186 |
+
"@id": "Relational_reasoning__proximity__assets_archive",
|
| 187 |
+
"name": "assets.tar.zst",
|
| 188 |
+
"contentUrl": "Relational reasoning/proximity/assets.tar.zst",
|
| 189 |
+
"encodingFormat": "application/zstd",
|
| 190 |
+
"contentSize": 1490183220,
|
| 191 |
+
"description": "Zstandard-compressed tar archive containing images/ and scene_graphs/ for Relational reasoning/proximity."
|
| 192 |
+
},
|
| 193 |
+
{
|
| 194 |
+
"@type": "cr:FileObject",
|
| 195 |
+
"@id": "Relational_reasoning__topological__records_json",
|
| 196 |
+
"name": "topological.json",
|
| 197 |
+
"contentUrl": "Relational reasoning/topological/topological.json",
|
| 198 |
+
"encodingFormat": "application/json",
|
| 199 |
+
"contentSize": 3784579,
|
| 200 |
+
"description": "Question-answer records for Relational reasoning/topological."
|
| 201 |
+
},
|
| 202 |
+
{
|
| 203 |
+
"@type": "cr:FileObject",
|
| 204 |
+
"@id": "Relational_reasoning__topological__assets_archive",
|
| 205 |
+
"name": "assets.tar.zst",
|
| 206 |
+
"contentUrl": "Relational reasoning/topological/assets.tar.zst",
|
| 207 |
+
"encodingFormat": "application/zstd",
|
| 208 |
+
"contentSize": 1238362287,
|
| 209 |
+
"description": "Zstandard-compressed tar archive containing images/ and scene_graphs/ for Relational reasoning/topological."
|
| 210 |
+
},
|
| 211 |
+
{
|
| 212 |
+
"@type": "cr:FileObject",
|
| 213 |
+
"@id": "Temporal_reasoning__category_turnover__records_json",
|
| 214 |
+
"name": "category_turnover.json",
|
| 215 |
+
"contentUrl": "Temporal reasoning/category_turnover/category_turnover.json",
|
| 216 |
+
"encodingFormat": "application/json",
|
| 217 |
+
"contentSize": 4609363,
|
| 218 |
+
"description": "Question-answer records for Temporal reasoning/category_turnover."
|
| 219 |
+
},
|
| 220 |
+
{
|
| 221 |
+
"@type": "cr:FileObject",
|
| 222 |
+
"@id": "Temporal_reasoning__category_turnover__assets_archive",
|
| 223 |
+
"name": "assets.tar.zst",
|
| 224 |
+
"contentUrl": "Temporal reasoning/category_turnover/assets.tar.zst",
|
| 225 |
+
"encodingFormat": "application/zstd",
|
| 226 |
+
"contentSize": 1103133575,
|
| 227 |
+
"description": "Zstandard-compressed tar archive containing images/ and scene_graphs/ for Temporal reasoning/category_turnover."
|
| 228 |
+
},
|
| 229 |
+
{
|
| 230 |
+
"@type": "cr:FileObject",
|
| 231 |
+
"@id": "Temporal_reasoning__net_change__records_json",
|
| 232 |
+
"name": "net_change.json",
|
| 233 |
+
"contentUrl": "Temporal reasoning/net_change/net_change.json",
|
| 234 |
+
"encodingFormat": "application/json",
|
| 235 |
+
"contentSize": 9985480,
|
| 236 |
+
"description": "Question-answer records for Temporal reasoning/net_change."
|
| 237 |
+
},
|
| 238 |
+
{
|
| 239 |
+
"@type": "cr:FileObject",
|
| 240 |
+
"@id": "Temporal_reasoning__net_change__assets_archive",
|
| 241 |
+
"name": "assets.tar.zst",
|
| 242 |
+
"contentUrl": "Temporal reasoning/net_change/assets.tar.zst",
|
| 243 |
+
"encodingFormat": "application/zstd",
|
| 244 |
+
"contentSize": 2299874948,
|
| 245 |
+
"description": "Zstandard-compressed tar archive containing images/ and scene_graphs/ for Temporal reasoning/net_change."
|
| 246 |
+
},
|
| 247 |
+
{
|
| 248 |
+
"@type": "cr:FileObject",
|
| 249 |
+
"@id": "Temporal_reasoning__semantic_transition__records_json",
|
| 250 |
+
"name": "semantic_transition.json",
|
| 251 |
+
"contentUrl": "Temporal reasoning/semantic_transition/semantic_transition.json",
|
| 252 |
+
"encodingFormat": "application/json",
|
| 253 |
+
"contentSize": 9784610,
|
| 254 |
+
"description": "Question-answer records for Temporal reasoning/semantic_transition."
|
| 255 |
+
},
|
| 256 |
+
{
|
| 257 |
+
"@type": "cr:FileObject",
|
| 258 |
+
"@id": "Temporal_reasoning__semantic_transition__assets_archive",
|
| 259 |
+
"name": "assets.tar.zst",
|
| 260 |
+
"contentUrl": "Temporal reasoning/semantic_transition/assets.tar.zst",
|
| 261 |
+
"encodingFormat": "application/zstd",
|
| 262 |
+
"contentSize": 689231000,
|
| 263 |
+
"description": "Zstandard-compressed tar archive containing images/ and scene_graphs/ for Temporal reasoning/semantic_transition."
|
| 264 |
+
}
|
| 265 |
+
],
|
| 266 |
+
"recordSet": [
|
| 267 |
+
{
|
| 268 |
+
"@type": "cr:RecordSet",
|
| 269 |
+
"name": "qa_records",
|
| 270 |
+
"description": "Flat metadata index over all RSFaith-Bench question-answer records.",
|
| 271 |
+
"source": {
|
| 272 |
+
"fileObject": {
|
| 273 |
+
"@id": "metadata_jsonl"
|
| 274 |
+
}
|
| 275 |
+
},
|
| 276 |
+
"field": [
|
| 277 |
+
{
|
| 278 |
+
"@type": "cr:Field",
|
| 279 |
+
"name": "question_id",
|
| 280 |
+
"description": "Anonymized question identifier.",
|
| 281 |
+
"dataType": "sc:Text",
|
| 282 |
+
"source": {
|
| 283 |
+
"fileObject": {
|
| 284 |
+
"@id": "metadata_jsonl"
|
| 285 |
+
},
|
| 286 |
+
"extract": {
|
| 287 |
+
"column": "question_id"
|
| 288 |
+
}
|
| 289 |
+
}
|
| 290 |
+
},
|
| 291 |
+
{
|
| 292 |
+
"@type": "cr:Field",
|
| 293 |
+
"name": "scene_id",
|
| 294 |
+
"description": "Anonymized scene identifier.",
|
| 295 |
+
"dataType": "sc:Text",
|
| 296 |
+
"source": {
|
| 297 |
+
"fileObject": {
|
| 298 |
+
"@id": "metadata_jsonl"
|
| 299 |
+
},
|
| 300 |
+
"extract": {
|
| 301 |
+
"column": "scene_id"
|
| 302 |
+
}
|
| 303 |
+
}
|
| 304 |
+
},
|
| 305 |
+
{
|
| 306 |
+
"@type": "cr:Field",
|
| 307 |
+
"name": "level",
|
| 308 |
+
"description": "High-level reasoning category.",
|
| 309 |
+
"dataType": "sc:Text",
|
| 310 |
+
"source": {
|
| 311 |
+
"fileObject": {
|
| 312 |
+
"@id": "metadata_jsonl"
|
| 313 |
+
},
|
| 314 |
+
"extract": {
|
| 315 |
+
"column": "level"
|
| 316 |
+
}
|
| 317 |
+
}
|
| 318 |
+
},
|
| 319 |
+
{
|
| 320 |
+
"@type": "cr:Field",
|
| 321 |
+
"name": "subcategory",
|
| 322 |
+
"description": "Fine-grained reasoning category.",
|
| 323 |
+
"dataType": "sc:Text",
|
| 324 |
+
"source": {
|
| 325 |
+
"fileObject": {
|
| 326 |
+
"@id": "metadata_jsonl"
|
| 327 |
+
},
|
| 328 |
+
"extract": {
|
| 329 |
+
"column": "subcategory"
|
| 330 |
+
}
|
| 331 |
+
}
|
| 332 |
+
},
|
| 333 |
+
{
|
| 334 |
+
"@type": "cr:Field",
|
| 335 |
+
"name": "question",
|
| 336 |
+
"description": "Natural-language question.",
|
| 337 |
+
"dataType": "sc:Text",
|
| 338 |
+
"source": {
|
| 339 |
+
"fileObject": {
|
| 340 |
+
"@id": "metadata_jsonl"
|
| 341 |
+
},
|
| 342 |
+
"extract": {
|
| 343 |
+
"column": "question"
|
| 344 |
+
}
|
| 345 |
+
}
|
| 346 |
+
},
|
| 347 |
+
{
|
| 348 |
+
"@type": "cr:Field",
|
| 349 |
+
"name": "answer",
|
| 350 |
+
"description": "Correct answer.",
|
| 351 |
+
"dataType": "sc:Text",
|
| 352 |
+
"source": {
|
| 353 |
+
"fileObject": {
|
| 354 |
+
"@id": "metadata_jsonl"
|
| 355 |
+
},
|
| 356 |
+
"extract": {
|
| 357 |
+
"column": "answer"
|
| 358 |
+
}
|
| 359 |
+
}
|
| 360 |
+
},
|
| 361 |
+
{
|
| 362 |
+
"@type": "cr:Field",
|
| 363 |
+
"name": "answer_type",
|
| 364 |
+
"description": "Answer representation.",
|
| 365 |
+
"dataType": "sc:Text",
|
| 366 |
+
"source": {
|
| 367 |
+
"fileObject": {
|
| 368 |
+
"@id": "metadata_jsonl"
|
| 369 |
+
},
|
| 370 |
+
"extract": {
|
| 371 |
+
"column": "answer_type"
|
| 372 |
+
}
|
| 373 |
+
}
|
| 374 |
+
},
|
| 375 |
+
{
|
| 376 |
+
"@type": "cr:Field",
|
| 377 |
+
"name": "choices",
|
| 378 |
+
"description": "Multiple-choice options encoded as JSON.",
|
| 379 |
+
"dataType": "sc:Text",
|
| 380 |
+
"source": {
|
| 381 |
+
"fileObject": {
|
| 382 |
+
"@id": "metadata_jsonl"
|
| 383 |
+
},
|
| 384 |
+
"extract": {
|
| 385 |
+
"column": "choices"
|
| 386 |
+
}
|
| 387 |
+
}
|
| 388 |
+
},
|
| 389 |
+
{
|
| 390 |
+
"@type": "cr:Field",
|
| 391 |
+
"name": "image_t1",
|
| 392 |
+
"description": "Relative path to the first image after archive extraction.",
|
| 393 |
+
"dataType": "sc:Text",
|
| 394 |
+
"source": {
|
| 395 |
+
"fileObject": {
|
| 396 |
+
"@id": "metadata_jsonl"
|
| 397 |
+
},
|
| 398 |
+
"extract": {
|
| 399 |
+
"column": "image_t1"
|
| 400 |
+
}
|
| 401 |
+
}
|
| 402 |
+
},
|
| 403 |
+
{
|
| 404 |
+
"@type": "cr:Field",
|
| 405 |
+
"name": "image_t2",
|
| 406 |
+
"description": "Relative path to the second image for temporal examples, if present.",
|
| 407 |
+
"dataType": "sc:Text",
|
| 408 |
+
"source": {
|
| 409 |
+
"fileObject": {
|
| 410 |
+
"@id": "metadata_jsonl"
|
| 411 |
+
},
|
| 412 |
+
"extract": {
|
| 413 |
+
"column": "image_t2"
|
| 414 |
+
}
|
| 415 |
+
}
|
| 416 |
+
},
|
| 417 |
+
{
|
| 418 |
+
"@type": "cr:Field",
|
| 419 |
+
"name": "scene_graph",
|
| 420 |
+
"description": "Relative path to the compact scene graph after archive extraction.",
|
| 421 |
+
"dataType": "sc:Text",
|
| 422 |
+
"source": {
|
| 423 |
+
"fileObject": {
|
| 424 |
+
"@id": "metadata_jsonl"
|
| 425 |
+
},
|
| 426 |
+
"extract": {
|
| 427 |
+
"column": "scene_graph"
|
| 428 |
+
}
|
| 429 |
+
}
|
| 430 |
+
},
|
| 431 |
+
{
|
| 432 |
+
"@type": "cr:Field",
|
| 433 |
+
"name": "record_json",
|
| 434 |
+
"description": "Relative path to the subcategory JSON file.",
|
| 435 |
+
"dataType": "sc:Text",
|
| 436 |
+
"source": {
|
| 437 |
+
"fileObject": {
|
| 438 |
+
"@id": "metadata_jsonl"
|
| 439 |
+
},
|
| 440 |
+
"extract": {
|
| 441 |
+
"column": "record_json"
|
| 442 |
+
}
|
| 443 |
+
}
|
| 444 |
+
},
|
| 445 |
+
{
|
| 446 |
+
"@type": "cr:Field",
|
| 447 |
+
"name": "support_json",
|
| 448 |
+
"description": "Grounded support evidence encoded as JSON.",
|
| 449 |
+
"dataType": "sc:Text",
|
| 450 |
+
"source": {
|
| 451 |
+
"fileObject": {
|
| 452 |
+
"@id": "metadata_jsonl"
|
| 453 |
+
},
|
| 454 |
+
"extract": {
|
| 455 |
+
"column": "support_json"
|
| 456 |
+
}
|
| 457 |
+
}
|
| 458 |
+
},
|
| 459 |
+
{
|
| 460 |
+
"@type": "cr:Field",
|
| 461 |
+
"name": "program_json",
|
| 462 |
+
"description": "Executable reasoning program encoded as JSON.",
|
| 463 |
+
"dataType": "sc:Text",
|
| 464 |
+
"source": {
|
| 465 |
+
"fileObject": {
|
| 466 |
+
"@id": "metadata_jsonl"
|
| 467 |
+
},
|
| 468 |
+
"extract": {
|
| 469 |
+
"column": "program_json"
|
| 470 |
+
}
|
| 471 |
+
}
|
| 472 |
+
}
|
| 473 |
+
]
|
| 474 |
+
}
|
| 475 |
+
]
|
| 476 |
+
}
|
dataset_manifest.json
ADDED
|
@@ -0,0 +1,135 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"dataset_name": "RSFaith-Bench",
|
| 3 |
+
"release_version": "1.0.0",
|
| 4 |
+
"schema_version": "1.0.0",
|
| 5 |
+
"schema_flavor": "compact_verifiable_archived_assets",
|
| 6 |
+
"total_records": 13511,
|
| 7 |
+
"unique_scenes": 11702,
|
| 8 |
+
"copied_images": 16288,
|
| 9 |
+
"scene_graph_files": 12876,
|
| 10 |
+
"levels": {
|
| 11 |
+
"Perception": {
|
| 12 |
+
"object_presence": 969,
|
| 13 |
+
"object_counting": 1085,
|
| 14 |
+
"fine_grained_recognition": 1228,
|
| 15 |
+
"object_localization": 1298
|
| 16 |
+
},
|
| 17 |
+
"Relational reasoning": {
|
| 18 |
+
"directional": 1166,
|
| 19 |
+
"topological": 921,
|
| 20 |
+
"proximity": 987,
|
| 21 |
+
"projective_ordering": 944,
|
| 22 |
+
"aggregate_distribution": 866
|
| 23 |
+
},
|
| 24 |
+
"Temporal reasoning": {
|
| 25 |
+
"category_turnover": 1515,
|
| 26 |
+
"net_change": 1273,
|
| 27 |
+
"semantic_transition": 1259
|
| 28 |
+
}
|
| 29 |
+
},
|
| 30 |
+
"public_id_policy": "sequential_anonymous_ids",
|
| 31 |
+
"contains_original_dataset_names": false,
|
| 32 |
+
"asset_packaging": {
|
| 33 |
+
"type": "per_subcategory_tar_zst",
|
| 34 |
+
"archive_name": "assets.tar.zst",
|
| 35 |
+
"archives": [
|
| 36 |
+
{
|
| 37 |
+
"subcategory_dir": "Perception/fine_grained_recognition",
|
| 38 |
+
"archive": "Perception/fine_grained_recognition/assets.tar.zst",
|
| 39 |
+
"images": 1228,
|
| 40 |
+
"scene_graphs": 1228,
|
| 41 |
+
"archive_bytes": 1417644607,
|
| 42 |
+
"source_asset_bytes": 2104835424
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"subcategory_dir": "Perception/object_counting",
|
| 46 |
+
"archive": "Perception/object_counting/assets.tar.zst",
|
| 47 |
+
"images": 1085,
|
| 48 |
+
"scene_graphs": 1085,
|
| 49 |
+
"archive_bytes": 2131892267,
|
| 50 |
+
"source_asset_bytes": 2585156509
|
| 51 |
+
},
|
| 52 |
+
{
|
| 53 |
+
"subcategory_dir": "Perception/object_localization",
|
| 54 |
+
"archive": "Perception/object_localization/assets.tar.zst",
|
| 55 |
+
"images": 1298,
|
| 56 |
+
"scene_graphs": 1298,
|
| 57 |
+
"archive_bytes": 2495263000,
|
| 58 |
+
"source_asset_bytes": 3098535552
|
| 59 |
+
},
|
| 60 |
+
{
|
| 61 |
+
"subcategory_dir": "Perception/object_presence",
|
| 62 |
+
"archive": "Perception/object_presence/assets.tar.zst",
|
| 63 |
+
"images": 969,
|
| 64 |
+
"scene_graphs": 969,
|
| 65 |
+
"archive_bytes": 1554646842,
|
| 66 |
+
"source_asset_bytes": 1938570476
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"subcategory_dir": "Relational reasoning/aggregate_distribution",
|
| 70 |
+
"archive": "Relational reasoning/aggregate_distribution/assets.tar.zst",
|
| 71 |
+
"images": 866,
|
| 72 |
+
"scene_graphs": 866,
|
| 73 |
+
"archive_bytes": 5081368574,
|
| 74 |
+
"source_asset_bytes": 6921795889
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"subcategory_dir": "Relational reasoning/directional",
|
| 78 |
+
"archive": "Relational reasoning/directional/assets.tar.zst",
|
| 79 |
+
"images": 1166,
|
| 80 |
+
"scene_graphs": 1166,
|
| 81 |
+
"archive_bytes": 1833785421,
|
| 82 |
+
"source_asset_bytes": 2418984946
|
| 83 |
+
},
|
| 84 |
+
{
|
| 85 |
+
"subcategory_dir": "Relational reasoning/projective_ordering",
|
| 86 |
+
"archive": "Relational reasoning/projective_ordering/assets.tar.zst",
|
| 87 |
+
"images": 944,
|
| 88 |
+
"scene_graphs": 944,
|
| 89 |
+
"archive_bytes": 1473705066,
|
| 90 |
+
"source_asset_bytes": 2320686389
|
| 91 |
+
},
|
| 92 |
+
{
|
| 93 |
+
"subcategory_dir": "Relational reasoning/proximity",
|
| 94 |
+
"archive": "Relational reasoning/proximity/assets.tar.zst",
|
| 95 |
+
"images": 987,
|
| 96 |
+
"scene_graphs": 987,
|
| 97 |
+
"archive_bytes": 1490183220,
|
| 98 |
+
"source_asset_bytes": 1789255608
|
| 99 |
+
},
|
| 100 |
+
{
|
| 101 |
+
"subcategory_dir": "Relational reasoning/topological",
|
| 102 |
+
"archive": "Relational reasoning/topological/assets.tar.zst",
|
| 103 |
+
"images": 921,
|
| 104 |
+
"scene_graphs": 921,
|
| 105 |
+
"archive_bytes": 1238362287,
|
| 106 |
+
"source_asset_bytes": 1359977655
|
| 107 |
+
},
|
| 108 |
+
{
|
| 109 |
+
"subcategory_dir": "Temporal reasoning/category_turnover",
|
| 110 |
+
"archive": "Temporal reasoning/category_turnover/assets.tar.zst",
|
| 111 |
+
"images": 2740,
|
| 112 |
+
"scene_graphs": 1370,
|
| 113 |
+
"archive_bytes": 1103133575,
|
| 114 |
+
"source_asset_bytes": 1123776543
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"subcategory_dir": "Temporal reasoning/net_change",
|
| 118 |
+
"archive": "Temporal reasoning/net_change/assets.tar.zst",
|
| 119 |
+
"images": 2424,
|
| 120 |
+
"scene_graphs": 1212,
|
| 121 |
+
"archive_bytes": 2299874948,
|
| 122 |
+
"source_asset_bytes": 2345848846
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"subcategory_dir": "Temporal reasoning/semantic_transition",
|
| 126 |
+
"archive": "Temporal reasoning/semantic_transition/assets.tar.zst",
|
| 127 |
+
"images": 1660,
|
| 128 |
+
"scene_graphs": 830,
|
| 129 |
+
"archive_bytes": 689231000,
|
| 130 |
+
"source_asset_bytes": 708993429
|
| 131 |
+
}
|
| 132 |
+
],
|
| 133 |
+
"note": "Each archive contains images/ and scene_graphs/ relative to the subcategory directory."
|
| 134 |
+
}
|
| 135 |
+
}
|
metadata.jsonl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9a1fbd8044fe473de7928d6cc5a37632a6ca7e9353c94ee23fa3427cdb13c7bb
|
| 3 |
+
size 42232446
|