NQ320K-NCI / README.md
dqmis's picture
Upload folder using huggingface_hub
be6e04e verified
---
license: cc-by-sa-3.0
language:
- en
task_categories:
- text-retrieval
- question-answering
size_categories:
- 100K<n<1M
source_datasets:
- extended|natural_questions
tags:
- generative-retrieval
- dsi
- nci
- ripor
- semantic-id
configs:
- config_name: corpus
data_files:
- split: corpus
path: data/corpus.jsonl
- split: corpus_summary
path: data/corpus_summary.jsonl
- config_name: pairs
data_files:
- split: train
path: data/train.jsonl
- split: validation
path: data/valid.jsonl
dataset_info:
- config_name: corpus
features:
- name: docid
dtype: int64
- name: document
dtype: string
splits:
- name: corpus
num_examples: 109650
- name: corpus_summary
num_examples: 21119
- config_name: pairs
features:
- name: query
dtype: string
- name: docid
dtype: int64
- name: nq_id
dtype: string # IMPORTANT: string, not int – uint64 IDs overflow signed int64
- name: url
dtype: string
- name: title
dtype: string
- name: long_answer
dtype: string
- name: short_answer
dtype: string
splits:
- name: train
num_examples: 307373
- name: validation
num_examples: 7830
---
# NQ320K (NCI-style preprocessing)
A reproduction of the **NQ320K** corpus used by generative-retrieval papers
(DSI, NCI, GenRet, RIPOR, Ultron, LTRGR, …) built directly from the
Hugging Face `google-research-datasets/natural_questions` snapshot.
## At a glance
| Split | Rows |
|------------------|--------------|
| `corpus` | 109,650 |
| `corpus_summary` | 21,119 |
| `train` pairs | 307,373 |
| `validation` pairs | 7,830 |
* Train pairs with non-empty `long_answer`: **152,148 / 307,373** (49.5%)
* Train pairs with non-empty `short_answer`: **106,926 / 307,373** (34.8%)
* Date built: 2026-05-06
## Schema
### `corpus.jsonl`
```json
{"docid": 0, "document": "<NCI doc_tac string, ~5K-50K chars>"}
```
### `train.jsonl` / `valid.jsonl`
```json
{
"query": "when is the last episode of season 8 of the walking dead",
"docid": 0,
"nq_id": "5225754983651766092",
"url": "https://en.wikipedia.org//w/index.php?title=The_Walking_Dead_(season_8)&oldid=...",
"title": "The Walking Dead (season 8)",
"long_answer": "List of The Walking Dead episodes ...",
"short_answer": ""
}
```
`docid` is a stable integer that **joins** to `corpus.jsonl`. To materialise
a (query, document) pair:
```python
from datasets import load_dataset
corpus = load_dataset("<your-username>/NQ320K-NCI", "corpus", split="corpus")
pairs = load_dataset("<your-username>/NQ320K-NCI", "pairs", split="train")
doc_lookup = {r["docid"]: r["document"] for r in corpus}
for p in pairs:
document = doc_lookup[p["docid"]]
# ... feed (p["query"], document) to your model
```
## Preprocessing
Faithful port of the official NCI notebook
([Wang et al., NeurIPS 2022](https://arxiv.org/abs/2206.02743),
[Data_process/NQ_dataset/NQ_dataset_Process.ipynb](https://github.com/solidsea98/Neural-Corpus-Indexer-NCI/blob/main/Data_process/NQ_dataset/NQ_dataset_Process.ipynb)
in the released code). Each NQ row produces one record:
1. Reconstruct `document_text = " ".join(document.tokens.token)`
(HTML tags appear as their own tokens).
2. `title = document.title`
3. `abs = document_text[<P>+3 : </P>]`**HTML tags inside `<P>` are kept**, matching NCI.
4. `content = document_text[</P>+4 : second-to-last </Ul>]`,
then HTML stripped, `\n` deleted, multiple spaces collapsed.
5. `doc_tac = title + abs + content`**no separators**.
6. `long_answer` / `short_answer`: token-span slices from the
first annotator (`annotations[0]`), HTML stripped.
Documents are de-duplicated by their **BERT-uncased-tokenizer-normalised
title** (`tokenizer.tokenize(title) → convert_to_ids → decode`), exactly as
in NCI's released notebook. Concatenating `train + validation` and dropping
duplicates yields **109,650** unique documents (NCI reports 109,739; the
~80-doc delta comes from a slightly newer Hugging Face snapshot of NQ).
## Known formatting characteristics
These are **inherited from NCI's preprocessing** and intentional:
* **Token-joined whitespace**: `"AMC ,"` instead of `"AMC,"`. NCI's `doc_tac`
is built by `" ".join(tokens)`, leaving a space before every punctuation
mark. NCI's downstream BERT/T5 tokenizer absorbs these correctly; you may
want to detokenize when feeding into other encoders.
* **HTML tags inside `abs`**: e.g. `"<Table><Tr>…<P>The eighth season…</P>"`.
Only `content` has its tags stripped. This is the canonical NCI format.
* **Non-detokenized hyphenation**: `"post - apocalyptic"`, `"Spider - Man"`.
## Caveat: `nq_id` is a string
NQ's original `example_id` is a **uint64**, and roughly half of the IDs
exceed `2^63 = 9.22 × 10^18`. They fit unsigned but overflow signed int64.
`nq_id` is therefore stored as a **string**, exactly as Google publishes it.
**Do not auto-cast it to int64** — about 50% of the values would silently
wrap to negative numbers. If you load with pandas:
```python
import pandas as pd
df = pd.read_json("train.jsonl", lines=True, dtype={"nq_id": str})
```
If you load with `datasets`, the typed `dataset_info` in this card already
enforces `string`, so you don't need to do anything extra:
```python
from datasets import load_dataset
ds = load_dataset("<your-username>/NQ320K-NCI", "pairs")
print(ds["train"].features["nq_id"]) # Value(dtype='string', id=None)
```
## Corpus Summary
We additionally present a subset of corpus as a summarized text. We use `sshleifer/distilbart-cnn-12-6` model for the summarization task.
## License & attribution
This dataset is a derivative of the Natural Questions dataset by Google
([Kwiatkowski et al., TACL 2019](https://aclanthology.org/Q19-1026/)),
released under **CC BY-SA 3.0**. This derivative dataset is therefore also
released under **CC BY-SA 3.0** (ShareAlike).
The preprocessing recipe is from
[Neural Corpus Indexer (Wang et al., NeurIPS 2022)](https://arxiv.org/abs/2206.02743);
see their released
[notebook](https://github.com/solidsea98/Neural-Corpus-Indexer-NCI/blob/main/Data_process/NQ_dataset/NQ_dataset_Process.ipynb).
## Citation
If you use this dataset, please cite:
```bibtex
@article{kwiatkowski2019natural,
author = {Kwiatkowski, Tom and Palomaki, Jennimaria and Redfield, Olivia
and Collins, Michael and Parikh, Ankur and Alberti, Chris and
Epstein, Danielle and Polosukhin, Illia and Devlin, Jacob and
Lee, Kenton and Toutanova, Kristina and Jones, Llion and Kelcey,
Matthew and Chang, Ming-Wei and Dai, Andrew and Uszkoreit, Jakob
and Le, Quoc and Petrov, Slav},
title = {Natural Questions: a Benchmark for Question Answering Research},
journal = {Transactions of the Association for Computational Linguistics},
year = {2019}
}
@inproceedings{wang2022neural,
author = {Wang, Yujing and Hou, Yingyan and Wang, Haonan and Miao, Ziming
and Wu, Shibin and Sun, Hao and Chen, Qi and Xia, Yuqing and
Chi, Chengmin and Zhao, Guoshuai and Liu, Zheng and Xie, Xing
and Sun, Hao Allen and Deng, Weiwei and Zhang, Qi and Yang,
Mao},
title = {A Neural Corpus Indexer for Document Retrieval},
booktitle = {Advances in Neural Information Processing Systems (NeurIPS)},
year = {2022}
}
```