Dataset Viewer
Auto-converted to Parquet Duplicate
Search is not available for this dataset
id
int64
0
1.14M
emb
listlengths
1.02k
1.02k
0
[ -0.0022642090916633606, -0.06003301590681076, -0.012406417168676853, 0.030092809349298477, -0.00928573776036501, -0.0703793317079544, -0.010247120633721352, -0.033388979732990265, -0.027437562122941017, 0.03177141398191452, 0.01088804192841053, -0.0033457649406045675, 0.014863284304738045, ...
1
[ 0.0744967982172966, -0.010944195091724396, 0.028706585988402367, 0.038346629589796066, 0.052105046808719635, -0.018593693152070045, 0.020241040736436844, 0.01949363388121128, -0.008007948286831379, 0.030948812142014503, 0.018532678484916687, 0.039475370198488235, 0.014421932399272919, 0.05...
2
[ 0.026934873312711716, 0.01764119789004326, -0.038914408534765244, -0.035343438386917114, 0.008149134926497936, -0.04755187779664993, -0.023348644375801086, -0.01613040268421173, -0.004303475841879845, 0.020678047090768814, -0.04089827835559845, 0.0335732139647007, -0.024691572412848473, 0....
3
[ 0.05187099799513817, -0.0423816554248333, -0.022121455520391464, -0.024837054312229156, 0.03350256010890007, -0.05159638449549675, 0.01266262587159872, -0.06529643386602402, -0.03716404363512993, 0.015614695847034454, -0.02770521491765976, -0.051809974014759064, -0.001479848986491561, -0.0...
4
[ 0.00046515860594809055, 0.008644177578389645, -0.06799374520778656, 0.003959689289331436, 0.04516639932990074, 0.01980607770383358, 0.029754891991615295, 0.018280185759067535, 0.08801344782114029, -0.04321325942873955, 0.07598941773176193, 0.001795783988200128, -0.01979081891477108, 0.0448...
5
[ 0.027655381709337234, 0.025230007246136665, 0.037402644753456116, -0.06211401894688606, 0.002427282277494669, -0.021508047357201576, 0.022469043731689453, -0.04939223825931549, 0.0479583665728569, 0.06327331811189651, 0.03935514762997627, -0.016474248841404915, 0.03386373445391655, 0.00073...
6
[ 0.04636586457490921, -0.016131902113556862, 0.019489534199237823, 0.01237745676189661, -0.02032894268631935, -0.05906382203102112, 0.04749524965882301, -0.011301488615572453, 0.05219593644142151, -0.025777466595172882, 0.04664057865738869, -0.03311847522854805, 0.061780452728271484, -0.008...
7
[ -0.028809092938899994, -0.00968186929821968, 0.08105609565973282, -0.022812284529209137, -0.0016889865510165691, 0.035645149648189545, 0.0035401005297899246, -0.05682471767067909, 0.007751599419862032, 0.01045245211571455, -0.0008440164383500814, 0.03512634336948395, 0.03369199112057686, 0...
8
[ 0.021394047886133194, 0.06155748292803764, 0.04141472652554512, 0.036592673510313034, 0.0009637430775910616, -0.035707611590623856, 0.0550568662583828, 0.042482901364564896, 0.07648143172264099, -0.019303474575281143, 0.022660600021481514, -0.008949792012572289, 0.04071278125047684, -0.020...
9
[-0.022997364401817322,0.036807991564273834,0.004890945740044117,0.028384272009134293,0.043858278542(...TRUNCATED)
End of preview. Expand in Data Studio

Arctic-Wiki-English-5M

VDBBench-compatible vector benchmark case published as a Hugging Face dataset repository.

What’s in this repo

  • vdbbench/: a Parquet-only folder intended to be downloaded as-is for VDBBench.
  • The Dataset Viewer indexes only train/test (and optionally shuffled), and intentionally ignores neighbors.parquet.

vdbbench/ file contract (VDBBench)

  • Train vectors: train.parquet or train-*-of-*.parquet
  • Query vectors: test.parquet or test-*-of-*.parquet
  • Ground truth: neighbors.parquet with columns id, neighbors_id (list[int])
  • Optional shuffled train: shuffle_train.parquet or shuffle_train-*-of-*.parquet

Important: do not add non-parquet files inside vdbbench/ (VDBBench requires the folder to be parquet-only).

Download for VDBBench

Download only the parquet files and point VDBBench at the vdbbench/ folder.

hf download "AI71ai/Arctic-Wiki-English-5M" \
  --repo-type dataset \
  --include "vdbbench/*.parquet" \
  --local-dir ./my_case

Then set VDBBench /customFolder Path to:

  • ./my_case/vdbbench

How this case was created

  • Source: wikimedia/wikipedia at revision b04c8d1ceb2f5cd4588862100d08de323dccfbaa (pinned for reproducibility when set).
  • Embedding model: Snowflake/snowflake-arctic-embed-l-v2.0 (dimension 1024).
  • Languages: en
  • Train size: 5,000,000 vectors
  • Test size: 1,000 query vectors
  • Phase 2 seed: 42

Size filtering (chosen to embed well)

We filter Wikipedia articles by character length of the article text to avoid pages that are too short (low-signal) or too long (risking truncation / poor embedding behavior).

Approx token ranges below are computed from the Snowflake Arctic tokenizer’s observed chars/token ratios (using the per-language ~8k-token max lengths we chose).

Language Text length filter (chars) Approx token range
en 400–32,605 chars ~98–~8,000 tokens

Test set (queries)

  • Queries are derived from Wikipedia titles of documents that appear in the training stream.
  • Phase 2 Stage 1 selects test_size titles using reservoir sampling (bounded memory) while writing train.
  • Phase 2 Stage 2 embeds those titles to produce test*.parquet (query vectors).

Shuffled train (optional)

  • Some vector DBs are sensitive to ingestion order; this benchmark optionally provides a shuffled training set to evaluate that behavior.
  • This repo includes shuffle_train*.parquet for VDBBench “Use Shuffled Data” mode.

Ground truth (neighbors.parquet)

  • neighbors.parquet contains the exact top-400 nearest train IDs for each test query ID.
  • Similarity is computed with cosine (implemented as L2-normalize + inner product).

Provenance: ID map sidecars (maps/)

This repo also includes optional ID mapping sidecars under maps/ that let you trace VDBBench IDs back to the original Wikipedia/source IDs used during embedding.

  • maps/ is intentionally not part of any Dataset Viewer split.
  • Files are Parquet (same sharding convention as vectors when applicable).

Files

  • train_id_map*.parquet: train VDBBench id → source id mapping
  • test_id_map*.parquet: test query id → train_id + source id mapping
  • shuffle_train_id_map*.parquet: shuffled-train VDBBench id → source id mapping

Schemas

  • train_id_map: id, source_id
  • test_id_map: id, train_id, source_id
  • shuffle_train_id_map: id, source_id

Notes:

  • source_id refers to the upstream article ID from the embedded Wikipedia dump (see the pipeline repo for details).
  • For mixed-language cases, the language column indicates which language shard the source came from.

Design decisions

  • Per-case repos: keeps each benchmark case independently versionable and easily downloadable without filtering huge multi-case repos.
  • Parquet-only vdbbench/ folder: matches VDBBench’s strict folder selection rule and enables “download folder → run benchmark”.
  • Neighbors kept next to vectors, but not indexed: neighbors.parquet is required by VDBBench, but it’s not useful to display as a dataset split in the viewer.
  • Prefer single-file test.parquet: VDBBench duplicates test queries across processes; keeping queries small and single-file reduces overhead.
  • Embedding model choice: Snowflake Arctic Embed L v2.0 is high on the MTEB-R leaderboard (see https://huggingface.co/spaces/mteb/leaderboard) and is a strong fit for evaluating compressed representations (including binarized/quantized vectors).

Usage with datasets

from datasets import load_dataset

ds = load_dataset("AI71ai/Arctic-Wiki-English-5M")
print(ds)

Note: neighbors.parquet is intentionally not part of any split.

Downloads last month
664