id int64 0 1.14M | emb listlengths 1.02k 1.02k |
|---|---|
0 | [
-0.0022642090916633606,
-0.06003301590681076,
-0.012406417168676853,
0.030092809349298477,
-0.00928573776036501,
-0.0703793317079544,
-0.010247120633721352,
-0.033388979732990265,
-0.027437562122941017,
0.03177141398191452,
0.01088804192841053,
-0.0033457649406045675,
0.014863284304738045,
... |
1 | [
0.0744967982172966,
-0.010944195091724396,
0.028706585988402367,
0.038346629589796066,
0.052105046808719635,
-0.018593693152070045,
0.020241040736436844,
0.01949363388121128,
-0.008007948286831379,
0.030948812142014503,
0.018532678484916687,
0.039475370198488235,
0.014421932399272919,
0.05... |
2 | [
0.026934873312711716,
0.01764119789004326,
-0.038914408534765244,
-0.035343438386917114,
0.008149134926497936,
-0.04755187779664993,
-0.023348644375801086,
-0.01613040268421173,
-0.004303475841879845,
0.020678047090768814,
-0.04089827835559845,
0.0335732139647007,
-0.024691572412848473,
0.... |
3 | [
0.05187099799513817,
-0.0423816554248333,
-0.022121455520391464,
-0.024837054312229156,
0.03350256010890007,
-0.05159638449549675,
0.01266262587159872,
-0.06529643386602402,
-0.03716404363512993,
0.015614695847034454,
-0.02770521491765976,
-0.051809974014759064,
-0.001479848986491561,
-0.0... |
4 | [
0.00046515860594809055,
0.008644177578389645,
-0.06799374520778656,
0.003959689289331436,
0.04516639932990074,
0.01980607770383358,
0.029754891991615295,
0.018280185759067535,
0.08801344782114029,
-0.04321325942873955,
0.07598941773176193,
0.001795783988200128,
-0.01979081891477108,
0.0448... |
5 | [
0.027655381709337234,
0.025230007246136665,
0.037402644753456116,
-0.06211401894688606,
0.002427282277494669,
-0.021508047357201576,
0.022469043731689453,
-0.04939223825931549,
0.0479583665728569,
0.06327331811189651,
0.03935514762997627,
-0.016474248841404915,
0.03386373445391655,
0.00073... |
6 | [
0.04636586457490921,
-0.016131902113556862,
0.019489534199237823,
0.01237745676189661,
-0.02032894268631935,
-0.05906382203102112,
0.04749524965882301,
-0.011301488615572453,
0.05219593644142151,
-0.025777466595172882,
0.04664057865738869,
-0.03311847522854805,
0.061780452728271484,
-0.008... |
7 | [
-0.028809092938899994,
-0.00968186929821968,
0.08105609565973282,
-0.022812284529209137,
-0.0016889865510165691,
0.035645149648189545,
0.0035401005297899246,
-0.05682471767067909,
0.007751599419862032,
0.01045245211571455,
-0.0008440164383500814,
0.03512634336948395,
0.03369199112057686,
0... |
8 | [
0.021394047886133194,
0.06155748292803764,
0.04141472652554512,
0.036592673510313034,
0.0009637430775910616,
-0.035707611590623856,
0.0550568662583828,
0.042482901364564896,
0.07648143172264099,
-0.019303474575281143,
0.022660600021481514,
-0.008949792012572289,
0.04071278125047684,
-0.020... |
9 | [-0.022997364401817322,0.036807991564273834,0.004890945740044117,0.028384272009134293,0.043858278542(...TRUNCATED) |
Arctic-Wiki-English-5M
VDBBench-compatible vector benchmark case published as a Hugging Face dataset repository.
What’s in this repo
vdbbench/: a Parquet-only folder intended to be downloaded as-is for VDBBench.- The Dataset Viewer indexes only train/test (and optionally shuffled), and intentionally ignores
neighbors.parquet.
vdbbench/ file contract (VDBBench)
- Train vectors:
train.parquetortrain-*-of-*.parquet - Query vectors:
test.parquetortest-*-of-*.parquet - Ground truth:
neighbors.parquetwith columnsid,neighbors_id(list[int]) - Optional shuffled train:
shuffle_train.parquetorshuffle_train-*-of-*.parquet
Important: do not add non-parquet files inside vdbbench/ (VDBBench requires the folder to be parquet-only).
Download for VDBBench
Download only the parquet files and point VDBBench at the vdbbench/ folder.
hf download "AI71ai/Arctic-Wiki-English-5M" \
--repo-type dataset \
--include "vdbbench/*.parquet" \
--local-dir ./my_case
Then set VDBBench /custom → Folder Path to:
./my_case/vdbbench
How this case was created
- Source:
wikimedia/wikipediaat revisionb04c8d1ceb2f5cd4588862100d08de323dccfbaa(pinned for reproducibility when set). - Embedding model:
Snowflake/snowflake-arctic-embed-l-v2.0(dimension 1024). - Languages:
en - Train size:
5,000,000vectors - Test size:
1,000query vectors - Phase 2 seed:
42
Size filtering (chosen to embed well)
We filter Wikipedia articles by character length of the article text to avoid pages that are too short (low-signal) or too long (risking truncation / poor embedding behavior).
Approx token ranges below are computed from the Snowflake Arctic tokenizer’s observed chars/token ratios (using the per-language ~8k-token max lengths we chose).
| Language | Text length filter (chars) | Approx token range |
|---|---|---|
en |
400–32,605 chars |
~98–~8,000 tokens |
Test set (queries)
- Queries are derived from Wikipedia titles of documents that appear in the training stream.
- Phase 2 Stage 1 selects
test_sizetitles using reservoir sampling (bounded memory) while writing train. - Phase 2 Stage 2 embeds those titles to produce
test*.parquet(query vectors).
Shuffled train (optional)
- Some vector DBs are sensitive to ingestion order; this benchmark optionally provides a shuffled training set to evaluate that behavior.
- This repo includes
shuffle_train*.parquetfor VDBBench “Use Shuffled Data” mode.
Ground truth (neighbors.parquet)
neighbors.parquetcontains the exact top-400nearest train IDs for each test query ID.- Similarity is computed with cosine (implemented as L2-normalize + inner product).
Provenance: ID map sidecars (maps/)
This repo also includes optional ID mapping sidecars under maps/ that let you trace VDBBench IDs back to the original Wikipedia/source IDs used during embedding.
maps/is intentionally not part of any Dataset Viewer split.- Files are Parquet (same sharding convention as vectors when applicable).
Files
train_id_map*.parquet: train VDBBench id → source id mappingtest_id_map*.parquet: test query id → train_id + source id mappingshuffle_train_id_map*.parquet: shuffled-train VDBBench id → source id mapping
Schemas
train_id_map:id,source_idtest_id_map:id,train_id,source_idshuffle_train_id_map:id,source_id
Notes:
source_idrefers to the upstream article ID from the embedded Wikipedia dump (see the pipeline repo for details).- For mixed-language cases, the
languagecolumn indicates which language shard the source came from.
Design decisions
- Per-case repos: keeps each benchmark case independently versionable and easily downloadable without filtering huge multi-case repos.
- Parquet-only
vdbbench/folder: matches VDBBench’s strict folder selection rule and enables “download folder → run benchmark”. - Neighbors kept next to vectors, but not indexed:
neighbors.parquetis required by VDBBench, but it’s not useful to display as a dataset split in the viewer. - Prefer single-file
test.parquet: VDBBench duplicates test queries across processes; keeping queries small and single-file reduces overhead. - Embedding model choice: Snowflake Arctic Embed L v2.0 is high on the MTEB-R leaderboard (see
https://huggingface.co/spaces/mteb/leaderboard) and is a strong fit for evaluating compressed representations (including binarized/quantized vectors).
Usage with datasets
from datasets import load_dataset
ds = load_dataset("AI71ai/Arctic-Wiki-English-5M")
print(ds)
Note: neighbors.parquet is intentionally not part of any split.
- Downloads last month
- 664