Maris Basha commited on
Commit
0db24f3
·
1 Parent(s): 8d59b89

De-anonymize and polish README: add ICML 2026 paper info, arXiv DOI, cross-links to vocsim/* repos

Browse files
Files changed (1) hide show
  1. README.md +51 -30
README.md CHANGED
@@ -29,10 +29,10 @@ dataset_info:
29
  dtype: string
30
  splits:
31
  - name: train
32
- num_bytes: 5452179735 # ~5.45 GB
33
  num_examples: 114641
34
- download_size: 5500616162 # ~5.5 GB
35
- dataset_size: 5452179735 # ~5.45 GB
36
  configs:
37
  - config_name: default
38
  data_files:
@@ -40,47 +40,68 @@ configs:
40
  path: data/train-*
41
  ---
42
 
43
- # VocSim: A Training-Free Benchmark for Content Identity in Single-Source Audio Embeddings
44
- [![GitHub Repository](https://img.shields.io/badge/GitHub-vocsim-black?logo=github)](https://github.com/anonymoussubmission0000/vocsim)
45
- [![Paper: arXiv](https://img.shields.io/badge/Paper-arXiv%20(TBD)-red.svg)](#)
 
46
  [![License: CC BY 4.0](https://img.shields.io/badge/License-CC%20BY%204.0-blue.svg)](https://creativecommons.org/licenses/by/4.0/)
47
 
48
- **VocSim** evaluates how well neural audio embeddings generalize for **zero-shot audio similarity**. It tests recognizing fine-grained acoustic similarity without specific similarity training.
49
 
50
- ---
51
 
52
- ## Key Features
53
 
54
- * **Diverse Sources:** Human speech (phones, words, utterances), birdsong, otter calls, environmental sounds.
55
- * **Varied Conditions:** Spans clean to noisy recordings, short (<100ms) to long durations, few to many classes per subset.
56
- * **Standardized:** All audio is 16kHz mono.
 
 
57
 
58
- ---
59
 
60
- ## Data Format
61
 
62
  ```python
63
  {
64
- 'audio': {'array': array([...], dtype=float32), 'sampling_rate': 16000},
65
- 'subset': 'HW1', # Origin identifier
66
- 'speaker': 'spk_id', # Speaker/Animal/Source ID or 'N/A'
67
- 'label': 'hello' # Ground truth class for similarity
68
  }
69
  ```
70
- Train split: 114,641 public examples from 15 subsets for evaluation.
71
 
72
- Blind Test Sets: 4 additional subsets held out privately.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
73
 
74
  ## Citation
75
- ```bib
76
- @inproceedings{vocsim_authors_2025,
77
- title={VocSim: A Training-Free Benchmark for Content Identity in Single-Source Audio Embeddings},
78
- author={Anonymous Authors},
79
- booktitle={Conference/Journal},
80
- year={2025},
81
- url={[Link to paper upon DOI]}
 
82
  }
83
  ```
84
- ## License
85
-
86
- CC BY 4.0 - Creative Commons Attribution 4.0 International.
 
29
  dtype: string
30
  splits:
31
  - name: train
32
+ num_bytes: 5452179735
33
  num_examples: 114641
34
+ download_size: 5500616162
35
+ dataset_size: 5452179735
36
  configs:
37
  - config_name: default
38
  data_files:
 
40
  path: data/train-*
41
  ---
42
 
43
+ # VocSim Public Benchmark
44
+
45
+ [![GitHub](https://img.shields.io/badge/GitHub-vocsim%2Fbenchmark-black?logo=github)](https://github.com/vocsim/benchmark)
46
+ [![Leaderboard](https://img.shields.io/badge/%F0%9F%A4%97%20Leaderboard-vocsim%2FVocSim-green)](https://huggingface.co/spaces/vocsim/VocSim)
47
  [![License: CC BY 4.0](https://img.shields.io/badge/License-CC%20BY%204.0-blue.svg)](https://creativecommons.org/licenses/by/4.0/)
48
 
49
+ The public split of **VocSim**, a training-free benchmark for zero-shot content identity in single-source audio embeddings. VocSim probes the intrinsic geometric quality of frozen audio representations: do acoustically variable instances of the same content land near each other in embedding space, without any task-specific training?
50
 
51
+ > Basha, M., Zai, A. T., Stoll, S., & Hahnloser, R. H. R. *VocSim: A Training-free Benchmark for Zero-shot Content Identity in Single-source Audio.* ICML 2026. [arXiv:2512.10120](https://doi.org/10.48550/arXiv.2512.10120)
52
 
53
+ ## What's here
54
 
55
+ - **114,641 clips** across **15 public subsets**, drawn from 19 source corpora.
56
+ - Domains: human speech (phones, words, utterances), animal vocalizations (birdsong, otter calls), environmental sounds.
57
+ - Conditions: clean to noisy, sub-100ms to multi-second, few to thousands of classes per subset.
58
+ - All audio standardized to **16 kHz mono**.
59
+ - Single-source only — no overlapping speakers or simultaneous sources — so evaluation isolates content representation from source separation.
60
 
61
+ Four additional **blind out-of-distribution subsets** (low-resource speech in Shipibo-Conibo and Chintang) are held out for server-side evaluation via the [leaderboard](https://huggingface.co/spaces/vocsim/VocSim).
62
 
63
+ ## Schema
64
 
65
  ```python
66
  {
67
+ "audio": {"array": np.ndarray, "sampling_rate": 16000},
68
+ "subset": "HW1", # source-corpus tag (see paper for the full list)
69
+ "speaker": "spk_042", # speaker / animal / source ID, or "N/A"
70
+ "label": "hello", # ground-truth class for similarity
71
  }
72
  ```
 
73
 
74
+ ## Quick start
75
+
76
+ ```python
77
+ from datasets import load_dataset
78
+
79
+ ds = load_dataset("vocsim/public", split="train")
80
+ print(ds[0])
81
+ ```
82
+
83
+ For end-to-end evaluation (feature extraction, distance computation, P@k / GSR), use the reference pipeline at [github.com/vocsim/benchmark](https://github.com/vocsim/benchmark).
84
+
85
+ ## Companion datasets
86
+
87
+ | Dataset | Purpose |
88
+ |---|---|
89
+ | [`vocsim/avian-perception-benchmark`](https://huggingface.co/datasets/vocsim/avian-perception-benchmark) | Alignment of embeddings with zebra-finch perceptual judgments |
90
+ | [`vocsim/mouse-strain-classification-benchmark`](https://huggingface.co/datasets/vocsim/mouse-strain-classification-benchmark) | C57 vs DBA USV classification |
91
+ | [`vocsim/mouse-identity-classification-benchmark`](https://huggingface.co/datasets/vocsim/mouse-identity-classification-benchmark) | Individual-mouse identification from USVs |
92
+
93
+ ## Licensing
94
+
95
+ Aggregation and metadata are released under CC BY 4.0. Each source corpus retains its original license; see Appendix A.1.1 of the paper for a per-source breakdown.
96
 
97
  ## Citation
98
+
99
+ ```bibtex
100
+ @inproceedings{basha2026vocsim,
101
+ title = {VocSim: A Training-free Benchmark for Zero-shot Content Identity in Single-source Audio},
102
+ author = {Basha, Maris and Zai, Anja T. and Stoll, Sabine and Hahnloser, Richard H. R.},
103
+ booktitle = {Proceedings of the 43rd International Conference on Machine Learning (ICML)},
104
+ year = {2026},
105
+ doi = {10.48550/arXiv.2512.10120}
106
  }
107
  ```