Add task category and link to paper and GitHub repository

#2
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +18 -3
README.md CHANGED
@@ -1,12 +1,27 @@
1
  ---
2
  license: apache-2.0
 
 
3
  ---
4
 
5
- - Before using the dataset, please visit: https://github.com/LMMMEng/CaRE.
6
 
7
- - If you find this dataset useful for your research, please cite:
8
 
9
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  @inproceedings{lou2026scaling,
11
  title={Scaling Continual Learning to 300+ Tasks with Bi-Level Routing Mixture-of-Experts},
12
  author={Lou, Meng and Fu, Yunxiang and Yu, Yizhou},
 
1
  ---
2
  license: apache-2.0
3
+ task_categories:
4
+ - image-classification
5
  ---
6
 
7
+ # OmniBenchmark-1K
8
 
9
+ OmniBenchmark-1K is a challenging benchmark for Class-Incremental Learning (CIL) designed to evaluate performance on very long task sequences, ranging from 100 to over 300 non-overlapping tasks.
10
 
11
+ The dataset was introduced in the paper [Scaling Continual Learning to 300+ Tasks with Bi-Level Routing Mixture-of-Experts](https://huggingface.co/papers/2602.03473).
12
+
13
+ - **GitHub:** [https://github.com/LMMMEng/CaRE](https://github.com/LMMMEng/CaRE)
14
+ - **Paper:** [https://huggingface.co/papers/2602.03473](https://huggingface.co/papers/2602.03473)
15
+
16
+ ## Description
17
+
18
+ OmniBenchmark-1K provides a large-scale evaluation protocol for comprehensively assessing CIL methods. While standard benchmarks often focus on 5-20 tasks, this dataset allows for performance evaluation on extremely long sequences, testing the stability and plasticity of models over time.
19
+
20
+ ## Citations
21
+
22
+ If you find this dataset useful for your research, please cite:
23
+
24
+ ```bibtex
25
  @inproceedings{lou2026scaling,
26
  title={Scaling Continual Learning to 300+ Tasks with Bi-Level Routing Mixture-of-Experts},
27
  author={Lou, Meng and Fu, Yunxiang and Yu, Yizhou},