nielsr HF Staff commited on
Commit
ed0d24c
·
verified ·
1 Parent(s): f099ac0

Add task category and link to paper and GitHub repository

Browse files

Hi! I'm Niels, part of the community science team at Hugging Face.

This PR improves the dataset card by:
- Adding the `image-classification` task category to the metadata.
- Linking the dataset to its official paper: [Scaling Continual Learning to 300+ Tasks with Bi-Level Routing Mixture-of-Experts](https://huggingface.co/papers/2602.03473).
- Providing a link to the official GitHub repository.
- Adding a brief description of the dataset's purpose for Class-Incremental Learning (CIL).

Files changed (1) hide show
  1. README.md +18 -3
README.md CHANGED
@@ -1,12 +1,27 @@
1
  ---
2
  license: apache-2.0
 
 
3
  ---
4
 
5
- - Before using the dataset, please visit: https://github.com/LMMMEng/CaRE.
6
 
7
- - If you find this dataset useful for your research, please cite:
8
 
9
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  @inproceedings{lou2026scaling,
11
  title={Scaling Continual Learning to 300+ Tasks with Bi-Level Routing Mixture-of-Experts},
12
  author={Lou, Meng and Fu, Yunxiang and Yu, Yizhou},
 
1
  ---
2
  license: apache-2.0
3
+ task_categories:
4
+ - image-classification
5
  ---
6
 
7
+ # OmniBenchmark-1K
8
 
9
+ OmniBenchmark-1K is a challenging benchmark for Class-Incremental Learning (CIL) designed to evaluate performance on very long task sequences, ranging from 100 to over 300 non-overlapping tasks.
10
 
11
+ The dataset was introduced in the paper [Scaling Continual Learning to 300+ Tasks with Bi-Level Routing Mixture-of-Experts](https://huggingface.co/papers/2602.03473).
12
+
13
+ - **GitHub:** [https://github.com/LMMMEng/CaRE](https://github.com/LMMMEng/CaRE)
14
+ - **Paper:** [https://huggingface.co/papers/2602.03473](https://huggingface.co/papers/2602.03473)
15
+
16
+ ## Description
17
+
18
+ OmniBenchmark-1K provides a large-scale evaluation protocol for comprehensively assessing CIL methods. While standard benchmarks often focus on 5-20 tasks, this dataset allows for performance evaluation on extremely long sequences, testing the stability and plasticity of models over time.
19
+
20
+ ## Citations
21
+
22
+ If you find this dataset useful for your research, please cite:
23
+
24
+ ```bibtex
25
  @inproceedings{lou2026scaling,
26
  title={Scaling Continual Learning to 300+ Tasks with Bi-Level Routing Mixture-of-Experts},
27
  author={Lou, Meng and Fu, Yunxiang and Yu, Yizhou},