| --- |
| license: apache-2.0 |
| task_categories: |
| - image-classification |
| --- |
| |
| # OmniBenchmark-1K |
|
|
| OmniBenchmark-1K is a challenging benchmark for Class-Incremental Learning (CIL) designed to evaluate performance on very long task sequences, ranging from 100 to over 300 non-overlapping tasks. |
|
|
| The dataset was introduced in the paper [Scaling Continual Learning to 300+ Tasks with Bi-Level Routing Mixture-of-Experts](https://huggingface.co/papers/2602.03473). |
|
|
| - **GitHub:** [https://github.com/LMMMEng/CaRE](https://github.com/LMMMEng/CaRE) |
| - **Paper:** [https://huggingface.co/papers/2602.03473](https://huggingface.co/papers/2602.03473) |
|
|
| ## Description |
|
|
| OmniBenchmark-1K provides a large-scale evaluation protocol for comprehensively assessing CIL methods. While standard benchmarks often focus on 5-20 tasks, this dataset allows for performance evaluation on extremely long sequences, testing the stability and plasticity of models over time. |
|
|
| ## Citations |
|
|
| If you find this dataset useful for your research, please cite: |
|
|
| ```bibtex |
| @inproceedings{lou2026scaling, |
| title={Scaling Continual Learning to 300+ Tasks with Bi-Level Routing Mixture-of-Experts}, |
| author={Lou, Meng and Fu, Yunxiang and Yu, Yizhou}, |
| booktitle={International Conference on Machine Learning}, |
| year={2026}, |
| } |
| |
| @inproceedings{zhang2022benchmarking, |
| title={Benchmarking omni-vision representation through the lens of visual realms}, |
| author={Zhang, Yuanhan and Yin, Zhenfei and Shao, Jing and Liu, Ziwei}, |
| booktitle={European Conference on Computer Vision}, |
| year={2022}, |
| } |
| ``` |