Datasets:
dataset string | model_name string | model_links list | paper_title string | paper_date timestamp[ns] | paper_url string | code_links list | metrics string | table_metrics list | prompts string | paper_text string | compute_hours float64 | num_gpus int64 | reasoning string | trainable_single_gpu_8h string | verified string | modality string | paper_title_drop string | paper_date_drop string | code_links_drop string | num_gpus_drop int64 | dataset_link string | time_and_compute_verification string | link_to_colab_notebook string | run_possible string | notes string |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
PDBbind | BAPULM | [] | BAPULM: Binding Affinity Prediction using Language Models | 2024-11-06T00:00:00 | https://arxiv.org/abs/2411.04150v1 | [
"https://github.com/radh55sh/BAPULM"
] | {'RMSE': '0.898±0.0172'} | [
"RMSE"
] | Given the following paper and codebase:
Paper: BAPULM: Binding Affinity Prediction using Language Models
Codebase: https://github.com/radh55sh/BAPULM
Improve the BAPULM model on the PDBbind dataset. The result
should improve on the following metrics: {'RMSE': '0.898±0.0172'}. You must use only the code... | BAPULM: Binding Affinity Prediction using Language Models Radheesh Sharma Meda†and Amir Barati Farimani∗,‡,¶,†,§ †Department of Chemical Engineering, Carnegie Mellon University, 15213, USA ‡Department of Mechanical Engineering, Carnegie Mellon University, 15213, USA ¶Department of Biomedical Engineering, Carnegie Mello... | 1 | 1 | The model uses ProtT5-XL-U50 and MolFormer architectures, which are large transformer-based models. Given that training on an Nvidia RTX 2080 Ti took approximately 4 minutes, and assuming training occurs over a reduced dataset with 100k sequences, with a complex architecture having a moderate number of parameters, a si... | yes | Yes | Bioinformatics | BAPULM: Binding Affinity Prediction using Language Models | 2024-11-06 0:00:00 | https://github.com/radh55sh/BAPULM | 1 | https://huggingface.co/datasets/radh25sh/BAPULM/resolve/main/prottrans_molformer_tensor_dataset100k.json?download=true | 16sec * 60 epochs = 16 minutes | https://colab.research.google.com/drive/1--rNlCN01wUgN_6cTTuiVcusqSP9vGlG?usp=sharing | Yes | -- no pdbind dataset.Specifices to use prottrans malformer |
Digital twin-supported deep learning for fault diagnosis | DANN | [] | A domain adaptation neural network for digital twin-supported fault diagnosis | 2025-05-27T00:00:00 | https://arxiv.org/abs/2505.21046v1 | [
"https://github.com/JialingRichard/Digital-Twin-Fault-Diagnosis"
] | {'Accuray': '80.22'} | [
"Accuray"
] | Given the following paper and codebase:
Paper: A domain adaptation neural network for digital twin-supported fault diagnosis
Codebase: https://github.com/JialingRichard/Digital-Twin-Fault-Diagnosis
Improve the DANN model on the Digital twin-supported deep learning for fault diagnosis dataset. The result
... | A domain adaptation neural network for digital twin-supported fault diagnosis Zhenling Chen CentraleSupélec, Université Paris-Saclay, Gif-sur-Yvette, 91190, FranceHaiwei Fu CentraleSupélec, Université Paris-Saclay, Gif-sur-Yvette, 91190, France Zhiguo Zeng Chair on Risk and Resilience of Complex Systems, Laboratoie Gen... | 2 | 1 | The DANN model employs a CNN architecture with two convolutional layers. Given the specified batch size of 32 and 250 training epochs on a dataset with 3,600 samples (360 samples per class for 9 distinct labels, plus a significantly smaller test set of 90 samples), the total iterations required for training would be (3... | yes | Yes | Time Series | A domain adaptation neural network for digital twin-supported fault diagnosis | 2025-05-27T00:00:00.000Z | [https://github.com/JialingRichard/Digital-Twin-Fault-Diagnosis] | 1 | Included in Repo | 3 Hours | Copy of train_ai_pytorch_DANN.ipynb | Yes | It starts and runs successfully |
MNIST | GatedGCN+ | [] | Unlocking the Potential of Classic GNNs for Graph-level Tasks: Simple Architectures Meet Excellence | 2025-02-13T00:00:00 | https://arxiv.org/abs/2502.09263v1 | [
"https://github.com/LUOyk1999/GNNPlus"
] | {'Accuracy': '98.712 ± 0.137'} | [
"Accuracy"
] | Given the following paper and codebase:
Paper: Unlocking the Potential of Classic GNNs for Graph-level Tasks: Simple Architectures Meet Excellence
Codebase: https://github.com/LUOyk1999/GNNPlus
Improve the GatedGCN+ model on the MNIST dataset. The result
should improve on the following metrics: {'Accur... | Unlocking the Potential of Classic GNNs for Graph-level Tasks: Simple Architectures Meet Excellence Yuankai Luo1 2Lei Shi*1Xiao-Ming Wu*2 Abstract Message-passing Graph Neural Networks (GNNs) are often criticized for their limited expres- siveness, issues like over-smoothing and over- squashing, and challenges in captu... | 4 | 1 | The GNN models (GCN, GIN, and GatedGCN) enhanced with GNN+ have approximately 500K parameters each, which is moderate for graph neural networks. The datasets used involve a variety of sizes, but the mentioned ones have a maximum of around 500K graphs (like the OGB datasets). Given the average training time of these mod... | yes | Yes | Graph | Unlocking the Potential of Classic GNNs for Graph-level Tasks: Simple Architectures Meet Excellence | 2025-02-13T00:00:00.000Z | [https://github.com/LUOyk1999/GNNPlus] | 1 | https://data.pyg.org/datasets/benchmarking-gnns/MNIST_v2.zip | 9 hour approx - ( 200 epochs * avg 157.2 sec) | https://drive.google.com/file/d/1Y7jMNhNybbdgrUJa_MxcOrbwpJNkDPav/view?usp=sharing | Yes | null |
ogbg-molhiv | GatedGCN+ | [] | "Unlocking the Potential of Classic GNNs for Graph-level Tasks: Simple Architectures Meet Excellence(...TRUNCATED) | 2025-02-13T00:00:00 | https://arxiv.org/abs/2502.09263v1 | [
"https://github.com/LUOyk1999/GNNPlus"
] | "{'Test ROC-AUC': '0.8040 ± 0.0164', 'Validation ROC-AUC': '0.8329 ± 0.0158', 'Number of params': (...TRUNCATED) | [
"Test ROC-AUC",
"Ext. data",
"Validation ROC-AUC",
"Number of params"
] | "Given the following paper and codebase:\n Paper: Unlocking the Potential of Classic GNNs for Gra(...TRUNCATED) | "Unlocking the Potential of Classic GNNs for Graph-level Tasks: Simple Architectures Meet Excellence(...TRUNCATED) | 4 | 1 | "The paper describes training across 14 well-known graph-level datasets with a mean parameter count (...TRUNCATED) | yes | Yes | Graph | "Unlocking the Potential of Classic GNNs for Graph-level Tasks: Simple Architectures Meet Excellence(...TRUNCATED) | 2025-02-13T00:00:00.000Z | [https://github.com/LUOyk1999/GNNPlus] | 1 | http://snap.stanford.edu/ogb/data/graphproppred/csv_mol_download/hiv.zip | approx 40 min - ( 100 epochs * 22.8s) | https://drive.google.com/file/d/1Y7jMNhNybbdgrUJa_MxcOrbwpJNkDPav/view?usp=sharing | Yes | null |
Fashion-MNIST | Continued fraction of straight lines | [] | Real-valued continued fraction of straight lines | 2024-12-16T00:00:00 | https://arxiv.org/abs/2412.16191v1 | ["https://github.com/grasshopper14/Continued-fraction-of-straight-lines/blob/main/continued_fraction(...TRUNCATED) | {'Accuracy': '84.12', 'Trainable Parameters': '7870', 'NMI': '74.4'} | [
"Percentage error",
"Accuracy",
"Trainable Parameters",
"NMI",
"Power consumption"
] | "Given the following paper and codebase:\n Paper: Real-valued continued fraction of straight line(...TRUNCATED) | "Real-valued continued fraction of straight lines Vijay Prakash S Alappuzha, Kerala, India. prakash.(...TRUNCATED) | 4 | 1 | "The model is trained on the Fashion-MNIST dataset, which consists of 60,000 training images and 10,(...TRUNCATED) | yes | Yes | CV | Real-valued continued fraction of straight lines | 2024-12-16T00:00:00.000Z | [https://github.com/grasshopper14/Continued-fraction-of-straight-lines/blob/main/continued_fraction_reg.py] | 1 | https://github.com/zalandoresearch/fashion-mnist | 20 min | https://colab.research.google.com/drive/1LNMCRLMIWN5U_9WDeRxYmcbnAgaNadSd?usp=sharing | Yes | Yes Everythng is running successfully |
Traffic | GLinear | [] | "Bridging Simplicity and Sophistication using GLinear: A Novel Architecture for Enhanced Time Series(...TRUNCATED) | 2025-01-02T00:00:00 | https://arxiv.org/abs/2501.01087v3 | [
"https://github.com/t-rizvi/GLinear"
] | {'MSE ': '0.3222'} | [
"MSE "
] | "Given the following paper and codebase:\n Paper: Bridging Simplicity and Sophistication using GL(...TRUNCATED) | "IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE JOURNAL, 2025 1 Bridging Simplic(...TRUNCATED) | 4 | 1 | "The GLinear model, being a simplified architecture without complex components like Transformers, sh(...TRUNCATED) | yes | Yes | Time Series | "Bridging Simplicity and Sophistication using GLinear: A Novel Architecture for Enhanced Time Series(...TRUNCATED) | 2025-01-02 0:00:00 | https://github.com/t-rizvi/GLinear | 1 | Inside the repo in dataset folder | 193 sec * 4 = 12.9 minutes | https://colab.research.google.com/drive/1sI72VSxjN4cyQR7UrueWfBXwoFi9Y9Qr?usp=sharing | Yes | "-- Training on all data set is included inside the scripts/EXP-LookBackWindow_\\&_LongForecasting/(...TRUNCATED) |
BTAD | URD | [] | Unlocking the Potential of Reverse Distillation for Anomaly Detection | 2024-12-10T00:00:00 | https://arxiv.org/abs/2412.07579v1 | [
"https://github.com/hito2448/urd"
] | "{'Segmentation AUROC': '98.1', 'Detection AUROC': '93.9', 'Segmentation AUPRO': '78.5', 'Segmentati(...TRUNCATED) | [
"Detection AUROC",
"Segmentation AUROC",
"Segmentation AP",
"Segmentation AUPRO"
] | "Given the following paper and codebase:\n Paper: Unlocking the Potential of Reverse Distillation(...TRUNCATED) | "Unlocking the Potential of Reverse Distillation for Anomaly Detection Xinyue Liu1, Jianyuan Wang2*,(...TRUNCATED) | 4 | 1 | "The proposed method utilizes a WideResNet50 architecture as a teacher network which typically has a(...TRUNCATED) | yes | Yes | CV | Unlocking the Potential of Reverse Distillation for Anomaly Detection | 2024-12-10 0:00:00 | https://github.com/hito2448/urd | 1 | https://www.mydrive.ch/shares/38536/3830184030e49fe74747669442f0f282/download/420938113-1629952094/mvtec_anomaly_detection.tar.xz; https://www.robots.ox.ac.uk/~vgg/data/dtd/download/dtd-r1.0.1.tar.gz | 8 hours for one folder. There are 11 folders. | https://drive.google.com/file/d/1OLbo3FifM1a7-wbCtfpjZrZLr0K5bS87/view?usp=sharing | Yes | -- Just need to change the num_workers in train.py according to system |
York Urban Dataset | DT-LSD | [] | DT-LSD: Deformable Transformer-based Line Segment Detection | 2024-11-20T00:00:00 | https://arxiv.org/abs/2411.13005v1 | [
"https://github.com/SebastianJanampa/DT-LSD"
] | {'sAP5': '30.2', 'sAP10': '33.2', 'sAP15': '35.1'} | [
"sAP5",
"sAP10",
"sAP15",
"FH"
] | "Given the following paper and codebase:\n Paper: DT-LSD: Deformable Transformer-based Line Segme(...TRUNCATED) | "DT-LSD: Deformable Transformer-based Line Segment Detection Sebastian Janampa The University of New(...TRUNCATED) | 4 | 1 | "The proposed DT-LSD model has a relatively small batch size of 2 and uses a single Nvidia RTX A5500(...TRUNCATED) | yes | Yes | CV | DT-LSD: Deformable Transformer-based Line Segment Detection | 2024-11-20 0:00:00 | https://github.com/SebastianJanampa/DT-LSD | 1 | script to download is provided in colab file. | uses cpu to trainf or some reason 8hr per epoch | https://colab.research.google.com/drive/1XPiW-hDq6q8HNZ4yVP0oAn-3a1_ay5rG?usp=sharing | Yes | -- Trains but uses cpu for some reason |
UCR Anomaly Archive | KAN | [] | KAN-AD: Time Series Anomaly Detection with Kolmogorov-Arnold Networks | 2024-11-01T00:00:00 | https://arxiv.org/abs/2411.00278v1 | [
"https://github.com/issaccv/KAN-AD"
] | {'AUC ROC ': '0.7489'} | [
"Average F1",
"AUC ROC "
] | "Given the following paper and codebase:\n Paper: KAN-AD: Time Series Anomaly Detection with Kolm(...TRUNCATED) | "KAN-AD: Time Series Anomaly Detection with Kolmogorov–Arnold Networks Quan Zhou*, Changhua Pei, H(...TRUNCATED) | 4 | 1 | "The KAN-AD model is based on a novel architecture that leverages Fourier series for anomaly detecti(...TRUNCATED) | yes | Yes | Time Series | KAN-AD: Time Series Anomaly Detection with Kolmogorov-Arnold Networks | 2024-11-01 0:00:00 | https://github.com/issaccv/KAN-AD | 1 | Downloaded when running prepeare_env.sh from repository & uses UTS dataset, https://github.com/CSTCloudOps/datasets | "There are 5 folders. May take around 2 hours or more no idea as time was not specified and traing w(...TRUNCATED) | https://colab.research.google.com/drive/1sE1mKwy3n9yameE-JG27Oa_HI-q8lFn9?usp=sharing | Yes | "-- After the installation of environment.sh. I changed a line of code to run matplot lib on colab (...TRUNCATED) |
Chameleon | CoED | [] | Improving Graph Neural Networks by Learning Continuous Edge Directions | 2024-10-18T00:00:00 | https://arxiv.org/abs/2410.14109v1 | [
"https://github.com/hormoz-lab/coed-gnn"
] | {'Accuracy': '79.69±1.35'} | [
"Accuracy"
] | "Given the following paper and codebase:\n Paper: Improving Graph Neural Networks by Learning Con(...TRUNCATED) | "Preprint IMPROVING GRAPH NEURAL NETWORKS BY LEARN - INGCONTINUOUS EDGE DIRECTIONS Seong Ho Pahng1, (...TRUNCATED) | 4 | 1 | "The proposed CoED GNN is a graph neural network architecture that utilizes a complex-valued Laplaci(...TRUNCATED) | yes | Yes | Graph | Improving Graph Neural Networks by Learning Continuous Edge Directions | 2024-10-18 0:00:00 | https://github.com/hormoz-lab/coed-gnn | 1 | specify on the classification.py and it handles itself | 2 min | https://colab.research.google.com/drive/1FiCFbVmQhjIqcCdViYynfEb9mWtJkB09?usp=sharing | Yes | -- I have put the best parameter with advice of "Gemini" Can change accordingly. |
End of preview. Expand in Data Studio
ARIA Repo Benchmark
The ARIA Repo Benchmark is part of the ARIA benchmark suite, a collection of closed-book benchmarks probing the ML knowledge that frontier models have internalized during training. This dataset contains 58 curated research paper implementations with metadata for evaluating whether ML experiments described in papers can be reproduced.
Dataset Summary
- Size: 58 entries
- Coverage: Computer Vision, NLP, Time Series, Graph, Bioinformatics
- Purpose: Evaluate AI agents on their ability to locate, understand, and reproduce ML research experiments
Each entry links a research paper to its code repository, dataset, metrics, and compute requirements, along with verification of whether the experiment is reproducible on constrained hardware.
Dataset Structure
Key Fields
| Field | Type | Description |
|---|---|---|
paper_title |
string | Title of the research paper |
paper_url |
string | ArXiv URL |
paper_date |
timestamp | Publication date |
paper_text |
string | Full paper text |
dataset |
string | Dataset used in the paper |
dataset_link |
string | Link to the dataset |
model_name |
string | Model name |
code_links |
list[string] | GitHub repository links |
metrics |
string | Performance metrics reported |
table_metrics |
list[string] | Detailed metrics from tables |
prompts |
string | Evaluation prompts |
modality |
string | Data modality (CV, NLP, Time Series, Graph, etc.) |
Compute & Reproducibility Fields
| Field | Type | Description |
|---|---|---|
compute_hours |
float64 | Estimated training compute hours |
num_gpus |
int64 | Number of GPUs required |
reasoning |
string | Reasoning about compute estimates |
trainable_single_gpu_8h |
string | Trainable on a single GPU in 8 hours |
verified |
string | Verification status |
time_and_compute_verification |
string | Compute verification notes |
link_to_colab_notebook |
string | Google Colab notebook link |
run_possible |
string | Whether the code runs successfully |
notes |
string | Additional notes |
Usage
from datasets import load_dataset
ds = load_dataset("AlgorithmicResearchGroup/aria-repo-benchmark", split="train")
for entry in ds:
print(f"{entry['paper_title']} - {entry['modality']} - Reproducible: {entry['run_possible']}")
Related Resources
Citation
@misc{aria_repo_benchmark,
title={ARIA Repo Benchmark},
author={Algorithmic Research Group},
year={2024},
publisher={Hugging Face},
url={https://huggingface.co/datasets/AlgorithmicResearchGroup/aria-repo-benchmark}
}
- Downloads last month
- 19