Add metadata and improve model card
Browse filesHi there! I'm Niels, part of the community science team at Hugging Face. I've opened this PR to improve the model card for AROMA.
This PR adds:
- Metadata to the YAML header: `pipeline_tag: text-generation`, `library_name: transformers`, and `base_model: Qwen/Qwen3-8B`.
- Links to the paper on the Hugging Face Hub for better discoverability.
- A BibTeX citation section from your GitHub repository.
These additions help users find your model more easily and provide a standard way for them to cite your work.
README.md
CHANGED
|
@@ -1,3 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
<p align="center">
|
| 2 |
<img src="figures/logo.jpg" alt="AROMA Logo" width="120">
|
| 3 |
</p>
|
|
@@ -5,11 +11,11 @@
|
|
| 5 |
<h2 align="center"> 𧬠AROMA: Augmented Reasoning Over a Multimodal Architecture for Virtual Cell Genetic Perturbation Modeling<br>(ACL 2026 Findings)</h2>
|
| 6 |
|
| 7 |
<p align="center">
|
| 8 |
-
π <a href="https://
|
| 9 |
</p>
|
| 10 |
</p>
|
| 11 |
|
| 12 |
-
> Please refer to our [repository](https://github.com/blazerye/AROMA) and [paper](https://
|
| 13 |
|
| 14 |
## π Overview
|
| 15 |
|
|
@@ -25,4 +31,16 @@ The overall AROMA pipeline is illustrated in the figure above and is divided int
|
|
| 25 |
|
| 26 |
- **Modeling stage.** AROMA adopts a retrieval-augmented strategy to incorporate query-relevant information, thereby providing explicit evidence cues for prediction. In addition, it jointly leverages topological representations learned from graph neural networks (GNN) and protein sequence representations encoded by ESM-2, and applies a cross-attention module to explicitly model perturbation-target gene dependencies across modalities.
|
| 27 |
|
| 28 |
-
- **Training stage.** AROMA first performs multimodal supervised fine-tuning (SFT), and is then further optimized with Group Relative Policy Optimization (GRPO) reinforcement learning to enhance predictive performance while generating biologically meaningful explanations.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
library_name: transformers
|
| 3 |
+
pipeline_tag: text-generation
|
| 4 |
+
base_model: Qwen/Qwen3-8B
|
| 5 |
+
---
|
| 6 |
+
|
| 7 |
<p align="center">
|
| 8 |
<img src="figures/logo.jpg" alt="AROMA Logo" width="120">
|
| 9 |
</p>
|
|
|
|
| 11 |
<h2 align="center"> 𧬠AROMA: Augmented Reasoning Over a Multimodal Architecture for Virtual Cell Genetic Perturbation Modeling<br>(ACL 2026 Findings)</h2>
|
| 12 |
|
| 13 |
<p align="center">
|
| 14 |
+
π <a href="https://huggingface.co/papers/2604.20263" target="_blank">Paper</a> β’ π <a href="https://github.com/blazerye/AROMA" target="_blank">Code</a> β’ ποΈ <a href="https://huggingface.co/datasets/blazerye/PerturbReason" target="_blank">Datasets</a><br>
|
| 15 |
</p>
|
| 16 |
</p>
|
| 17 |
|
| 18 |
+
> Please refer to our [repository](https://github.com/blazerye/AROMA) and [paper](https://huggingface.co/papers/2604.20263) for more details.
|
| 19 |
|
| 20 |
## π Overview
|
| 21 |
|
|
|
|
| 31 |
|
| 32 |
- **Modeling stage.** AROMA adopts a retrieval-augmented strategy to incorporate query-relevant information, thereby providing explicit evidence cues for prediction. In addition, it jointly leverages topological representations learned from graph neural networks (GNN) and protein sequence representations encoded by ESM-2, and applies a cross-attention module to explicitly model perturbation-target gene dependencies across modalities.
|
| 33 |
|
| 34 |
+
- **Training stage.** AROMA first performs multimodal supervised fine-tuning (SFT), and is then further optimized with Group Relative Policy Optimization (GRPO) reinforcement learning to enhance predictive performance while generating biologically meaningful explanations.
|
| 35 |
+
|
| 36 |
+
## π Citation
|
| 37 |
+
If you find AROMA useful for your research and applications, please cite using this BibTeX:
|
| 38 |
+
```bibtex
|
| 39 |
+
@inproceedings{wang2026aroma,
|
| 40 |
+
title="{AROMA}: Augmented Reasoning Over a Multimodal Architecture for Virtual Cell Genetic Perturbation Modeling",
|
| 41 |
+
author="Wang, Zhenyu and Ye, Geyan and Liu, Wei and Ng, Man Tat Alexander",
|
| 42 |
+
booktitle="Findings of the Association for Computational Linguistics: ACL 2026",
|
| 43 |
+
year="2026",
|
| 44 |
+
publisher="Association for Computational Linguistics"
|
| 45 |
+
}
|
| 46 |
+
```
|