| --- |
| datasets: |
| - Quinn777/AMATH-SFT |
| base_model: |
| - liuhaotian/llava-v1.5-7b |
| --- |
| # Model Card for AtomThink-LlamaV |
|
|
| The model is post-trained based on llava-v1.5-7b and the AtomThink framework, and can be used to solve complex multimodal mathematical problems. |
|
|
| # Citation |
| If you use this dataset in your research, please cite: |
| ```text |
| @article{xiang2025can, |
| title={Can Atomic Step Decomposition Enhance the Self-structured Reasoning of Multimodal Large Models?}, |
| author={Xiang, Kun and Liu, Zhili and Jiang, Zihao and Nie, Yunshuang and Cai, Kaixin and Yin, Yiyang and Huang, Runhui and Fan, Haoxiang and Li, Hanhui and Huang, Weiran and others}, |
| journal={arXiv preprint arXiv:2503.06252}, |
| year={2025} |
| } |
| |
| @article{xiang2024atomthink, |
| title={AtomThink: A Slow Thinking Framework for Multimodal Mathematical Reasoning}, |
| author={Xiang, Kun and Liu, Zhili and Jiang, Zihao and Nie, Yunshuang and Huang, Runhui and Fan, Haoxiang and Li, Hanhui and Huang, Weiran and Zeng, Yihan and Han, Jianhua and others}, |
| journal={arXiv preprint arXiv:2411.11930}, |
| year={2024} |
| } |
| ``` |
|
|
| # License |
| The checkpoint is released under the Apache 2.0 license. Please ensure proper attribution when using this checkpoint. |