--- license: cc-by-nc-4.0 language: - en base_model: - deepseek-ai/deepseek-coder-6.7b-instruct - google/siglip-so400m-patch14-384 pipeline_tag: image-text-to-text --- # Aligned Multi-View Scripts for Universal Chart-to-Code Generation CharLuMA-1.3B generates a plotting script in Python, R, or LaTeX from a chart image. This 1.3B variant is from the paper *"Aligned Multi-View Scripts for Universal Chart-to-Code Generation"* (ACL 2026 Main Conference). - **Paper:** https://arxiv.org/abs/2604.24559 - **Code (required for inference):** https://github.com/zhihan72/CharLuMA - **6.7B Variant:** [CharLuMA-6.7B](https://huggingface.co/Zhihan/CharLuMA-6.7B) - **Training data:** [Chart2NCode](https://huggingface.co/datasets/Zhihan/Chart2NCode) | Backbone | Vision encoder | Output languages | dtype | |---|---|---|---| | DeepSeek-Coder-1.3B-Instruct | SigLIP-SO400M-patch14-384 | Python, R, LaTeX | bfloat16 | ## Usage This is a custom architecture, so `AutoModel.from_pretrained` will not load it. Clone the codebase and use its loader: ```bash git clone https://github.com/Zhihan72/CharLuMA ``` Before loading, replace the placeholder paths in `config.json` (`/your_local_path/...`) with `deepseek-ai/deepseek-coder-1.3b-instruct` (or a local copy) and `google/siglip-so400m-patch14-384` (or a local copy). Inference example: `scripts/inference_charluma.py` in the repo. ## Citation If you find our work useful, consider citing our paper as follows: ```bibtex @misc{zhang2026aligned, title = {Aligned Multi-View Scripts for Universal Chart-to-Code Generation}, author = {Zhihan Zhang and Lizi Liao}, year = {2026}, eprint = {2604.24559}, archivePrefix = {arXiv}, primaryClass = {cs.CL}, url = {https://arxiv.org/abs/2604.24559} } ```