--- license: cc-by-nc-4.0 language: - it tags: - llama - llama-3 - meta - medical-qa - italian - biomedical - question-answering - fine-tuning - unsloth - bnb - 4bit - imb - Ortopedia datasets: - praiselab-picuslab/IMB base_model: - unsloth/Llama-3.2-1B-Instruct-unsloth-bnb-4bit --- # ๐Ÿง  Llama-3.2-1B-Instruct โ€” IMB Ortopedia Fine-Tuned Model This model is a fine-tuned version of [`unsloth/Llama-3.2-1B-Instruct-unsloth-bnb-4bit`](https://huggingface.co/unsloth/Llama-3.2-1B-Instruct-unsloth-bnb-4bit), optimized for **Italian medical question answering**, with a specific focus on **Ortopedia**. The fine-tuning was performed using a **subset of the IMB (Italian Medical Benchmark) dataset**, specifically: - Ortopedia category only - ~10,000 training samples The training was performed using the **Unsloth** library with LoRA fine-tuning, and the adapter weights were later merged into the base model to provide a standalone checkpoint. This model relies on data from the IMB dataset. **If you use this model in research or applications, you must cite the IMB paper (see Citation section below).** --- ## ๐Ÿ“š Training Dataset โ€” IMB (Italian Medical Benchmark) IMB is an Italian benchmark for medical question answering, designed to evaluate and improve LLM performance in clinical-domain Italian language understanding and reasoning. The full dataset includes: - **IMB-QA**: 782,644 doctor-patient conversations collected from Italian online medical forums - **IMB-MCQA**: 25,862 multiple-choice questions derived from Italian medical specialization exams โš ๏ธ **Important:** This model was trained **only on the Ortopedia subset (~10,000 samples)** of IMB, not on the full dataset. Dataset repository: ๐Ÿ‘‰ https://github.com/PRAISELab-PicusLab/IMB --- ## ๐Ÿงช Usage Example ```python from transformers import AutoModelForCausalLM, AutoTokenizer model = AutoModelForCausalLM.from_pretrained("praiselab-picuslab/Llama-3.2-1B-Instruct-Ortopedia") tokenizer = AutoTokenizer.from_pretrained("praiselab-picuslab/Llama-3.2-1B-Instruct-Ortopedia") prompt = "[Example question in Italian about Ortopedia]" inputs = tokenizer(prompt, return_tensors="pt") outputs = model.generate(**inputs, max_new_tokens=150) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ``` --- ## โš ๏ธ Usage Restrictions * Allowed use: **Non-commercial research only** * Redistribution: Not allowed without explicit authorization * Mandatory citation: The IMB dataset paper must be cited in any publication or derived work --- ## ๐Ÿ“„ Citation If you use this model, the IMB dataset, or derived outputs in research, please cite: ```bibtex @inproceedings{DBLP:conf/clic-it/RomanoRBPM25, author = {Antonio Romano and Giuseppe Riccio and Mariano Barone and Marco Postiglione and Vincenzo Moscato}, editor = {Cristina Bosco and Elisabetta Jezek and Marco Polignano and Manuela Sanguinetti}, title = {{IMB:} An Italian Medical Benchmark for Question Answering}, booktitle = {Proceedings of the Eleventh Italian Conference on Computational Linguistics (CLiC-it 2025), Cagliari, Italy, September 24-26, 2025}, series = {{CEUR} Workshop Proceedings}, volume = {4112}, publisher = {CEUR-WS.org}, year = {2025}, url = {https://ceur-ws.org/Vol-4112/92_main_long.pdf} } ``` --- ## ๐Ÿ— Training Details * Base model: `unsloth/Llama-3.2-1B-Instruct-unsloth-bnb-4bit` * Fine-tuning method: LoRA (Unsloth) * Quantization: 4-bit (BitsAndBytes) * Adapter merging: Yes (Full merged model) * Language: Italian * Domain: Medical โ€” Ortopedia * Training size: ~10,000 samples --- ## ๐Ÿ“œ License This work is licensed under a [Creative Commons Attribution-NonCommercial-NoDerivs 4.0 International License][cc-by-nc-nd]. [![CC BY-NC-ND 4.0][cc-by-nc-nd-image]][cc-by-nc-nd] [cc-by-nc-nd]: http://creativecommons.org/licenses/by-nc-nd/4.0/ [cc-by-nc-nd-image]: https://licensebuttons.net/l/by-nc-nd/4.0/88x31.png [cc-by-nc-nd-shield]: https://img.shields.io/badge/License-CC%20BY--NC--ND%204.0-lightgrey.svg --- ## ๐Ÿค Acknowledgements ๐Ÿ‘จโ€๐Ÿ’ป This project was developed by Mariano Barone, Roberta Di Marino, Francesco Di Serio, Giovanni Dioguardi, Marco Postiglione, Antonio Romano, Giuseppe Riccio, and Vincenzo Moscato at University of Naples, Federico II