| --- |
| library_name: transformers |
| tags: [fine-tuning, medical, oncology, reasoning, chain-of-thought, Qwen-1.5B] |
| --- |
| |
| # Model Card for Medical Oncology Reasoning Fine-Tuned Model |
|
|
| This is a fine-tuned version of the **DeepSeek-R1-Distill-Qwen-1.5B** model, specifically adapted for medical oncology reasoning tasks using chain-of-thought prompting. The fine-tuning was performed on a curated subset of the [FreedomIntelligence/medical-o1-reasoning-SFT](https://huggingface.co/datasets/FreedomIntelligence/medical-o1-reasoning-SFT) dataset. This model is designed to provide detailed, step-by-step reasoning when answering medical questions, making it suitable for clinical decision support, medical education, and research. |
|
|
| ## Model Details |
|
|
| ### Model Description |
|
|
| This model leverages the capabilities of **DeepSeek-R1-Distill-Qwen-1.5B** and has been fine-tuned to enhance its performance in medical reasoning. It incorporates chain-of-thought prompting to produce detailed explanations for clinical queries. The fine-tuning process focused on tasks related to clinical diagnostics, treatment planning, and medical reasoning. |
|
|
| - **Developed by:** Arihant Tripathi |
| - **Model type:** Causal Language Model (LLM) |
| - **Language(s):** English |
| - **Finetuned from model:** [deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B) |
|
|
| ### Model Sources |
|
|
| - **Repository:** [EndOfLe/OncoFineTuned](https://huggingface.co/EndOfLe/OncoFineTuned) |
|
|
| ## Uses |
|
|
| ### Direct Use |
|
|
| This model can be used directly to generate detailed, step-by-step medical reasoning in response to clinical queries. It is especially useful for: |
| - Medical diagnosis support. |
| - Clinical reasoning and treatment planning. |
| - Medical education and research. |
|
|
| ### Downstream Use |
|
|
| The model can be integrated into larger clinical decision support systems or educational tools that require natural language understanding and detailed reasoning for medical queries. |
|
|
| ### Out-of-Scope Use |
|
|
| The model is not intended for: |
| - Replacing expert medical advice or making final clinical decisions. |
| - General-purpose language generation without domain adaptation. |
| - High-stakes applications where errors could have severe consequences without expert oversight. |
|
|
| ## Bias, Risks, and Limitations |
|
|
| The model’s output should be treated as an aid to human decision-making rather than a substitute for professional medical advice. Key considerations include: |
| - The model may generate outdated or incorrect medical information. |
| - The reasoning is based on the training data and might reflect its inherent biases. |
| - Use in clinical settings should always involve human review and validation. |
|
|
| ### Recommendations |
|
|
| Users should verify the model’s outputs with expert medical knowledge and ensure its use complies with clinical standards and ethical guidelines. |
|
|
| ## How to Get Started with the Model |
|
|
| You can get started with the model using the following code snippet: |
|
|
| ```python |
| from transformers import AutoTokenizer, AutoModelForCausalLM |
| |
| model_name = "EndOfLe/OncoFineTuned" |
| tokenizer = AutoTokenizer.from_pretrained(model_name) |
| model = AutoModelForCausalLM.from_pretrained(model_name) |
| |
| input_text = "What are the causes of colon cancer?" |
| inputs = tokenizer(input_text, return_tensors="pt") |
| outputs = model.generate(**inputs, max_new_tokens=256) |
| print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
| |