finsql-mlx-nemotron-nano-9b-v2-4bits

This is a LoRA adapter for financial SQL generation, fine-tuned on mlx-community/NVIDIA-Nemotron-Nano-9B-v2-4bits.

Latest Finetuning

Screenshot 2025-10-11 at 13.28.30

HF Model Path: gccmorgoth/finsql-mlx-nemotron-nano-9b-v2-4bits

Finetuning Details

  • Method: Direct Preference Optimization (DPO)
  • Checkpoint: Iteration 600
  • Validation Loss: 0.069
  • Training Loss: 0.0002
  • Learning Rate: Cosine decay with warmup
  • LoRA Rank: 16
  • Beta: 0.1

Performance

  • Validation loss: 0.069 (optimal convergence point)
  • Selected at iteration 600 to prevent overfitting
  • DPO training for improved preference alignment on financial SQL tasks

Model Selection

  • Checkpoint: Iteration 600 selected based on validation loss curve
  • Rationale: Optimal balance between training convergence and generalization
  • Training Dynamics: Early stopping before overfitting (val loss increased at iter 700+)

Dataset

This model was fine-tuned on financial text-to-sql data pairs, specifically the FinSQLBull dataset, to improve SQL query generation for financial databases and tables.

Usage

Recommended prompt format to specify:

Database: [database_name]

[Schema information]

Task

[Natural language question about the data]

Constraint: [Any specific constraints]

SQL: [Model Generated SQL Query]

Sample Prompt Format

Database: company_financials

Table: revenue (id, company, year, revenue, profit)

Task

What was the total revenue for all companies in 2023?

SQL: [Model Generated SQL Query]

Python

from mlx_lm import load, generate

model, tokenizer = load("your-username/your-model-name")
response = generate(model, tokenizer, prompt="Your prompt here")
Downloads last month

-

Downloads are not tracked for this model. How to track
MLX
Hardware compatibility
Log In to add your hardware

Quantized

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for gccmorgoth/finsql-mlx-nvidia-nemotron-nano-9b-v2-4bits