YAML Metadata Warning:empty or missing yaml metadata in repo card
Check out the documentation for more information.
SpatialVLA Merged Model
This model is a merged version of:
- Base Model:
/remote-home/share/chenglong/Workplace/SpatialVLA/ckpts_pretrained/spatialvla-4b-224-pt - LoRA Adapter:
/remote-home/share/chenglong/Workplace/SpatialVLA/outputs/spatialvla_4b_finetune/2025-10-21/03-55-14_glasses_sigma5_dataset_spatialvla-4b-224-pt_lr5e-6_bs32_node1_gpu2_r32_a32_ep10_linear+emb+h/checkpoint-30000
Merge Details
- LoRA Rank (r): 32
- LoRA Alpha: 32
- Target Modules: o_proj, position_embedding_head.0, k_proj, v_proj, fc1, up_proj, out_proj, spatial_embed_tokens, fc2, q_proj, linear, down_proj, gate_proj, position_embedding_head.3, lm_head
- Merge Date: <function get_file_binaries_from_pathnames at 0x7f1cb6c5b0a0>
Usage
This merged model can be used directly without PEFT:
import torch
from transformers import AutoModel, AutoProcessor
model_path = "/remote-home/share/chenglong/Workplace/SpatialVLA/ckpts_merged/glassvla-4b-sam2-lora-percent10-30k-sigma-5"
processor = AutoProcessor.from_pretrained(model_path, trust_remote_code=True)
model = AutoModel.from_pretrained(
model_path,
trust_remote_code=True,
torch_dtype=torch.bfloat16
).eval().cuda()
# Use the model for inference
# ... your inference code here ...
Notes
- This is a fully merged model, so the LoRA adapter is no longer needed.
- The model can be used just like the original base model.
- All weights have been merged into a single set of parameters.
- Downloads last month
- 1
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support