Image-Text-to-Text
Transformers
Diffusers
Safetensors
qwen3_vl
vision-language-model
image-decomposition
conversational
Instructions to use SynLayers/Bbox-caption-8b with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use SynLayers/Bbox-caption-8b with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("image-text-to-text", model="SynLayers/Bbox-caption-8b") messages = [ { "role": "user", "content": [ {"type": "image", "url": "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/p-blog/candy.JPG"}, {"type": "text", "text": "What animal is on the candy?"} ] }, ] pipe(text=messages)# Load model directly from transformers import AutoProcessor, AutoModelForImageTextToText processor = AutoProcessor.from_pretrained("SynLayers/Bbox-caption-8b") model = AutoModelForImageTextToText.from_pretrained("SynLayers/Bbox-caption-8b") messages = [ { "role": "user", "content": [ {"type": "image", "url": "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/p-blog/candy.JPG"}, {"type": "text", "text": "What animal is on the candy?"} ] }, ] inputs = processor.apply_chat_template( messages, add_generation_prompt=True, tokenize=True, return_dict=True, return_tensors="pt", ).to(model.device) outputs = model.generate(**inputs, max_new_tokens=40) print(processor.decode(outputs[0][inputs["input_ids"].shape[-1]:])) - Notebooks
- Google Colab
- Kaggle
- Local Apps
- vLLM
How to use SynLayers/Bbox-caption-8b with vLLM:
Install from pip and serve model
# Install vLLM from pip: pip install vllm # Start the vLLM server: vllm serve "SynLayers/Bbox-caption-8b" # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:8000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "SynLayers/Bbox-caption-8b", "messages": [ { "role": "user", "content": [ { "type": "text", "text": "Describe this image in one sentence." }, { "type": "image_url", "image_url": { "url": "https://cdn.britannica.com/61/93061-050-99147DCE/Statue-of-Liberty-Island-New-York-Bay.jpg" } } ] } ] }'Use Docker
docker model run hf.co/SynLayers/Bbox-caption-8b
- SGLang
How to use SynLayers/Bbox-caption-8b with SGLang:
Install from pip and serve model
# Install SGLang from pip: pip install sglang # Start the SGLang server: python3 -m sglang.launch_server \ --model-path "SynLayers/Bbox-caption-8b" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "SynLayers/Bbox-caption-8b", "messages": [ { "role": "user", "content": [ { "type": "text", "text": "Describe this image in one sentence." }, { "type": "image_url", "image_url": { "url": "https://cdn.britannica.com/61/93061-050-99147DCE/Statue-of-Liberty-Island-New-York-Bay.jpg" } } ] } ] }'Use Docker images
docker run --gpus all \ --shm-size 32g \ -p 30000:30000 \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HF_TOKEN=<secret>" \ --ipc=host \ lmsysorg/sglang:latest \ python3 -m sglang.launch_server \ --model-path "SynLayers/Bbox-caption-8b" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "SynLayers/Bbox-caption-8b", "messages": [ { "role": "user", "content": [ { "type": "text", "text": "Describe this image in one sentence." }, { "type": "image_url", "image_url": { "url": "https://cdn.britannica.com/61/93061-050-99147DCE/Statue-of-Liberty-Island-New-York-Bay.jpg" } } ] } ] }' - Docker Model Runner
How to use SynLayers/Bbox-caption-8b with Docker Model Runner:
docker model run hf.co/SynLayers/Bbox-caption-8b
File size: 12,551 Bytes
b752efc | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 | import argparse
import json
import logging
import os
import re
import sys
from pathlib import Path
import numpy as np
import torch
from PIL import Image
PROJECT_ROOT = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
if PROJECT_ROOT not in sys.path:
sys.path.insert(0, PROJECT_ROOT)
logging.getLogger("transformers.tokenization_utils_base").setLevel(logging.ERROR)
os.environ["CUDA_VISIBLE_DEVICES"] = os.environ.get("CUDA_VISIBLE_DEVICES", "0")
from infer.common_infer import initialize_pipeline, quantize_box_16, scale_box_xyxy
from tools.tools import load_config, seed_everything
def load_real_metadata(jsonl_path: str):
"""Load real-test metadata from JSONL."""
items = []
with open(jsonl_path, "r", encoding="utf-8") as f:
for line in f:
line = line.strip()
if line:
items.append(json.loads(line))
return items
def extract_checkpoint_tag(path: str):
"""Extract a checkpoint tag like scaleup_1024_20k or original_1024_512seq."""
if not path:
return None
match = re.search(r"ckpt_prism_([^/]+)", path)
if match:
return match.group(1)
return None
def derive_run_name(config: dict) -> str:
"""Derive the result subfolder name from the active checkpoint setup."""
checkpoint_tags = {}
for key in ("lora_ckpt", "layer_ckpt", "adapter_lora_dir"):
tag = extract_checkpoint_tag(config.get(key, ""))
if tag:
checkpoint_tags[key] = tag
if checkpoint_tags:
unique_tags = sorted(set(checkpoint_tags.values()))
if len(unique_tags) != 1:
details = ", ".join(f"{key}={value}" for key, value in checkpoint_tags.items())
raise ValueError(
"Checkpoint paths are inconsistent. "
"Please switch lora_ckpt, layer_ckpt, and adapter_lora_dir together. "
f"Current tags: {details}"
)
inferred_tag = unique_tags[0]
else:
inferred_tag = "real_infer"
if config.get("run_name"):
return config["run_name"]
return inferred_tag
def build_run_save_dir(config: dict):
"""Build the final save directory as <save_dir>/<run_name>."""
save_root = config.get("save_dir", "./real_inference_output")
run_name = derive_run_name(config)
return os.path.join(save_root, run_name), run_name
def resolve_image_path(sample: dict, data_dir: str, image_dir: str = None) -> str:
"""Resolve the input image path, preferring local files_real_test images."""
sample_name = sample.get("sample_or_stem", "")
image_path = sample.get("image", "")
if image_dir is None and data_dir:
image_dir = os.path.join(data_dir, "layers_real_test_1024")
candidates = []
if image_dir:
if sample_name:
candidates.extend(
[
os.path.join(image_dir, f"{sample_name}.png"),
os.path.join(image_dir, f"{sample_name}.jpg"),
os.path.join(image_dir, f"{sample_name}.jpeg"),
]
)
if image_path:
candidates.append(os.path.join(image_dir, os.path.basename(image_path)))
if image_path:
candidates.append(image_path)
if data_dir and not os.path.isabs(image_path):
candidates.append(os.path.join(data_dir, image_path))
seen = set()
for candidate in candidates:
if not candidate or candidate in seen:
continue
seen.add(candidate)
if os.path.exists(candidate):
return candidate
raise FileNotFoundError(
f"Could not resolve image for sample '{sample_name}'. "
f"Tried local image_dir='{image_dir}' and json path '{image_path}'."
)
def quantize_box_16_safe(box: tuple, target_size: int) -> tuple:
"""Quantize a box to the 16-pixel grid and keep at least one latent cell."""
x0_q, y0_q, x1_q, y1_q = quantize_box_16(box, target_size)
if x1_q <= x0_q:
if x0_q + 16 <= target_size:
x1_q = x0_q + 16
else:
x0_q = max(0, target_size - 16)
x1_q = target_size
if y1_q <= y0_q:
if y0_q + 16 <= target_size:
y1_q = y0_q + 16
else:
y0_q = max(0, target_size - 16)
y1_q = target_size
return (x0_q, y0_q, x1_q, y1_q)
def get_real_boxes(sample: dict, source_size: int, target_size: int) -> list:
"""Scale and quantize real-test boxes from JSON metadata."""
boxes = []
for box in sample.get("bboxes", []):
if not isinstance(box, (list, tuple)) or len(box) != 4:
continue
scaled_box = scale_box_xyxy(box, source_size, target_size)
boxes.append(quantize_box_16_safe(scaled_box, target_size))
return boxes
def load_adapter_image(sample: dict, target_size: int, config: dict):
"""Load and resize the real-test image used as adapter input."""
image_path = resolve_image_path(
sample,
data_dir=config.get("data_dir", ""),
image_dir=config.get("image_dir"),
)
img = Image.open(image_path).convert("RGB")
if img.size != (target_size, target_size):
img = img.resize((target_size, target_size), Image.LANCZOS)
return img, image_path
def format_source_image_path(image_path: str, config: dict) -> str:
path = Path(image_path)
for key in ("image_dir", "data_dir"):
root = config.get(key)
if not root:
continue
try:
return path.relative_to(Path(root)).as_posix()
except ValueError:
continue
return path.name
@torch.no_grad()
def inference_real(config):
"""Main inference function for the real-test dataset."""
if config.get("seed") is not None:
seed_everything(config["seed"])
source_size = config.get("source_size", 1024)
target_size = config.get("target_size", 1024)
max_layer_num = config.get("max_layer_num", 52)
print(f"[INFO] Source size: {source_size}, Target size: {target_size}", flush=True)
save_dir, run_name = build_run_save_dir(config)
os.makedirs(save_dir, exist_ok=True)
os.makedirs(os.path.join(save_dir, "merged"), exist_ok=True)
os.makedirs(os.path.join(save_dir, "merged_rgba"), exist_ok=True)
print(f"[INFO] Run name: {run_name}", flush=True)
print(f"[INFO] Results will be saved to: {save_dir}", flush=True)
pipeline, transp_vae = initialize_pipeline(config)
test_jsonl = config.get("test_jsonl", "")
if not test_jsonl or not os.path.exists(test_jsonl):
raise ValueError(f"Test JSONL not found: {test_jsonl}")
all_samples = load_real_metadata(test_jsonl)
total_available = len(all_samples)
start_idx = config.get("start_idx", 1)
end_idx = config.get("end_idx", total_available)
max_samples = config.get("max_samples", None)
if max_samples and not config.get("end_idx"):
end_idx = min(start_idx + max_samples - 1, total_available)
start_idx = max(1, min(start_idx, total_available))
end_idx = max(start_idx, min(end_idx, total_available))
samples = all_samples[start_idx - 1 : end_idx]
print(f"[INFO] Total samples in dataset: {total_available}", flush=True)
print(
f"[INFO] Processing samples {start_idx} to {end_idx} ({len(samples)} samples)",
flush=True,
)
generator = torch.Generator(device=torch.device("cuda")).manual_seed(
config.get("seed", 42)
)
for local_idx, sample in enumerate(samples):
idx_zero_based = start_idx - 1 + local_idx
sample_name = sample.get("sample_or_stem", f"real_{idx_zero_based:06d}")
print(
f"Processing [{local_idx + 1}/{len(samples)}] idx={idx_zero_based} ({sample_name})...",
flush=True,
)
try:
layer_boxes = get_real_boxes(sample, source_size, target_size)
adapter_img, image_path = load_adapter_image(sample, target_size, config)
except Exception as e:
print(f" Error preparing sample: {e}", flush=True)
continue
whole_box = (0, 0, target_size, target_size)
bg_box = (0, 0, target_size, target_size)
all_boxes = [whole_box, bg_box] + layer_boxes
if len(all_boxes) > max_layer_num:
print(
f" Skipping sample because num_layers={len(all_boxes)} exceeds max_layer_num={max_layer_num}",
flush=True,
)
continue
caption = sample.get("whole_caption", "")
print(f" Size: {target_size}x{target_size}, Layers: {len(all_boxes)}", flush=True)
try:
x_hat, image, _ = pipeline(
prompt=caption,
adapter_image=adapter_img,
adapter_conditioning_scale=config.get("adapter_scale", 0.9),
validation_box=all_boxes,
generator=generator,
height=target_size,
width=target_size,
guidance_scale=config.get("cfg", 4.0),
num_layers=len(all_boxes),
sdxl_vae=transp_vae,
)
except Exception as e:
print(f" Error during inference: {e}", flush=True)
continue
x_hat = (x_hat + 1) / 2
x_hat = x_hat.squeeze(0).permute(1, 0, 2, 3).to(torch.float32)
case_dir = os.path.join(save_dir, sample_name)
os.makedirs(case_dir, exist_ok=True)
whole_image_layer = (
x_hat[0].permute(1, 2, 0).cpu().numpy() * 255
).astype(np.uint8)
Image.fromarray(whole_image_layer, "RGBA").save(
os.path.join(case_dir, "whole_image_rgba.png")
)
background_layer = (
x_hat[1].permute(1, 2, 0).cpu().numpy() * 255
).astype(np.uint8)
Image.fromarray(background_layer, "RGBA").save(
os.path.join(case_dir, "background_rgba.png")
)
adapter_img.save(os.path.join(case_dir, "origin.png"))
merged_image = image[1]
for layer_idx in range(2, x_hat.shape[0]):
rgba_layer = (
x_hat[layer_idx].permute(1, 2, 0).cpu().numpy() * 255
).astype(np.uint8)
rgba_image = Image.fromarray(rgba_layer, "RGBA")
rgba_image.save(os.path.join(case_dir, f"layer_{layer_idx - 2}_rgba.png"))
merged_image = Image.alpha_composite(merged_image.convert("RGBA"), rgba_image)
merged_image.convert("RGB").save(
os.path.join(save_dir, "merged", f"{sample_name}.png")
)
merged_image.convert("RGB").save(os.path.join(case_dir, "merged.png"))
merged_image.save(os.path.join(save_dir, "merged_rgba", f"{sample_name}.png"))
case_meta = {
"sample_idx_zero_based": idx_zero_based,
"sample_idx_one_based": idx_zero_based + 1,
"sample_name": sample_name,
"source_image_path": format_source_image_path(image_path, config),
"target_size": target_size,
"source_size": source_size,
"raw_num_layers": sample.get("num_layers"),
"num_layers": len(all_boxes),
"raw_boxes": sample.get("bboxes", []),
"boxes": all_boxes,
"caption": caption,
"run_name": run_name,
}
with open(os.path.join(case_dir, "inference_meta.json"), "w", encoding="utf-8") as f:
json.dump(case_meta, f, indent=2)
if idx_zero_based % 10 == 0:
torch.cuda.empty_cache()
print(f"[INFO] Inference complete. Results saved to {save_dir}", flush=True)
del pipeline
if torch.cuda.is_available():
torch.cuda.empty_cache()
def main():
parser = argparse.ArgumentParser()
parser.add_argument(
"--config_path",
"-c",
type=str,
required=True,
help="Path to the YAML configuration file.",
)
parser.add_argument(
"--start_idx",
type=int,
default=None,
help="1-based start index for the JSONL entries.",
)
parser.add_argument(
"--end_idx",
type=int,
default=None,
help="1-based end index for the JSONL entries (inclusive).",
)
args = parser.parse_args()
config = load_config(args.config_path)
if args.start_idx is not None:
config["start_idx"] = args.start_idx
if args.end_idx is not None:
config["end_idx"] = args.end_idx
inference_real(config)
if __name__ == "__main__":
main()
|