SynLayers commited on
Commit
e0543ce
·
verified ·
1 Parent(s): 12e6363

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +41 -0
README.md ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - vision-language-model
5
+ - image-decomposition
6
+ ---
7
+
8
+ # SynLayers
9
+
10
+ This repository contains the assets behind SynLayers, our two-stage image decomposition system.
11
+
12
+ At the root is the bbox-caption model. Given one image, it predicts:
13
+
14
+ - a whole-image caption
15
+ - bounding boxes for visible objects or layers
16
+
17
+ The same repo also includes the Stage 2 SynLayers pipeline to do layer decomposition.
18
+
19
+ If you want the easiest way to try the full system, please use our public demo:
20
+ [SynLayers/synlayers](https://huggingface.co/spaces/SynLayers/synlayers)
21
+
22
+ This repo is not meant to be used as a single generic `DiffusionPipeline(prompt)` model.
23
+ The full SynLayers pipeline is:
24
+
25
+ 1. bbox + whole-caption prediction
26
+ 2. layer decomposition into transparent RGBA outputs
27
+
28
+ If you only want the Stage 1 model at the repo root, you can load it with `transformers`.
29
+
30
+ ```python
31
+ from transformers import AutoProcessor, Qwen3VLForConditionalGeneration
32
+
33
+ model = Qwen3VLForConditionalGeneration.from_pretrained(
34
+ "SynLayers/Bbox-caption-8b",
35
+ torch_dtype="auto",
36
+ device_map="auto",
37
+ )
38
+ processor = AutoProcessor.from_pretrained("SynLayers/Bbox-caption-8b")
39
+ ```
40
+
41
+ Thanks for trying SynLayers.