lijiayangCS commited on
Commit
d1ed0f3
·
verified ·
1 Parent(s): 93b1197

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +70 -1
README.md CHANGED
@@ -10,4 +10,73 @@ new_version: lijiayangCS/StableI2I_PLUS
10
  pipeline_tag: image-text-to-text
11
  ---
12
 
13
- # StableI2I
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  pipeline_tag: image-text-to-text
11
  ---
12
 
13
+ # StableI2I
14
+
15
+ Official implementation of **StableI2I: Spotting Unintended Changes in Image-to-Image Transition**
16
+ **ICML 2026**
17
+
18
+ > This HuggingFace repository provides the checkpoint used in the paper.
19
+ > For the latest code, demo, inference scripts, and score-supported version, please refer to the official GitHub repository:
20
+ > https://github.com/Henry-Lee-real/StableI2I
21
+
22
+ Any questions can be consulted via email: **lijiayang.cs@gmail.com**
23
+
24
+ Looking forward to your ⭐!
25
+
26
+ ---
27
+
28
+ ## 📌 TODOs
29
+
30
+ - [x] Release code
31
+ - [x] Release checkpoint
32
+ - [ ] Release pip package
33
+ - [ ] Release arXiv version
34
+ - [ ] Release ICML camera-ready paper
35
+ - [ ] Release HuggingFace project page
36
+
37
+ ---
38
+
39
+ ## 🔥 News
40
+
41
+ - **StableI2I** is accepted by **ICML 2026**.
42
+ - This HuggingFace repository hosts the checkpoint used in the paper.
43
+ - The latest codebase is maintained in the official GitHub repository.
44
+ - If you need the version with explicit **score output**, please use the latest GitHub code.
45
+
46
+ ---
47
+
48
+ ## Core Concept
49
+
50
+ In most real-world image-to-image (I2I) scenarios, existing evaluations primarily focus on instruction following and perceptual quality or aesthetics of the generated images. However, they often fail to assess whether the output image faithfully preserves the semantic correspondence, spatial structure, and low-level appearance of the input image.
51
+
52
+ To address this limitation, we propose **StableI2I**, a unified and dynamic evaluation framework for measuring content fidelity and pre--post consistency in image-to-image transitions. StableI2I does not require reference images and can be applied to a wide range of I2I tasks, including image editing and image restoration.
53
+
54
+ StableI2I evaluates unintended changes from three complementary perspectives:
55
+
56
+ 1. **Semantic Level**
57
+ Checks whether the output introduces unintended object-level or meaning-level changes, such as object addition, removal, replacement, or identity drift.
58
+
59
+ 2. **Structure Level**
60
+ Checks whether the output preserves spatial layout and geometric consistency, including misalignment, deformation, repainting, and structural distortion.
61
+
62
+ 3. **Low-level Appearance**
63
+ Checks whether the output introduces unintended visual degradation, such as blur, noise, color cast, exposure degradation, or artifacts.
64
+
65
+ In addition, we construct **StableI2I-Bench**, a benchmark designed to systematically evaluate the ability of MLLMs to judge content fidelity and consistency in image-to-image tasks.
66
+
67
+ Extensive experiments show that StableI2I provides accurate, fine-grained, and interpretable evaluations, with strong correlations to human subjective judgments. It serves as a practical evaluation tool for diagnosing content consistency and benchmarking real-world I2I systems.
68
+
69
+ ---
70
+
71
+ ## Model Checkpoint
72
+
73
+ This HuggingFace repository provides the checkpoint used in the StableI2I paper.
74
+
75
+ Please note:
76
+
77
+ - The checkpoint corresponds to the paper version.
78
+ - For the latest inference pipeline, API interface, and score-supported output format, please refer to the official GitHub repository.
79
+ - The model is built upon the Qwen3-VL environment and follows the Qwen3-VL inference style.
80
+
81
+ ---
82
+