diff --git "a/jsonl/Edit-Reward-Bench/labeled_data/0829_labels_0-104.jsonl" "b/jsonl/Edit-Reward-Bench/labeled_data/0829_labels_0-104.jsonl" new file mode 100644--- /dev/null +++ "b/jsonl/Edit-Reward-Bench/labeled_data/0829_labels_0-104.jsonl" @@ -0,0 +1,104 @@ +{"task_type": "ps_human", "data_source": "GEdit-Bench", "instruction": "Retouch this image.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/c9101db419bbcd1b258ed367dc09b986_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/ps_human/en/c9101db419bbcd1b258ed367dc09b986.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/ps_human/en/c9101db419bbcd1b258ed367dc09b986.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/ps_human/en/c9101db419bbcd1b258ed367dc09b986.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/ps_human/en/c9101db419bbcd1b258ed367dc09b986.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/ps_human/en/c9101db419bbcd1b258ed367dc09b986.png"], "label": "bad", "timestamp": "2025-08-29T13:37:04.203657"} +{"task_type": "ps_human", "data_source": "GEdit-Bench", "instruction": "Make me look like a handsome guy in this photo.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/6cabdec52f6113e0a365332f323053b1_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/ps_human/en/6cabdec52f6113e0a365332f323053b1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/6cabdec52f6113e0a365332f323053b1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/ps_human/en/6cabdec52f6113e0a365332f323053b1.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/ps_human/en/6cabdec52f6113e0a365332f323053b1.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/ps_human/en/6cabdec52f6113e0a365332f323053b1.png"], "label": "triple_ranking", "timestamp": "2025-08-29T13:41:04.482750", "instruction_following": {"ranking_text": "124|35"}, "consistency": {"ranking_text": "125|4|3"}, "overall": {"ranking_text": "12|4|5|3"}} +{"task_type": "ps_human", "data_source": "GEdit-Bench", "instruction": "How can I fix these facial imperfections?", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/47e2e49cf8b662a7493ddad42334b6e1_SRCIMG.png"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image/fullset/ps_human/en/47e2e49cf8b662a7493ddad42334b6e1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/ps_human/en/47e2e49cf8b662a7493ddad42334b6e1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/ps_human/en/47e2e49cf8b662a7493ddad42334b6e1.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/ps_human/en/47e2e49cf8b662a7493ddad42334b6e1.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/ps_human/en/47e2e49cf8b662a7493ddad42334b6e1.png"], "label": "triple_ranking", "timestamp": "2025-08-29T13:42:49.977263", "instruction_following": {"ranking_text": "3|45|12"}, "consistency": {"ranking_text": "1345|2"}, "overall": {"ranking_text": "3|45|1|2"}} +{"task_type": "ps_human", "data_source": "GEdit-Bench", "instruction": "Make the allergic reaction on my face look more severe.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/f0995ee97b33e6ca5effc808be8e4ac2_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/ps_human/en/f0995ee97b33e6ca5effc808be8e4ac2.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/ps_human/en/f0995ee97b33e6ca5effc808be8e4ac2.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/f0995ee97b33e6ca5effc808be8e4ac2.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/ps_human/en/f0995ee97b33e6ca5effc808be8e4ac2.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/ps_human/en/f0995ee97b33e6ca5effc808be8e4ac2.png"], "label": "triple_ranking", "timestamp": "2025-08-29T13:46:44.955855", "instruction_following": {"ranking_text": "134|25"}, "consistency": {"ranking_text": "2|4|13|5"}, "overall": {"ranking_text": "4|13|2|5"}} +{"task_type": "ps_human", "data_source": "GEdit-Bench", "instruction": "Make him look more handsome with sharper eyes", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/1f1c9a1e6ce6899d91abcb9a67922758_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/ps_human/en/1f1c9a1e6ce6899d91abcb9a67922758.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/ps_human/en/1f1c9a1e6ce6899d91abcb9a67922758.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/ps_human/en/1f1c9a1e6ce6899d91abcb9a67922758.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/ps_human/en/1f1c9a1e6ce6899d91abcb9a67922758.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/ps_human/en/1f1c9a1e6ce6899d91abcb9a67922758.png"], "label": "triple_ranking", "timestamp": "2025-08-29T13:50:28.039998", "instruction_following": {"ranking_text": "25|134"}, "consistency": {"ranking_text": "1234|5"}, "overall": {"ranking_text": "25|134"}} +{"task_type": "ps_human", "data_source": "GEdit-Bench", "instruction": "Make my face look better.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/f81d8419a96bac5878844b85e21a938c_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/ps_human/en/f81d8419a96bac5878844b85e21a938c.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/ps_human/en/f81d8419a96bac5878844b85e21a938c.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/ps_human/en/f81d8419a96bac5878844b85e21a938c.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/ps_human/en/f81d8419a96bac5878844b85e21a938c.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/ps_human/en/f81d8419a96bac5878844b85e21a938c.png"], "label": "triple_ranking", "timestamp": "2025-08-29T13:51:42.084775", "instruction_following": {"ranking_text": "23|145"}, "consistency": {"ranking_text": "1|3|245"}, "overall": {"ranking_text": "3|2|1|45"}} +{"task_type": "ps_human", "data_source": "GEdit-Bench", "instruction": "This is my photo—please make me look more handsome.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/038013b7852ce014b254effb307ec5de_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/ps_human/en/038013b7852ce014b254effb307ec5de.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/ps_human/en/038013b7852ce014b254effb307ec5de.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/ps_human/en/038013b7852ce014b254effb307ec5de.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/ps_human/en/038013b7852ce014b254effb307ec5de.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/038013b7852ce014b254effb307ec5de.png"], "label": "triple_ranking", "timestamp": "2025-08-29T13:54:37.808801", "instruction_following": {"ranking_text": "15|4|23"}, "consistency": {"ranking_text": "235|1|4"}, "overall": {"ranking_text": "15|4|23"}} +{"task_type": "ps_human", "data_source": "GEdit-Bench", "instruction": "Remove acne and blemishes from my face, slim down my nose and face.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/c18b9ea3a82a132108bb19942258fae1_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/ps_human/en/c18b9ea3a82a132108bb19942258fae1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/ps_human/en/c18b9ea3a82a132108bb19942258fae1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/ps_human/en/c18b9ea3a82a132108bb19942258fae1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/ps_human/en/c18b9ea3a82a132108bb19942258fae1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/ps_human/en/c18b9ea3a82a132108bb19942258fae1.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:03:14.614106", "instruction_following": {"ranking_text": "2|135|4"}, "overall": {"ranking_text": "2|13|5|4"}} +{"task_type": "subject-add", "data_source": "GEdit-Bench", "instruction": "Place a wine glass in the hand.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-add/en/9435ef3cbe961ecde654fdde42598cb1_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/subject-add/en/9435ef3cbe961ecde654fdde42598cb1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/subject-add/en/9435ef3cbe961ecde654fdde42598cb1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/subject-add/en/9435ef3cbe961ecde654fdde42598cb1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/subject-add/en/9435ef3cbe961ecde654fdde42598cb1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-add/en/9435ef3cbe961ecde654fdde42598cb1.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:05:32.825387", "instruction_following": {"ranking_text": "3|5|1|24"}, "overall": {"ranking_text": "3|5|1|24"}} +{"task_type": "subject-add", "data_source": "GEdit-Bench", "instruction": "Add two small dogs sitting face-to-face in the foreground.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-add/en/bcb9d7a80eaf8a5f630cc78b6bce0b6c_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/subject-add/en/bcb9d7a80eaf8a5f630cc78b6bce0b6c.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/subject-add/en/bcb9d7a80eaf8a5f630cc78b6bce0b6c.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/subject-add/en/bcb9d7a80eaf8a5f630cc78b6bce0b6c.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/subject-add/en/bcb9d7a80eaf8a5f630cc78b6bce0b6c.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/subject-add/en/bcb9d7a80eaf8a5f630cc78b6bce0b6c.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:09:11.870067", "instruction_following": {"ranking_text": "2|135|4"}, "consistency": {"ranking_text": "1345|2"}, "overall": {"ranking_text": "2|135|4"}} +{"task_type": "subject-add", "data_source": "GEdit-Bench", "instruction": "Can you add penguin eyes to this image?", "input_images": ["/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/subject-add/en/794bc25fba24e9c7546c7ffed818fba1_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/subject-add/en/794bc25fba24e9c7546c7ffed818fba1.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/subject-add/en/794bc25fba24e9c7546c7ffed818fba1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/subject-add/en/794bc25fba24e9c7546c7ffed818fba1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/subject-add/en/794bc25fba24e9c7546c7ffed818fba1.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/subject-add/en/794bc25fba24e9c7546c7ffed818fba1.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:11:31.493427", "instruction_following": {"ranking_text": "4|25|1|3"}, "consistency": {"ranking_text": "124|35"}, "overall": {"ranking_text": "4|2|1|5|3"}} +{"task_type": "subject-add", "data_source": "GEdit-Bench", "instruction": "Add a painting to the easel.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-add/en/61e0b78dbfbf640f62447931c8c45a9a_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/subject-add/en/61e0b78dbfbf640f62447931c8c45a9a.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/subject-add/en/61e0b78dbfbf640f62447931c8c45a9a.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/subject-add/en/61e0b78dbfbf640f62447931c8c45a9a.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/subject-add/en/61e0b78dbfbf640f62447931c8c45a9a.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-add/en/61e0b78dbfbf640f62447931c8c45a9a.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:14:21.689503", "instruction_following": {"ranking_text": "1245|3"}, "consistency": {"ranking_text": "2345|1"}, "overall": {"ranking_text": "245|13"}} +{"task_type": "subject-add", "data_source": "imgedit", "instruction": "Add a car in the foreground to the right side of the image.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/for_add/000081452.jpg"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image_imgedit/57.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/57.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/57.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/57.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/57.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:16:49.907743", "instruction_following": {"ranking_text": "2345|1"}, "consistency": {"ranking_text": "1345|2"}, "overall": {"ranking_text": "345|2|1"}} +{"task_type": "subject-add", "data_source": "imgedit", "instruction": "Add a person sitting near the small structure in the center of the image, facing the scenic landscape as if enjoying the view.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/for_add/000313906.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/109.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/109.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/109.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/109.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/109.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:19:57.803923", "instruction_following": {"ranking_text": "1|245|3"}, "consistency": {"ranking_text": "1245|3"}, "overall": {"ranking_text": "1|245|3"}} +{"task_type": "subject-add", "data_source": "imgedit", "instruction": "Add a hiker standing on one of the rocks near the river, wearing a backpack, and looking towards the mountain in the background.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/for_add/000060085.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/53.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/53.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/53.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/53.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/53.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:24:05.783230", "instruction_following": {"ranking_text": "3|245|1"}, "consistency": {"ranking_text": "2345|1"}, "overall": {"ranking_text": "3|245|1"}} +{"task_type": "subject-add", "data_source": "imgedit", "instruction": "Add a vintage suitcase inside the trunk of the car to emphasize the transport theme.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/for_add/000001314.jpg"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image_imgedit/23.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/23.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/23.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/23.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/23.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:26:29.256974", "consistency": {"ranking_text": "1234|5"}, "overall": {"ranking_text": "1234|5"}} +{"task_type": "extract", "data_source": "imgedit", "instruction": "Extract the animals present in the image.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/animal/000216869.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/imgedit/364.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/364.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/364.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/364.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/364.png"], "label": "bad", "timestamp": "2025-08-29T14:26:56.209039"} +{"task_type": "extract", "data_source": "imgedit", "instruction": "Extract the white T-shirt with a red \"Levi's\" logo worn by the person in the image.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/clothes/00000012.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/437.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/437.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/437.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/437.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/437.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:29:06.929550", "instruction_following": {"ranking_text": "1|2345"}, "consistency": {"ranking_text": "5|12|34"}, "overall": {"ranking_text": "1|2345"}} +{"task_type": "extract", "data_source": "imgedit", "instruction": "Extract the colorful striped top worn by the person in the image.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/clothes/00000051.jpg"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/450.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/450.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/450.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/450.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/450.png"], "label": "bad", "timestamp": "2025-08-29T14:29:29.128104"} +{"task_type": "extract", "data_source": "imgedit", "instruction": "Extract the animals present in the image.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/animal/000216869.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/364.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/364.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/364.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/364.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/364.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:30:46.899219", "instruction_following": {"ranking_text": "2|14|35"}, "consistency": {"ranking_text": "12|4|35"}, "overall": {"ranking_text": "2|14|35"}} +{"task_type": "extract", "data_source": "imgedit", "instruction": "Extract the black T-shirt worn by the person in the image", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/clothes/00000003.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/imgedit/434.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/434.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/434.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/434.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/434.png"], "label": "bad", "timestamp": "2025-08-29T14:31:00.384786"} +{"task_type": "extract", "data_source": "imgedit", "instruction": "Extract the helicopter flying over the waterfall in the image", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/transport/000270866.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/384.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/384.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/384.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/384.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/384.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:36:40.006282", "instruction_following": {"ranking_text": "2|345|1"}, "consistency": {"ranking_text": "23|45|1"}, "overall": {"ranking_text": "2|1345"}} +{"task_type": "extract", "data_source": "imgedit", "instruction": "Extract the daily objects visible in the image.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/daily object/000278574.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/426.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/426.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/426.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/426.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/426.png"], "label": "bad", "timestamp": "2025-08-29T14:37:04.939522"} +{"task_type": "extract", "data_source": "imgedit", "instruction": "Extract the transport object(s) in the image.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/transport/000076056.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/377.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/377.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/377.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/377.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/377.png"], "label": "bad", "timestamp": "2025-08-29T14:37:56.704512"} +{"task_type": "tone_transfer", "data_source": "GEdit-Bench", "instruction": "Apply a filter adjustment.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/3053a9287013dac68056dd7aefdced02_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/tone_transfer/en/3053a9287013dac68056dd7aefdced02.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/tone_transfer/en/3053a9287013dac68056dd7aefdced02.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/tone_transfer/en/3053a9287013dac68056dd7aefdced02.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/3053a9287013dac68056dd7aefdced02.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/tone_transfer/en/3053a9287013dac68056dd7aefdced02.png"], "label": "bad", "timestamp": "2025-08-29T14:38:20.971368"} +{"task_type": "tone_transfer", "data_source": "GEdit-Bench", "instruction": "Make the image brighter.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/e6b1eb3b883e718a85581c0d36727f24_SRCIMG.png"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image/fullset/tone_transfer/en/e6b1eb3b883e718a85581c0d36727f24.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/tone_transfer/en/e6b1eb3b883e718a85581c0d36727f24.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/tone_transfer/en/e6b1eb3b883e718a85581c0d36727f24.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/tone_transfer/en/e6b1eb3b883e718a85581c0d36727f24.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/e6b1eb3b883e718a85581c0d36727f24.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:40:59.635373", "instruction_following": {"ranking_text": "123|4|5"}, "consistency": {"ranking_text": "1245|3"}, "overall": {"ranking_text": "2|1|4|35"}} +{"task_type": "tone_transfer", "data_source": "GEdit-Bench", "instruction": "Make the image brighter.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/1db07f0d277222e32913bff2681faebb_SRCIMG.png"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image/fullset/tone_transfer/en/1db07f0d277222e32913bff2681faebb.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/tone_transfer/en/1db07f0d277222e32913bff2681faebb.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/tone_transfer/en/1db07f0d277222e32913bff2681faebb.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/tone_transfer/en/1db07f0d277222e32913bff2681faebb.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/tone_transfer/en/1db07f0d277222e32913bff2681faebb.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:43:39.913844", "instruction_following": {"ranking_text": "1245|3"}, "consistency": {"ranking_text": "234|1|5"}, "overall": {"ranking_text": "24|135"}} +{"task_type": "tone_transfer", "data_source": "GEdit-Bench", "instruction": "My photo looks a bit yellowish; please adjust the color.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/f57ec87ccf7bc1788dfd5be1da4dbe7a_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/tone_transfer/en/f57ec87ccf7bc1788dfd5be1da4dbe7a.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/tone_transfer/en/f57ec87ccf7bc1788dfd5be1da4dbe7a.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/tone_transfer/en/f57ec87ccf7bc1788dfd5be1da4dbe7a.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/tone_transfer/en/f57ec87ccf7bc1788dfd5be1da4dbe7a.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/tone_transfer/en/f57ec87ccf7bc1788dfd5be1da4dbe7a.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:47:50.783465", "instruction_following": {"ranking_text": "3|145|2"}, "consistency": {"ranking_text": "13|25|4"}, "overall": {"ranking_text": "3|145|2"}} +{"task_type": "tone_transfer", "data_source": "GEdit-Bench", "instruction": "change the weather to snow", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/82713e857fa4a3972bd3bd560ad45d70_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/tone_transfer/en/82713e857fa4a3972bd3bd560ad45d70.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/82713e857fa4a3972bd3bd560ad45d70.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/tone_transfer/en/82713e857fa4a3972bd3bd560ad45d70.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/tone_transfer/en/82713e857fa4a3972bd3bd560ad45d70.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/tone_transfer/en/82713e857fa4a3972bd3bd560ad45d70.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:50:35.570026", "consistency": {"ranking_text": "1245|3"}, "overall": {"ranking_text": "1245|3"}} +{"task_type": "tone_transfer", "data_source": "GEdit-Bench", "instruction": "Restore and colorize the image.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/e7652e4858f7d1f3b86a0de28c6cb8c1_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/tone_transfer/en/e7652e4858f7d1f3b86a0de28c6cb8c1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/tone_transfer/en/e7652e4858f7d1f3b86a0de28c6cb8c1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/tone_transfer/en/e7652e4858f7d1f3b86a0de28c6cb8c1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/e7652e4858f7d1f3b86a0de28c6cb8c1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/tone_transfer/en/e7652e4858f7d1f3b86a0de28c6cb8c1.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:51:53.587103", "instruction_following": {"ranking_text": "145|2|3"}, "consistency": {"ranking_text": "145|23"}, "overall": {"ranking_text": "145|2|3"}} +{"task_type": "tone_transfer", "data_source": "GEdit-Bench", "instruction": "change the time to nighttime", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/ad3ad5f80040286822ec035c8fcf6c0f_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/tone_transfer/en/ad3ad5f80040286822ec035c8fcf6c0f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/tone_transfer/en/ad3ad5f80040286822ec035c8fcf6c0f.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/tone_transfer/en/ad3ad5f80040286822ec035c8fcf6c0f.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/tone_transfer/en/ad3ad5f80040286822ec035c8fcf6c0f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/tone_transfer/en/ad3ad5f80040286822ec035c8fcf6c0f.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:23:05.053654", "consistency": {"ranking_text": "1245|3"}, "overall": {"ranking_text": "1245|3"}} +{"task_type": "tone_transfer", "data_source": "GEdit-Bench", "instruction": "Enhance it to super high quality.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/4b0700347e2ea2aef8f27a2cc2b9c370_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/tone_transfer/en/4b0700347e2ea2aef8f27a2cc2b9c370.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/tone_transfer/en/4b0700347e2ea2aef8f27a2cc2b9c370.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/tone_transfer/en/4b0700347e2ea2aef8f27a2cc2b9c370.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/tone_transfer/en/4b0700347e2ea2aef8f27a2cc2b9c370.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/tone_transfer/en/4b0700347e2ea2aef8f27a2cc2b9c370.png"], "label": "bad", "timestamp": "2025-08-29T15:23:42.726780"} +{"task_type": "background_change", "data_source": "GEdit-Bench", "instruction": "Adjust the background to a glass wall.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/background_change/en/f17eaba1650c7320694dd8a5493361b8_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/background_change/en/f17eaba1650c7320694dd8a5493361b8.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/background_change/en/f17eaba1650c7320694dd8a5493361b8.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/background_change/en/f17eaba1650c7320694dd8a5493361b8.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/background_change/en/f17eaba1650c7320694dd8a5493361b8.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/background_change/en/f17eaba1650c7320694dd8a5493361b8.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:25:43.751008", "instruction_following": {"ranking_text": "1245|3"}, "consistency": {"ranking_text": "35|12|4"}, "overall": {"ranking_text": "5|2|1|34"}} +{"task_type": "background_change", "data_source": "GEdit-Bench", "instruction": "Change the background to a forest.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/background_change/en/05040717fb0f2ac80083ef81ee206ace_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/background_change/en/05040717fb0f2ac80083ef81ee206ace.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/background_change/en/05040717fb0f2ac80083ef81ee206ace.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/background_change/en/05040717fb0f2ac80083ef81ee206ace.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/background_change/en/05040717fb0f2ac80083ef81ee206ace.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/background_change/en/05040717fb0f2ac80083ef81ee206ace.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:28:20.905704", "instruction_following": {"ranking_text": "145|23"}, "consistency": {"ranking_text": "124|5|3"}, "overall": {"ranking_text": "14|5|2|3"}} +{"task_type": "background_change", "data_source": "GEdit-Bench", "instruction": "Change the background to the ocean.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/background_change/en/f7d391ffa970e18fc8393888295899f8_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/background_change/en/f7d391ffa970e18fc8393888295899f8.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/background_change/en/f7d391ffa970e18fc8393888295899f8.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/background_change/en/f7d391ffa970e18fc8393888295899f8.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/background_change/en/f7d391ffa970e18fc8393888295899f8.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/background_change/en/f7d391ffa970e18fc8393888295899f8.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:29:11.591985", "overall": {"ranking_text": "1245|3"}} +{"task_type": "background_change", "data_source": "GEdit-Bench", "instruction": "Replace the sky in this image with blue skies and white clouds.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/background_change/en/165533290b7c205b0dd34d1053716dcb_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/background_change/en/165533290b7c205b0dd34d1053716dcb.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/background_change/en/165533290b7c205b0dd34d1053716dcb.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/background_change/en/165533290b7c205b0dd34d1053716dcb.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/background_change/en/165533290b7c205b0dd34d1053716dcb.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/background_change/en/165533290b7c205b0dd34d1053716dcb.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:31:14.493282", "instruction_following": {"ranking_text": "145|23"}, "consistency": {"ranking_text": "12|45|3"}, "overall": {"ranking_text": "1|45|2|3"}} +{"task_type": "background_change", "data_source": "imgedit", "instruction": "Change the snowy forest environment to a springtime forest with budding trees and wildflowers.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/architecture/000276347.jpg"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/273.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/273.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/273.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/273.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/273.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:37:35.335869", "instruction_following": {"ranking_text": "14|2|35"}, "consistency": {"ranking_text": "5|3|1|4|2"}, "overall": {"ranking_text": "5|13|4|2"}} +{"task_type": "background_change", "data_source": "imgedit", "instruction": "Change the interior environment of the image from a classic and elegant room with green chairs and ornate rugs to a modern minimalist setting with sleek furniture and a neutral color palette.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/human/000336112.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/imgedit/222.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/222.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/222.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/222.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/222.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:41:17.537420", "instruction_following": {"ranking_text": "2|35|1|4"}, "consistency": {"ranking_text": "5|2|1|34"}, "overall": {"ranking_text": "2|5|1|34"}} +{"task_type": "background_change", "data_source": "imgedit", "instruction": "Change the castle in the picture from the rocky landscape to a lush tropical forest", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/architecture/000280642.jpg"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/274.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/274.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/274.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/274.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/274.png"], "label": "bad", "timestamp": "2025-08-29T15:42:11.758575"} +{"task_type": "background_change", "data_source": "imgedit", "instruction": "Change the racetrack in the picture from an asphalt circuit to a desert track.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/transport/000210129.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/252.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/252.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/252.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/252.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/252.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:44:26.086955", "instruction_following": {"ranking_text": "1|234|5"}, "consistency": {"ranking_text": "134|2|5"}, "overall": {"ranking_text": "1|24|35"}} +{"task_type": "subject-remove", "data_source": "GEdit-Bench", "instruction": "Remove the elderly man wearing glasses", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-remove/en/8853ec3095105930363c6c8c988f55e6_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/subject-remove/en/8853ec3095105930363c6c8c988f55e6.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/subject-remove/en/8853ec3095105930363c6c8c988f55e6.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/subject-remove/en/8853ec3095105930363c6c8c988f55e6.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-remove/en/8853ec3095105930363c6c8c988f55e6.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/subject-remove/en/8853ec3095105930363c6c8c988f55e6.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:45:56.190738", "instruction_following": {"ranking_text": "34|125"}, "consistency": {"ranking_text": "145|3|2"}, "overall": {"ranking_text": "4|3|125"}} +{"task_type": "subject-remove", "data_source": "GEdit-Bench", "instruction": "remove the freight train", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-remove/en/fc610a23a5c9ac5c4a3c2cc0386bc8d2_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/subject-remove/en/fc610a23a5c9ac5c4a3c2cc0386bc8d2.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/subject-remove/en/fc610a23a5c9ac5c4a3c2cc0386bc8d2.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-remove/en/fc610a23a5c9ac5c4a3c2cc0386bc8d2.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/subject-remove/en/fc610a23a5c9ac5c4a3c2cc0386bc8d2.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/subject-remove/en/fc610a23a5c9ac5c4a3c2cc0386bc8d2.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:49:37.091962", "instruction_following": {"ranking_text": "23|1|45"}, "consistency": {"ranking_text": "13|45|2"}, "overall": {"ranking_text": "3|12|45"}} +{"task_type": "subject-remove", "data_source": "GEdit-Bench", "instruction": "Remove the sticky notes next to the monitor.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-remove/en/f5d8129b33eaf3adcaad19ba2d471529_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/subject-remove/en/f5d8129b33eaf3adcaad19ba2d471529.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/subject-remove/en/f5d8129b33eaf3adcaad19ba2d471529.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/subject-remove/en/f5d8129b33eaf3adcaad19ba2d471529.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/subject-remove/en/f5d8129b33eaf3adcaad19ba2d471529.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/subject-remove/en/f5d8129b33eaf3adcaad19ba2d471529.png"], "label": "bad", "timestamp": "2025-08-29T15:50:17.482439"} +{"task_type": "subject-remove", "data_source": "GEdit-Bench", "instruction": "Remove the elderly man wearing glasses", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-remove/en/8853ec3095105930363c6c8c988f55e6_SRCIMG.png"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image/fullset/subject-remove/en/8853ec3095105930363c6c8c988f55e6.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/subject-remove/en/8853ec3095105930363c6c8c988f55e6.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/subject-remove/en/8853ec3095105930363c6c8c988f55e6.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/subject-remove/en/8853ec3095105930363c6c8c988f55e6.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/subject-remove/en/8853ec3095105930363c6c8c988f55e6.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:52:04.813649", "instruction_following": {"ranking_text": "14|235"}, "consistency": {"ranking_text": "5|14|23"}, "overall": {"ranking_text": "14|235"}} +{"task_type": "subject-remove", "data_source": "imgedit", "instruction": "Remove the human sitting on the bench in the foreground.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/human/000352814.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/475.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/475.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/475.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/475.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/475.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:53:31.447451", "instruction_following": {"ranking_text": "45|23|1"}, "overall": {"ranking_text": "45|23|1"}} +{"task_type": "subject-remove", "data_source": "imgedit", "instruction": "Remove the colorful inflatable bounce house with slides in the foreground.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/architecture/000186995.jpg"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image_imgedit/516.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/516.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/516.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/516.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/516.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:55:51.365448", "instruction_following": {"ranking_text": "4|1|2|53"}, "overall": {"ranking_text": "4|12|5|3"}} +{"task_type": "subject-remove", "data_source": "imgedit", "instruction": "Remove the person from the image.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/human/000286285.jpg"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/468.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/468.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/468.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/468.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/468.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:04:41.933288", "instruction_following": {"ranking_text": "2|134|5"}, "consistency": {"ranking_text": "1245|3"}, "overall": {"ranking_text": "2|14|35"}} +{"task_type": "subject-remove", "data_source": "imgedit", "instruction": "Remove the animal from the image.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/animal/000069528.jpg"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/482.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/482.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/482.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/482.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/482.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:06:09.964185", "instruction_following": {"ranking_text": "1345|2"}, "consistency": {"ranking_text": "1234|5"}, "overall": {"ranking_text": "134|5|2"}} +{"task_type": "compose", "data_source": "imgedit", "instruction": "Remove the plant on the right side of the image, and adjust the man's suit to a darker shade of blue.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/compose/human/6.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/1155.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/1155.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1155.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/1155.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/1155.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:08:50.309060", "instruction_following": {"ranking_text": "245|1|3"}, "consistency": {"ranking_text": "245|1|3"}, "overall": {"ranking_text": "245|1|3"}} +{"task_type": "compose", "data_source": "imgedit", "instruction": "Remove the object on the left side of the image, and adjust the brightness of the background.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/compose/objects/8.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/1172.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1172.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1172.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/1172.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/1172.png"], "label": "bad", "timestamp": "2025-08-29T16:09:21.915708"} +{"task_type": "compose", "data_source": "imgedit", "instruction": "Remove the gift box in the air held by the woman on the right, and adjust the lighting to brighten the scene.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/compose/human/14.jpg"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/1149.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/1149.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1149.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/1149.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1149.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:11:55.633463", "instruction_following": {"ranking_text": "45|2|1|3"}, "consistency": {"ranking_text": "145|2|3"}, "overall": {"ranking_text": "45|2|13"}} +{"task_type": "compose", "data_source": "imgedit", "instruction": "Remove the books from the shelf, and change the color of the couch pillow to blue.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/compose/human/8.jpg"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1157.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/1157.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/1157.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/1157.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/1157.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:16:44.319706", "instruction_following": {"ranking_text": "13|245"}, "consistency": {"ranking_text": "123|45"}, "overall": {"ranking_text": "1|35|24"}} +{"task_type": "compose", "data_source": "imgedit", "instruction": "Remove the object on the left side of the image, and adjust the lighting to brighten the right side.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/compose/human/4.jpg"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1153.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/1153.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1153.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/1153.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/1153.png"], "label": "bad", "timestamp": "2025-08-29T16:17:41.069059"} +{"task_type": "compose", "data_source": "imgedit", "instruction": "Remove the pink cushion on the ground, and change the drink in the woman's hand on the right to a green beverage.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/compose/human/15.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1150.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/1150.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/1150.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/1150.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/1150.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:20:56.571743", "instruction_following": {"ranking_text": "5|234|1"}, "consistency": {"ranking_text": "1345|2"}, "overall": {"ranking_text": "5|23|14"}} +{"task_type": "compose", "data_source": "imgedit", "instruction": "Remove the basket of fruit on the coffee table, and change the color of the left armchair cushion to dark green.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/compose/objects/11.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1164.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/1164.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1164.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/1164.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/1164.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:24:45.475580", "instruction_following": {"ranking_text": "4|25|13"}, "consistency": {"ranking_text": "45|3|12"}, "overall": {"ranking_text": "4|5|2|13"}} +{"task_type": "compose", "data_source": "imgedit", "instruction": "Remove the object in the top left corner of the image, and resize the object in the bottom right to be larger.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/compose/objects/4.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/1170.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/1170.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/1170.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/1170.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1170.png"], "label": "bad", "timestamp": "2025-08-29T16:29:46.429151"} +{"task_type": "style_change", "data_source": "GEdit-Bench", "instruction": "Adjust the image style to a watercolor effect.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/style_change/en/4e62777f17329aff2906ff86a217ccb6_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/style_change/en/4e62777f17329aff2906ff86a217ccb6.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/style_change/en/4e62777f17329aff2906ff86a217ccb6.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/style_change/en/4e62777f17329aff2906ff86a217ccb6.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/style_change/en/4e62777f17329aff2906ff86a217ccb6.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gpt_image_1/results/fullset/style_change/en/4e62777f17329aff2906ff86a217ccb6.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:30:54.929408", "instruction_following": {"ranking_text": "1345|2"}, "consistency": {"ranking_text": "134|5|2"}, "overall": {"ranking_text": "134|5|2"}} +{"task_type": "style_change", "data_source": "GEdit-Bench", "instruction": "Turn the image into an American comic style.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/style_change/en/07fc2fa9b1bbee0e9e37421fe3a6576b_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/style_change/en/07fc2fa9b1bbee0e9e37421fe3a6576b.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/style_change/en/07fc2fa9b1bbee0e9e37421fe3a6576b.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/style_change/en/07fc2fa9b1bbee0e9e37421fe3a6576b.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/style_change/en/07fc2fa9b1bbee0e9e37421fe3a6576b.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/style_change/en/07fc2fa9b1bbee0e9e37421fe3a6576b.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:31:41.587938", "instruction_following": {"ranking_text": "2345|1"}, "consistency": {"ranking_text": "2345|1"}, "overall": {"ranking_text": "2345|1"}} +{"task_type": "style_change", "data_source": "GEdit-Bench", "instruction": "Generate a monochrome-style animation.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/style_change/en/a50fdf85f87b7a11acc92335eaba1b6c_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/style_change/en/a50fdf85f87b7a11acc92335eaba1b6c.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/style_change/en/a50fdf85f87b7a11acc92335eaba1b6c.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/style_change/en/a50fdf85f87b7a11acc92335eaba1b6c.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/style_change/en/a50fdf85f87b7a11acc92335eaba1b6c.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/style_change/en/a50fdf85f87b7a11acc92335eaba1b6c.png"], "label": "bad", "timestamp": "2025-08-29T16:33:27.200460"} +{"task_type": "style_change", "data_source": "GEdit-Bench", "instruction": "Redraw it as a chibi-style illustration.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/style_change/en/ae112c98cae0bfd203af4da8ee3ad54f_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/style_change/en/ae112c98cae0bfd203af4da8ee3ad54f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/style_change/en/ae112c98cae0bfd203af4da8ee3ad54f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/style_change/en/ae112c98cae0bfd203af4da8ee3ad54f.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/style_change/en/ae112c98cae0bfd203af4da8ee3ad54f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/style_change/en/ae112c98cae0bfd203af4da8ee3ad54f.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:36:10.009381", "consistency": {"ranking_text": "5|1|24|3"}, "overall": {"ranking_text": "5|1|24|3"}} +{"task_type": "style_change", "data_source": "imgedit", "instruction": "Transfer the image into a faceted low-poly 3-D render style.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/style/000308736.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/imgedit/724.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/724.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/724.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/724.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/724.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:37:51.527014", "instruction_following": {"ranking_text": "1234|5"}, "consistency": {"ranking_text": "1234|5"}, "overall": {"ranking_text": "23|4|1|5"}} +{"task_type": "style_change", "data_source": "imgedit", "instruction": "Transfer the image into a stained-glass cathedral-window style.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/style/000308736.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/719.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/719.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/719.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/719.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/719.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:40:10.434931", "instruction_following": {"ranking_text": "3|4|25|1"}, "consistency": {"ranking_text": "1234|5"}, "overall": {"ranking_text": "3|4|25|1"}} +{"task_type": "style_change", "data_source": "imgedit", "instruction": "Transfer the image into a clean graphite pencil-sketch style.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/style/000308736.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/imgedit/717.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/717.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/717.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/717.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/717.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:44:13.860746", "instruction_following": {"ranking_text": "2345|1"}, "consistency": {"ranking_text": "15|3|24"}, "overall": {"ranking_text": "5|3|24|1"}} +{"task_type": "style_change", "data_source": "imgedit", "instruction": "Transfer the image into a Lego-brick stop-motion diorama style.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/style/000015317.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/imgedit/642.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/642.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/642.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/642.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/642.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:49:12.301176", "instruction_following": {"ranking_text": "2345|1"}, "consistency": {"ranking_text": "3|4|125"}, "overall": {"ranking_text": "3|4|125"}} +{"task_type": "material_alter", "data_source": "GEdit-Bench", "instruction": "Transform the clothing material into silk.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/641f39026c89fffaf60a4f0f50304d7d_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/material_alter/en/641f39026c89fffaf60a4f0f50304d7d.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/material_alter/en/641f39026c89fffaf60a4f0f50304d7d.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/material_alter/en/641f39026c89fffaf60a4f0f50304d7d.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/641f39026c89fffaf60a4f0f50304d7d.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/material_alter/en/641f39026c89fffaf60a4f0f50304d7d.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:51:26.786483", "instruction_following": {"ranking_text": "35|124"}, "consistency": {"ranking_text": "1345|2"}, "overall": {"ranking_text": "35|124"}} +{"task_type": "material_alter", "data_source": "GEdit-Bench", "instruction": "Craft the cat using cloisonné enamel.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/5dbff1a3b7d1fb890b72cef2f711a2ac_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/material_alter/en/5dbff1a3b7d1fb890b72cef2f711a2ac.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/material_alter/en/5dbff1a3b7d1fb890b72cef2f711a2ac.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/5dbff1a3b7d1fb890b72cef2f711a2ac.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/material_alter/en/5dbff1a3b7d1fb890b72cef2f711a2ac.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/material_alter/en/5dbff1a3b7d1fb890b72cef2f711a2ac.png"], "label": "bad", "timestamp": "2025-08-29T16:52:22.052182"} +{"task_type": "material_alter", "data_source": "GEdit-Bench", "instruction": "Make the clothing fabric from premium linen.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/6db0677c1fa5b1a266e9c078d5cb175d_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/material_alter/en/6db0677c1fa5b1a266e9c078d5cb175d.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/material_alter/en/6db0677c1fa5b1a266e9c078d5cb175d.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/material_alter/en/6db0677c1fa5b1a266e9c078d5cb175d.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/material_alter/en/6db0677c1fa5b1a266e9c078d5cb175d.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/material_alter/en/6db0677c1fa5b1a266e9c078d5cb175d.png"], "label": "bad", "timestamp": "2025-08-29T16:52:52.269774"} +{"task_type": "material_alter", "data_source": "GEdit-Bench", "instruction": "Craft the outerwear from full-grain calfskin leather.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/66fbc2d25acbb4b6542ba627c365bd4f_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/material_alter/en/66fbc2d25acbb4b6542ba627c365bd4f.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/material_alter/en/66fbc2d25acbb4b6542ba627c365bd4f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/material_alter/en/66fbc2d25acbb4b6542ba627c365bd4f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/material_alter/en/66fbc2d25acbb4b6542ba627c365bd4f.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/material_alter/en/66fbc2d25acbb4b6542ba627c365bd4f.png"], "label": "bad", "timestamp": "2025-08-29T16:58:59.488072"} +{"task_type": "material_alter", "data_source": "GEdit-Bench", "instruction": "Craft the outerwear from lambskin leather.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/803f665220970a442a420afb826b6747_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/803f665220970a442a420afb826b6747.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/material_alter/en/803f665220970a442a420afb826b6747.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/material_alter/en/803f665220970a442a420afb826b6747.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/material_alter/en/803f665220970a442a420afb826b6747.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/material_alter/en/803f665220970a442a420afb826b6747.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:00:23.774956", "instruction_following": {"ranking_text": "1|2345"}, "consistency": {"ranking_text": "1|2345"}, "overall": {"ranking_text": "1|2345"}} +{"task_type": "material_alter", "data_source": "GEdit-Bench", "instruction": "Swap the background plants for woven rattan.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/73875335f42e4154ece47b4a4fafd83e_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/material_alter/en/73875335f42e4154ece47b4a4fafd83e.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/material_alter/en/73875335f42e4154ece47b4a4fafd83e.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/material_alter/en/73875335f42e4154ece47b4a4fafd83e.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/material_alter/en/73875335f42e4154ece47b4a4fafd83e.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/material_alter/en/73875335f42e4154ece47b4a4fafd83e.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:04:43.753197", "instruction_following": {"ranking_text": "15|2|34"}, "consistency": {"ranking_text": "245|13"}, "overall": {"ranking_text": "5|1234"}} +{"task_type": "material_alter", "data_source": "GEdit-Bench", "instruction": "Craft the outerwear from full-grain calfskin leather.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/66fbc2d25acbb4b6542ba627c365bd4f_SRCIMG.png"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image/fullset/material_alter/en/66fbc2d25acbb4b6542ba627c365bd4f.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/material_alter/en/66fbc2d25acbb4b6542ba627c365bd4f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/material_alter/en/66fbc2d25acbb4b6542ba627c365bd4f.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/material_alter/en/66fbc2d25acbb4b6542ba627c365bd4f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/66fbc2d25acbb4b6542ba627c365bd4f.png"], "label": "bad", "timestamp": "2025-08-29T17:05:01.000065"} +{"task_type": "material_alter", "data_source": "GEdit-Bench", "instruction": "Reconstruct the bus body with solid wood panels.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/9d76287b0d48bcff3cdff69b198f569e_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/material_alter/en/9d76287b0d48bcff3cdff69b198f569e.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/material_alter/en/9d76287b0d48bcff3cdff69b198f569e.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/material_alter/en/9d76287b0d48bcff3cdff69b198f569e.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/material_alter/en/9d76287b0d48bcff3cdff69b198f569e.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/9d76287b0d48bcff3cdff69b198f569e.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:07:59.839220", "instruction_following": {"ranking_text": "2345|1"}, "consistency": {"ranking_text": "3|45|2|1"}, "overall": {"ranking_text": "3|45|2|1"}} +{"task_type": "color_alter", "data_source": "GEdit-Bench", "instruction": "change the color of goat to yellow", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/color_alter/en/981afb942cdf3cbacf7614f47ff21b2d_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/color_alter/en/981afb942cdf3cbacf7614f47ff21b2d.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/color_alter/en/981afb942cdf3cbacf7614f47ff21b2d.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/color_alter/en/981afb942cdf3cbacf7614f47ff21b2d.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/color_alter/en/981afb942cdf3cbacf7614f47ff21b2d.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/color_alter/en/981afb942cdf3cbacf7614f47ff21b2d.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:09:03.752395", "instruction_following": {"ranking_text": "2345|1"}, "consistency": {"ranking_text": "2345|1"}, "overall": {"ranking_text": "2345|1"}} +{"task_type": "color_alter", "data_source": "GEdit-Bench", "instruction": "alter the color of doughnut to silver", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/color_alter/en/41fbe7550d337d07d030b308f2099d1f_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/color_alter/en/41fbe7550d337d07d030b308f2099d1f.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/color_alter/en/41fbe7550d337d07d030b308f2099d1f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/color_alter/en/41fbe7550d337d07d030b308f2099d1f.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/color_alter/en/41fbe7550d337d07d030b308f2099d1f.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/color_alter/en/41fbe7550d337d07d030b308f2099d1f.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:12:06.338927", "instruction_following": {"ranking_text": "3|1245"}, "consistency": {"ranking_text": "1235|4"}, "overall": {"ranking_text": "3|125|4"}} +{"task_type": "color_alter", "data_source": "GEdit-Bench", "instruction": "Change the tablecloth color to bright red.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/color_alter/en/69d1ef2ac7a987ce31e0aa2d9e96beea_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/color_alter/en/69d1ef2ac7a987ce31e0aa2d9e96beea.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/color_alter/en/69d1ef2ac7a987ce31e0aa2d9e96beea.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/color_alter/en/69d1ef2ac7a987ce31e0aa2d9e96beea.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/color_alter/en/69d1ef2ac7a987ce31e0aa2d9e96beea.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/color_alter/en/69d1ef2ac7a987ce31e0aa2d9e96beea.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:12:49.811380", "instruction_following": {"ranking_text": "1345|2"}, "consistency": {"ranking_text": "1345|2"}, "overall": {"ranking_text": "1345|2"}} +{"task_type": "color_alter", "data_source": "GEdit-Bench", "instruction": "alter the color of plane to pink", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/color_alter/en/fb71870e760822d8674699ceb7034449_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/color_alter/en/fb71870e760822d8674699ceb7034449.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/color_alter/en/fb71870e760822d8674699ceb7034449.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/color_alter/en/fb71870e760822d8674699ceb7034449.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/color_alter/en/fb71870e760822d8674699ceb7034449.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/color_alter/en/fb71870e760822d8674699ceb7034449.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:14:11.711083", "instruction_following": {"ranking_text": "45|123"}, "consistency": {"ranking_text": "345|12"}, "overall": {"ranking_text": "45|3|12"}} +{"task_type": "color_alter", "data_source": "imgedit", "instruction": "Change the yacht's hull color to navy blue.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/transport/000156391.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/imgedit/1090.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/1090.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1090.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1090.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/1090.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:16:44.364218", "instruction_following": {"ranking_text": "1235|4"}, "consistency": {"ranking_text": "123|5|4"}, "overall": {"ranking_text": "123|5|4"}} +{"task_type": "color_alter", "data_source": "imgedit", "instruction": "Change the surface color of the mug to red.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/daily object/000109381.jpg"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1130.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/1130.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/1130.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/1130.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/1130.png"], "label": "bad", "timestamp": "2025-08-29T17:17:53.672513"} +{"task_type": "color_alter", "data_source": "imgedit", "instruction": "Change the color of the cup to blue.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/daily object/000368863.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/1142.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1142.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/1142.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/1142.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1142.png"], "label": "bad", "timestamp": "2025-08-29T17:18:24.137314"} +{"task_type": "color_alter", "data_source": "imgedit", "instruction": "Change the building's facade color to a lighter shade of blue.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/architecture/000369231.jpg"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/1120.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/1120.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1120.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/1120.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/1120.png"], "label": "bad", "timestamp": "2025-08-29T17:19:33.348041"} +{"task_type": "subject-replace", "data_source": "GEdit-Bench", "instruction": "Dress the person in the image in a Zhongshan suit.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-replace/en/b993cccb7c8b197175226e397a0f09a8_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/subject-replace/en/b993cccb7c8b197175226e397a0f09a8.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/subject-replace/en/b993cccb7c8b197175226e397a0f09a8.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-replace/en/b993cccb7c8b197175226e397a0f09a8.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/subject-replace/en/b993cccb7c8b197175226e397a0f09a8.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/subject-replace/en/b993cccb7c8b197175226e397a0f09a8.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:21:07.871305", "instruction_following": {"ranking_text": "3|1245"}, "overall": {"ranking_text": "3|1245"}} +{"task_type": "subject-replace", "data_source": "GEdit-Bench", "instruction": "Replace the person in the mirror wearing a white shirt with a wardrobe.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-replace/en/0064d30c8f40ddd94fa9bc564677498e_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/subject-replace/en/0064d30c8f40ddd94fa9bc564677498e.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/subject-replace/en/0064d30c8f40ddd94fa9bc564677498e.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/subject-replace/en/0064d30c8f40ddd94fa9bc564677498e.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/subject-replace/en/0064d30c8f40ddd94fa9bc564677498e.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/subject-replace/en/0064d30c8f40ddd94fa9bc564677498e.png"], "label": "bad", "timestamp": "2025-08-29T17:22:11.319376"} +{"task_type": "subject-replace", "data_source": "GEdit-Bench", "instruction": "Replace the cat with a fish.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-replace/en/1eaed2671a534749a7c6a02a3d3e5f82_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/subject-replace/en/1eaed2671a534749a7c6a02a3d3e5f82.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/subject-replace/en/1eaed2671a534749a7c6a02a3d3e5f82.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/subject-replace/en/1eaed2671a534749a7c6a02a3d3e5f82.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/subject-replace/en/1eaed2671a534749a7c6a02a3d3e5f82.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-replace/en/1eaed2671a534749a7c6a02a3d3e5f82.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:47:34.867504", "consistency": {"ranking_text": "2345|1"}, "overall": {"ranking_text": "2345|1"}} +{"task_type": "subject-replace", "data_source": "imgedit", "instruction": "Replace the motorcycle in the image with a grand piano.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/transport/000045228.jpg"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/581.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/581.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/581.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/581.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/581.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:51:13.826106", "instruction_following": {"ranking_text": "1245|3"}, "consistency": {"ranking_text": "45|3|12"}, "overall": {"ranking_text": "45|12|3"}} +{"task_type": "subject-replace", "data_source": "imgedit", "instruction": "Replace the child in the image with a large pumpkin. ", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/human/000286285.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/554.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/554.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/554.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/554.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/554.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:53:05.843958", "instruction_following": {"ranking_text": "1345|2"}, "consistency": {"ranking_text": "2345|1"}, "overall": {"ranking_text": "345|1|2"}} +{"task_type": "subject-replace", "data_source": "imgedit", "instruction": "Replace the yacht in the image with a hot air balloon floating just above the ocean surface.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/transport/000156391.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/586.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/586.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/586.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/586.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/586.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:54:40.527110", "instruction_following": {"ranking_text": "123|5|4"}, "overall": {"ranking_text": "123|5|4"}} +{"task_type": "subject-replace", "data_source": "imgedit", "instruction": "Replace the red trolley car in the image with a giant vintage radio, keeping it on the same train tracks in the urban environment.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/transport/000351716.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/597.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/597.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/597.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/597.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/597.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:56:47.496306", "instruction_following": {"ranking_text": "1345|2"}, "consistency": {"ranking_text": "1345|2"}, "overall": {"ranking_text": "1345|2"}} +{"task_type": "motion_change", "data_source": "GEdit-Bench", "instruction": "change the action of cat to jumping", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/motion_change/en/76a4af36b318953c8054fdd706e7294f_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/motion_change/en/76a4af36b318953c8054fdd706e7294f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/motion_change/en/76a4af36b318953c8054fdd706e7294f.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/motion_change/en/76a4af36b318953c8054fdd706e7294f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/motion_change/en/76a4af36b318953c8054fdd706e7294f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/motion_change/en/76a4af36b318953c8054fdd706e7294f.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:58:29.422945", "instruction_following": {"ranking_text": "345|1|2"}, "overall": {"ranking_text": "345|1|2"}} +{"task_type": "motion_change", "data_source": "GEdit-Bench", "instruction": "change the action of the horses to galloping", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/motion_change/en/8175d438e57f213c80425595063d053a_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/motion_change/en/8175d438e57f213c80425595063d053a.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/motion_change/en/8175d438e57f213c80425595063d053a.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/motion_change/en/8175d438e57f213c80425595063d053a.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/motion_change/en/8175d438e57f213c80425595063d053a.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/motion_change/en/8175d438e57f213c80425595063d053a.png"], "label": "triple_ranking", "timestamp": "2025-08-29T18:00:40.702973", "instruction_following": {"ranking_text": "15|234"}, "consistency": {"ranking_text": "12|5|4|3"}, "overall": {"ranking_text": "15|234"}} +{"task_type": "motion_change", "data_source": "GEdit-Bench", "instruction": "Make the person in the image wave.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/motion_change/en/fc228a38f175cad001bc8a409c76e63b_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/motion_change/en/fc228a38f175cad001bc8a409c76e63b.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/motion_change/en/fc228a38f175cad001bc8a409c76e63b.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/motion_change/en/fc228a38f175cad001bc8a409c76e63b.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/motion_change/en/fc228a38f175cad001bc8a409c76e63b.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/motion_change/en/fc228a38f175cad001bc8a409c76e63b.png"], "label": "triple_ranking", "timestamp": "2025-08-29T18:02:41.533171", "instruction_following": {"ranking_text": "24|15|3"}, "consistency": {"ranking_text": "1345|2"}, "overall": {"ranking_text": "24|15|3"}} +{"task_type": "motion_change", "data_source": "GEdit-Bench", "instruction": "Make the person in the image give a thumbs-up.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/motion_change/en/65368cbdae17f7c44cd4d8d1271f0bdf_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/motion_change/en/65368cbdae17f7c44cd4d8d1271f0bdf.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/motion_change/en/65368cbdae17f7c44cd4d8d1271f0bdf.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/motion_change/en/65368cbdae17f7c44cd4d8d1271f0bdf.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/motion_change/en/65368cbdae17f7c44cd4d8d1271f0bdf.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/motion_change/en/65368cbdae17f7c44cd4d8d1271f0bdf.png"], "label": "triple_ranking", "timestamp": "2025-08-29T18:04:44.205570", "instruction_following": {"ranking_text": "1345|2"}, "consistency": {"ranking_text": "145|3|2"}, "overall": {"ranking_text": "14|3|5|2"}} +{"task_type": "motion_change", "data_source": "imgedit", "instruction": "Raise the person's right arm.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/clothes/00000066.jpg"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1043.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/1043.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/1043.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/1043.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/1043.png"], "label": "triple_ranking", "timestamp": "2025-08-29T18:07:22.869475", "instruction_following": {"ranking_text": "2|5|14|3"}, "consistency": {"ranking_text": "1|245|3"}, "overall": {"ranking_text": "2|5|14|3"}} +{"task_type": "motion_change", "data_source": "imgedit", "instruction": "Raise the person's left arm.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/clothes/00000054.jpg"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/1039.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1039.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/1039.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1039.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/1039.png"], "label": "triple_ranking", "timestamp": "2025-08-29T18:09:31.067403", "instruction_following": {"ranking_text": "2|135|4"}, "consistency": {"ranking_text": "1|235|4"}, "overall": {"ranking_text": "2|135|4"}} +{"task_type": "motion_change", "data_source": "imgedit", "instruction": "Raise the person's left arm.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/clothes/00000009.jpg"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/1024.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/1024.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/1024.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1024.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1024.png"], "label": "bad", "timestamp": "2025-08-29T18:09:47.258323"} +{"task_type": "motion_change", "data_source": "imgedit", "instruction": "Make the person raise her right arm.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/clothes/00000003.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1022.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/1022.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/1022.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/1022.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/1022.png"], "label": "triple_ranking", "timestamp": "2025-08-29T18:12:05.086196", "instruction_following": {"ranking_text": "35|4|12"}, "consistency": {"ranking_text": "4|35|12"}, "overall": {"ranking_text": "35|4|12"}} +{"task_type": "text_change", "data_source": "GEdit-Bench", "instruction": "Replace the text 'NIPS' with 'CVPR'", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/0cf1208c4cfe6b460aaa6c4e01af30a3_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/text_change/en/0cf1208c4cfe6b460aaa6c4e01af30a3.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/text_change/en/0cf1208c4cfe6b460aaa6c4e01af30a3.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/text_change/en/0cf1208c4cfe6b460aaa6c4e01af30a3.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/text_change/en/0cf1208c4cfe6b460aaa6c4e01af30a3.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/0cf1208c4cfe6b460aaa6c4e01af30a3.png"], "label": "triple_ranking", "timestamp": "2025-08-29T20:08:03.343458", "instruction_following": {"ranking_text": "245|13"}, "consistency": {"ranking_text": "25|134"}, "overall": {"ranking_text": "25|4|13"}} +{"task_type": "text_change", "data_source": "GEdit-Bench", "instruction": "Add \"超级大\" before \"探秘\"", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/8ed283fe0c51659c06fd1de14420b544_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/text_change/en/8ed283fe0c51659c06fd1de14420b544.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/text_change/en/8ed283fe0c51659c06fd1de14420b544.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/text_change/en/8ed283fe0c51659c06fd1de14420b544.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/text_change/en/8ed283fe0c51659c06fd1de14420b544.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/text_change/en/8ed283fe0c51659c06fd1de14420b544.png"], "label": "bad", "timestamp": "2025-08-29T20:08:52.184627"} +{"task_type": "text_change", "data_source": "GEdit-Bench", "instruction": "Remove the text from the image", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/b43f20f42dbb3e169fabf75289627f98_SRCIMG.png"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image/fullset/text_change/en/b43f20f42dbb3e169fabf75289627f98.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/text_change/en/b43f20f42dbb3e169fabf75289627f98.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/text_change/en/b43f20f42dbb3e169fabf75289627f98.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/text_change/en/b43f20f42dbb3e169fabf75289627f98.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/text_change/en/b43f20f42dbb3e169fabf75289627f98.png"], "label": "triple_ranking", "timestamp": "2025-08-29T20:11:09.561785", "consistency": {"ranking_text": "14|2|3|5"}, "overall": {"ranking_text": "14|2|3|5"}} +{"task_type": "text_change", "data_source": "GEdit-Bench", "instruction": "Change the text 'COST' to 'FREE'", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/5efb5e25b477ca8d0f98f7774cb28ce0_SRCIMG.png"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image/fullset/text_change/en/5efb5e25b477ca8d0f98f7774cb28ce0.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/text_change/en/5efb5e25b477ca8d0f98f7774cb28ce0.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/text_change/en/5efb5e25b477ca8d0f98f7774cb28ce0.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/text_change/en/5efb5e25b477ca8d0f98f7774cb28ce0.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/text_change/en/5efb5e25b477ca8d0f98f7774cb28ce0.png"], "label": "triple_ranking", "timestamp": "2025-08-29T20:13:33.766035", "instruction_following": {"ranking_text": "1235|4"}, "consistency": {"ranking_text": "1|23|5|4"}, "overall": {"ranking_text": "1|23|5|4"}} +{"task_type": "text_change", "data_source": "GEdit-Bench", "instruction": "Replace the text 'SPA' with 'Relaxation Oasis'", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/f58bb1fd98acc1888a7272d0d0f4f2a7_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/text_change/en/f58bb1fd98acc1888a7272d0d0f4f2a7.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/text_change/en/f58bb1fd98acc1888a7272d0d0f4f2a7.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/text_change/en/f58bb1fd98acc1888a7272d0d0f4f2a7.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/text_change/en/f58bb1fd98acc1888a7272d0d0f4f2a7.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/text_change/en/f58bb1fd98acc1888a7272d0d0f4f2a7.png"], "label": "bad", "timestamp": "2025-08-29T20:13:59.490697"} +{"task_type": "text_change", "data_source": "GEdit-Bench", "instruction": "Write \"我们欢迎你\" in the bottom left corner", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/58cf4f4a16cd16ffef55c170804be136_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/text_change/en/58cf4f4a16cd16ffef55c170804be136.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/text_change/en/58cf4f4a16cd16ffef55c170804be136.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/text_change/en/58cf4f4a16cd16ffef55c170804be136.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gpt_image_1/results/fullset/text_change/en/58cf4f4a16cd16ffef55c170804be136.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/text_change/en/58cf4f4a16cd16ffef55c170804be136.png"], "label": "triple_ranking", "timestamp": "2025-08-29T20:16:18.095888", "instruction_following": {"ranking_text": "45|12|3"}, "consistency": {"ranking_text": "125|4|3"}, "overall": {"ranking_text": "5|4|12|3"}} +{"task_type": "text_change", "data_source": "GEdit-Bench", "instruction": "Write \"我们欢迎你\" in the bottom left corner", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/58cf4f4a16cd16ffef55c170804be136_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/text_change/en/58cf4f4a16cd16ffef55c170804be136.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/text_change/en/58cf4f4a16cd16ffef55c170804be136.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/text_change/en/58cf4f4a16cd16ffef55c170804be136.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/text_change/en/58cf4f4a16cd16ffef55c170804be136.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/text_change/en/58cf4f4a16cd16ffef55c170804be136.png"], "label": "triple_ranking", "timestamp": "2025-08-29T20:17:38.590218", "instruction_following": {"ranking_text": "25|134"}, "consistency": {"ranking_text": "25|14|3"}, "overall": {"ranking_text": "25|14|3"}} +{"task_type": "text_change", "data_source": "GEdit-Bench", "instruction": "Change the text 'SNP' to 'Call me'", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/de1a98de20909a104b97fc444fff100d_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/de1a98de20909a104b97fc444fff100d.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/text_change/en/de1a98de20909a104b97fc444fff100d.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/text_change/en/de1a98de20909a104b97fc444fff100d.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/text_change/en/de1a98de20909a104b97fc444fff100d.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/text_change/en/de1a98de20909a104b97fc444fff100d.png"], "label": "triple_ranking", "timestamp": "2025-08-29T20:19:16.540781", "instruction_following": {"ranking_text": "1|234|5"}, "consistency": {"ranking_text": "1|34|25"}, "overall": {"ranking_text": "1|34|2|5"}} +{"task_type": "subject-replace", "data_source": "GEdit-Bench", "instruction": "Replace the vase with a sculpture.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-replace/en/99303b3de6bf596a47ecf83b25b08db5_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/subject-replace/en/99303b3de6bf596a47ecf83b25b08db5.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/subject-replace/en/99303b3de6bf596a47ecf83b25b08db5.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/subject-replace/en/99303b3de6bf596a47ecf83b25b08db5.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/subject-replace/en/99303b3de6bf596a47ecf83b25b08db5.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/subject-replace/en/99303b3de6bf596a47ecf83b25b08db5.png"], "label": "bad", "timestamp": "2025-08-29T20:20:17.604330"}