diff --git "a/jsonl/Edit-Reward-Bench/labeled_data/0829_labels_104-208.jsonl" "b/jsonl/Edit-Reward-Bench/labeled_data/0829_labels_104-208.jsonl" new file mode 100644--- /dev/null +++ "b/jsonl/Edit-Reward-Bench/labeled_data/0829_labels_104-208.jsonl" @@ -0,0 +1,103 @@ +{"task_type": "subject-add", "data_source": "GEdit-Bench", "instruction": "Add a puppy to this picture leaning against the girl's legs.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-add/en/720454d83c65f03eefe4cb6da5d706df_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/subject-add/en/720454d83c65f03eefe4cb6da5d706df.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/subject-add/en/720454d83c65f03eefe4cb6da5d706df.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/subject-add/en/720454d83c65f03eefe4cb6da5d706df.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-add/en/720454d83c65f03eefe4cb6da5d706df.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/subject-add/en/720454d83c65f03eefe4cb6da5d706df.png"], "label": "bad", "timestamp": "2025-08-29T13:46:09.943996"} +{"task_type": "subject-add", "data_source": "GEdit-Bench", "instruction": "Add a puppy to this picture leaning against the girl's legs.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-add/en/720454d83c65f03eefe4cb6da5d706df_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-add/en/720454d83c65f03eefe4cb6da5d706df.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/subject-add/en/720454d83c65f03eefe4cb6da5d706df.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/subject-add/en/720454d83c65f03eefe4cb6da5d706df.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/subject-add/en/720454d83c65f03eefe4cb6da5d706df.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/subject-add/en/720454d83c65f03eefe4cb6da5d706df.png"], "label": "bad", "timestamp": "2025-08-29T13:48:04.503173"} +{"task_type": "subject-add", "data_source": "GEdit-Bench", "instruction": "Light the candle to enhance the candlelight.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-add/en/853784745a3c52dcfd24cf3a8dba1f56_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/subject-add/en/853784745a3c52dcfd24cf3a8dba1f56.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/subject-add/en/853784745a3c52dcfd24cf3a8dba1f56.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/subject-add/en/853784745a3c52dcfd24cf3a8dba1f56.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/subject-add/en/853784745a3c52dcfd24cf3a8dba1f56.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/subject-add/en/853784745a3c52dcfd24cf3a8dba1f56.png"], "label": "triple_ranking", "timestamp": "2025-08-29T13:51:10.221012", "instruction_following": {"ranking_text": "4|35|1|2"}, "consistency": {"ranking_text": "2345|1"}, "overall": {"ranking_text": "4|35|1|2"}} +{"task_type": "subject-add", "data_source": "GEdit-Bench", "instruction": "add a person standing next to the bus", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-add/en/e9ac3ec18e91f8bf73b340de1c2e459e_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/subject-add/en/e9ac3ec18e91f8bf73b340de1c2e459e.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-add/en/e9ac3ec18e91f8bf73b340de1c2e459e.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/subject-add/en/e9ac3ec18e91f8bf73b340de1c2e459e.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/subject-add/en/e9ac3ec18e91f8bf73b340de1c2e459e.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/subject-add/en/e9ac3ec18e91f8bf73b340de1c2e459e.png"], "label": "triple_ranking", "timestamp": "2025-08-29T13:53:20.404295", "instruction_following": {"ranking_text": "124|3|5"}, "consistency": {"ranking_text": "34|125"}, "overall": {"ranking_text": "12|34|5"}} +{"task_type": "subject-add", "data_source": "imgedit", "instruction": "Add a small stone gazebo with a tiled roof near the center-left area of the garden, blending naturally with the surrounding greenery and landscape.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/for_add/000145048.jpg"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/83.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/83.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/83.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/83.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/83.png"], "label": "triple_ranking", "timestamp": "2025-08-29T13:56:38.973047", "instruction_following": {"ranking_text": "45|12|3"}, "consistency": {"ranking_text": "4|1235"}, "overall": {"ranking_text": "4|5|2|1|3"}} +{"task_type": "subject-add", "data_source": "imgedit", "instruction": "Add Add a wooden a wooden bench along the path near the bench along the dirt path edge of the cliff, facing the on the right side of the image. ocean.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/for_add/000155856.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/92.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/92.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/92.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/92.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/92.png"], "label": "triple_ranking", "timestamp": "2025-08-29T13:59:30.031832", "instruction_following": {"ranking_text": "2|5|3|4|1"}, "consistency": {"ranking_text": "2|4|1|3|5"}, "overall": {"ranking_text": "2|3|5|4|1"}} +{"task_type": "subject-add", "data_source": "imgedit", "instruction": "Add a seagull flying above the lifeguard tower, near the clouds.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/for_add/000081452.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/56.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/56.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/56.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/56.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/56.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:02:03.240254", "instruction_following": {"ranking_text": "1|25|4|3"}, "consistency": {"ranking_text": "5|234|1"}, "overall": {"ranking_text": "25|1|4|3"}} +{"task_type": "subject-add", "data_source": "imgedit", "instruction": "Add a bicycle near the fence on the left side of the image.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/for_add/000313906.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/108.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/108.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/108.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/108.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/108.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:04:42.239704", "instruction_following": {"ranking_text": "1|3|2|5|4"}, "consistency": {"ranking_text": "123|4|5"}, "overall": {"ranking_text": "1|3|2|4|5"}} +{"task_type": "color_alter", "data_source": "GEdit-Bench", "instruction": "Alter the color of bus to lime", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/color_alter/en/a4ca581574347248e1762c4987c931aa_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/color_alter/en/a4ca581574347248e1762c4987c931aa.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/color_alter/en/a4ca581574347248e1762c4987c931aa.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/color_alter/en/a4ca581574347248e1762c4987c931aa.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/color_alter/en/a4ca581574347248e1762c4987c931aa.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/color_alter/en/a4ca581574347248e1762c4987c931aa.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:07:07.259975", "instruction_following": {"ranking_text": "3|245|1"}, "consistency": {"ranking_text": "34|5|2|1"}, "overall": {"ranking_text": "3|4|5|2|1"}} +{"task_type": "color_alter", "data_source": "GEdit-Bench", "instruction": "Change the car body color to gray.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/color_alter/en/eeab5f9b2f3a62deb674c7bc6af021fb_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/color_alter/en/eeab5f9b2f3a62deb674c7bc6af021fb.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/color_alter/en/eeab5f9b2f3a62deb674c7bc6af021fb.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/color_alter/en/eeab5f9b2f3a62deb674c7bc6af021fb.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/color_alter/en/eeab5f9b2f3a62deb674c7bc6af021fb.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/color_alter/en/eeab5f9b2f3a62deb674c7bc6af021fb.png"], "label": "bad", "timestamp": "2025-08-29T14:09:07.187785"} +{"task_type": "color_alter", "data_source": "GEdit-Bench", "instruction": "change the color of couch to yellow", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/color_alter/en/f437c7392b76ded921a0abc243f81290_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/color_alter/en/f437c7392b76ded921a0abc243f81290.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/color_alter/en/f437c7392b76ded921a0abc243f81290.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/color_alter/en/f437c7392b76ded921a0abc243f81290.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/color_alter/en/f437c7392b76ded921a0abc243f81290.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/color_alter/en/f437c7392b76ded921a0abc243f81290.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:11:28.228312", "instruction_following": {"ranking_text": "25|4|3|1"}, "consistency": {"ranking_text": "5|2|4|3|1"}, "overall": {"ranking_text": "5|2|4|3|1"}} +{"task_type": "color_alter", "data_source": "GEdit-Bench", "instruction": "change the color of elephant to pink", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/color_alter/en/174b49f45ca4ff5d1d3ea06096b78e57_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/color_alter/en/174b49f45ca4ff5d1d3ea06096b78e57.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/color_alter/en/174b49f45ca4ff5d1d3ea06096b78e57.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/color_alter/en/174b49f45ca4ff5d1d3ea06096b78e57.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/color_alter/en/174b49f45ca4ff5d1d3ea06096b78e57.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/color_alter/en/174b49f45ca4ff5d1d3ea06096b78e57.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:14:06.003896", "instruction_following": {"ranking_text": "234|1|5"}, "consistency": {"ranking_text": "1|23|4|5"}, "overall": {"ranking_text": "2|3|4|1|5"}} +{"task_type": "color_alter", "data_source": "imgedit", "instruction": "Change the building's wall color to light blue.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/architecture/000004499.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/1104.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/1104.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1104.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/1104.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/1104.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:17:09.477051", "instruction_following": {"ranking_text": "1|45|2|3"}, "consistency": {"ranking_text": "14|5|2|3"}, "overall": {"ranking_text": "1|4|5|2|3"}} +{"task_type": "color_alter", "data_source": "imgedit", "instruction": "Change the color of the vehicle to red.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/transport/000276156.jpg"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1096.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/1096.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/1096.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/1096.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/1096.png"], "label": "bad", "timestamp": "2025-08-29T14:17:48.579756"} +{"task_type": "color_alter", "data_source": "imgedit", "instruction": "Change the suit color to navy blue.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/human/000015317.jpg"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/1046.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1046.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/1046.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1046.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/1046.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:22:43.360856", "instruction_following": {"ranking_text": "1|23|4|5"}, "consistency": {"ranking_text": "145|2|3"}, "overall": {"ranking_text": "1|4|2|3|5"}} +{"task_type": "color_alter", "data_source": "imgedit", "instruction": "Change the surface texture of the object to a smooth, glossy finish.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/daily object/000117413.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/1131.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1131.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/1131.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/1131.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1131.png"], "label": "bad", "timestamp": "2025-08-29T14:24:27.309494"} +{"task_type": "compose", "data_source": "imgedit", "instruction": "Remove the object on the left side of the image, and increase the brightness of the central figure.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/compose/objects/3.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1169.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/1169.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/1169.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1169.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/1169.png"], "label": "bad", "timestamp": "2025-08-29T14:25:29.759884"} +{"task_type": "compose", "data_source": "imgedit", "instruction": "Remove the object on the table, and increase the brightness of the background.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/compose/human/9.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/imgedit/1158.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/1158.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/1158.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/1158.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/1158.png"], "label": "bad", "timestamp": "2025-08-29T14:27:10.894851"} +{"task_type": "compose", "data_source": "imgedit", "instruction": "Remove the smartphone from the man's hand, and adjust the woman's hair to appear more windblown.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/compose/human/5.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1154.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/1154.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/1154.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/1154.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/1154.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:30:03.247121", "instruction_following": {"ranking_text": "3|45|2|1"}, "consistency": {"ranking_text": "34|5|2|1"}, "overall": {"ranking_text": "3|4|5|2|1"}} +{"task_type": "compose", "data_source": "imgedit", "instruction": "Remove the laptop from the person's lap, and change the color of the couch to light blue.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/compose/human/10.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/1145.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1145.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/1145.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/1145.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/1145.png"], "label": "bad", "timestamp": "2025-08-29T14:35:58.404446"} +{"task_type": "compose", "data_source": "imgedit", "instruction": "Remove the radiator on the right side of the image, and change the color of the black and white cat to brown.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/compose/animal/7.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/imgedit/1161.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/1161.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1161.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/1161.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/1161.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:38:41.711094", "instruction_following": {"ranking_text": "2|13|4|5"}, "consistency": {"ranking_text": "134|5|2"}, "overall": {"ranking_text": "2|13|45"}} +{"task_type": "compose", "data_source": "imgedit", "instruction": "Remove the phone from the man on the left, and change the color of the woman's sweater to light blue.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/compose/human/13.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/1148.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/1148.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/1148.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/1148.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1148.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:41:44.017377", "instruction_following": {"ranking_text": "45|2|1|3"}, "consistency": {"ranking_text": "245|13"}, "overall": {"ranking_text": "5|24|3|1"}} +{"task_type": "compose", "data_source": "imgedit", "instruction": "Remove the person on the left side of the image, and adjust the brightness of the background to make it appear lighter.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/compose/human/2.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/1151.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/1151.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/1151.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1151.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1151.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:43:36.900151", "instruction_following": {"ranking_text": "4|2|3|15"}, "consistency": {"ranking_text": "34|2|15"}, "overall": {"ranking_text": "4|2|3|1|5"}} +{"task_type": "motion_change", "data_source": "GEdit-Bench", "instruction": "Make the person in the image smile.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/motion_change/en/353cf97ec89d2e51932763ae7538c4cc_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/motion_change/en/353cf97ec89d2e51932763ae7538c4cc.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/motion_change/en/353cf97ec89d2e51932763ae7538c4cc.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/motion_change/en/353cf97ec89d2e51932763ae7538c4cc.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/motion_change/en/353cf97ec89d2e51932763ae7538c4cc.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/motion_change/en/353cf97ec89d2e51932763ae7538c4cc.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:46:53.458057", "instruction_following": {"ranking_text": "235|14"}, "consistency": {"ranking_text": "135|2|4"}, "overall": {"ranking_text": "3|25|1|4"}} +{"task_type": "motion_change", "data_source": "GEdit-Bench", "instruction": "make the action of the man to cheering", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/motion_change/en/be4bb34c6d879f253a4b7c4f32fc333f_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/motion_change/en/be4bb34c6d879f253a4b7c4f32fc333f.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/motion_change/en/be4bb34c6d879f253a4b7c4f32fc333f.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/motion_change/en/be4bb34c6d879f253a4b7c4f32fc333f.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/motion_change/en/be4bb34c6d879f253a4b7c4f32fc333f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/motion_change/en/be4bb34c6d879f253a4b7c4f32fc333f.png"], "label": "bad", "timestamp": "2025-08-29T14:48:41.304249"} +{"task_type": "motion_change", "data_source": "GEdit-Bench", "instruction": "Animate the cat in the image.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/motion_change/en/e648d94e8f66940befa13e34039be176_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/motion_change/en/e648d94e8f66940befa13e34039be176.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/motion_change/en/e648d94e8f66940befa13e34039be176.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/motion_change/en/e648d94e8f66940befa13e34039be176.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/motion_change/en/e648d94e8f66940befa13e34039be176.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/motion_change/en/e648d94e8f66940befa13e34039be176.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:49:58.180830", "instruction_following": {"ranking_text": "3|5|2|14"}, "consistency": {"ranking_text": "1|4|23|5"}, "overall": {"ranking_text": "3|5|2|14"}} +{"task_type": "motion_change", "data_source": "GEdit-Bench", "instruction": "Make the girl in the photo blow a kiss.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/motion_change/en/9ce39582df9aaf3b21b39fb9627f7bb2_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/motion_change/en/9ce39582df9aaf3b21b39fb9627f7bb2.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/motion_change/en/9ce39582df9aaf3b21b39fb9627f7bb2.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/motion_change/en/9ce39582df9aaf3b21b39fb9627f7bb2.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/motion_change/en/9ce39582df9aaf3b21b39fb9627f7bb2.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/motion_change/en/9ce39582df9aaf3b21b39fb9627f7bb2.png"], "label": "triple_ranking", "timestamp": "2025-08-29T14:51:41.718336", "instruction_following": {"ranking_text": "23|45|1"}, "consistency": {"ranking_text": "1|4|2|35"}, "overall": {"ranking_text": "2|34|5|1"}} +{"task_type": "motion_change", "data_source": "imgedit", "instruction": "Make the person raise her right arm.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/clothes/00000006.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1023.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/1023.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/1023.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/1023.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/1023.png"], "label": "bad", "timestamp": "2025-08-29T14:55:56.797260"} +{"task_type": "motion_change", "data_source": "imgedit", "instruction": "Raise the person's left arm.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/clothes/00000048.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/imgedit/1037.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1037.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1037.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/1037.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/1037.png"], "label": "bad", "timestamp": "2025-08-29T14:56:37.597556"} +{"task_type": "motion_change", "data_source": "imgedit", "instruction": "Raise the person's left arm.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/human/000336112.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1019.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/1019.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/1019.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/1019.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/1019.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:09:35.794810", "instruction_following": {"ranking_text": "4|1235"}, "consistency": {"ranking_text": "5|234|1"}, "overall": {"ranking_text": "4|5|23|1"}} +{"task_type": "motion_change", "data_source": "imgedit", "instruction": "Raise the person's left arm.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/clothes/00000009.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/1024.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/1024.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/1024.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/1024.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/1024.png"], "label": "bad", "timestamp": "2025-08-29T15:10:25.966126"} +{"task_type": "text_change", "data_source": "GEdit-Bench", "instruction": "Replace the text 'TRAIN' with 'PLANE'", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/5e085566f105978483848cab2f3a7001_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/text_change/en/5e085566f105978483848cab2f3a7001.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/text_change/en/5e085566f105978483848cab2f3a7001.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/text_change/en/5e085566f105978483848cab2f3a7001.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/text_change/en/5e085566f105978483848cab2f3a7001.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gpt_image_1/results/fullset/text_change/en/5e085566f105978483848cab2f3a7001.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:13:25.929044", "instruction_following": {"ranking_text": "35|12|4"}, "consistency": {"ranking_text": "13|245"}, "overall": {"ranking_text": "3|5|12|4"}} +{"task_type": "text_change", "data_source": "GEdit-Bench", "instruction": "Replace the text 'McCONAUGHEY' with 'McDonald'", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/140cbdcd2cc6adccb374b62d40f41b9f_SRCIMG.png"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image/fullset/text_change/en/140cbdcd2cc6adccb374b62d40f41b9f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/text_change/en/140cbdcd2cc6adccb374b62d40f41b9f.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/text_change/en/140cbdcd2cc6adccb374b62d40f41b9f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/text_change/en/140cbdcd2cc6adccb374b62d40f41b9f.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/text_change/en/140cbdcd2cc6adccb374b62d40f41b9f.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:15:16.533796", "instruction_following": {"ranking_text": "45|3|12"}, "consistency": {"ranking_text": "1345|2"}, "overall": {"ranking_text": "4|5|3|1|2"}} +{"task_type": "text_change", "data_source": "GEdit-Bench", "instruction": "Change the text 'FAIRGROUNDS' to 'PARKWAY'", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/f00dfa62ea1474aeb985b5447dc8fa0c_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/text_change/en/f00dfa62ea1474aeb985b5447dc8fa0c.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/text_change/en/f00dfa62ea1474aeb985b5447dc8fa0c.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/f00dfa62ea1474aeb985b5447dc8fa0c.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/text_change/en/f00dfa62ea1474aeb985b5447dc8fa0c.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/text_change/en/f00dfa62ea1474aeb985b5447dc8fa0c.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:16:44.337371", "instruction_following": {"ranking_text": "235|4|1"}, "consistency": {"ranking_text": "2345|1"}, "overall": {"ranking_text": "35|2|4|1"}} +{"task_type": "text_change", "data_source": "GEdit-Bench", "instruction": "Change the text 'Bank' to 'Banks'", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/a9ae1402abe1d2624b7fce054edd7313_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/text_change/en/a9ae1402abe1d2624b7fce054edd7313.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/text_change/en/a9ae1402abe1d2624b7fce054edd7313.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/text_change/en/a9ae1402abe1d2624b7fce054edd7313.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/a9ae1402abe1d2624b7fce054edd7313.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gpt_image_1/results/fullset/text_change/en/a9ae1402abe1d2624b7fce054edd7313.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:18:56.898905", "instruction_following": {"ranking_text": "2|135|4"}, "consistency": {"ranking_text": "24|13|5"}, "overall": {"ranking_text": "2|13|5|4"}} +{"task_type": "text_change", "data_source": "GEdit-Bench", "instruction": "Help me insert the words \"时光与你共守情长\" into the image.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/be45c39b3bcc9d082051c13b5300dde1_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/text_change/en/be45c39b3bcc9d082051c13b5300dde1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/text_change/en/be45c39b3bcc9d082051c13b5300dde1.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/text_change/en/be45c39b3bcc9d082051c13b5300dde1.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/text_change/en/be45c39b3bcc9d082051c13b5300dde1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/text_change/en/be45c39b3bcc9d082051c13b5300dde1.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:20:44.772047", "instruction_following": {"ranking_text": "4|1|3|5|2"}, "consistency": {"ranking_text": "2345|1"}, "overall": {"ranking_text": "4|1|3|5|2"}} +{"task_type": "text_change", "data_source": "GEdit-Bench", "instruction": "Replace the text '95' with '123'", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/99500eadb2f363c2e26fcb501972c29f_SRCIMG.png"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image/fullset/text_change/en/99500eadb2f363c2e26fcb501972c29f.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/text_change/en/99500eadb2f363c2e26fcb501972c29f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/text_change/en/99500eadb2f363c2e26fcb501972c29f.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/text_change/en/99500eadb2f363c2e26fcb501972c29f.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/text_change/en/99500eadb2f363c2e26fcb501972c29f.png"], "label": "bad", "timestamp": "2025-08-29T15:21:55.017414"} +{"task_type": "text_change", "data_source": "GEdit-Bench", "instruction": "Replace the text 'Salmon' with 'Sandwich'", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/89714e3ea9345ea5483ac6d5856915fe_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/text_change/en/89714e3ea9345ea5483ac6d5856915fe.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/text_change/en/89714e3ea9345ea5483ac6d5856915fe.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/text_change/en/89714e3ea9345ea5483ac6d5856915fe.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/text_change/en/89714e3ea9345ea5483ac6d5856915fe.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/text_change/en/89714e3ea9345ea5483ac6d5856915fe.png"], "label": "bad", "timestamp": "2025-08-29T15:22:22.531370"} +{"task_type": "text_change", "data_source": "GEdit-Bench", "instruction": "I need you to change \"2024\" to \"2025\" in this image and replace the \"陈\" character inside the topmost heart with \"华\".", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/5b8717b2209b784940f388864d5520f3_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/text_change/en/5b8717b2209b784940f388864d5520f3.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/text_change/en/5b8717b2209b784940f388864d5520f3.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/text_change/en/5b8717b2209b784940f388864d5520f3.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/text_change/en/5b8717b2209b784940f388864d5520f3.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/text_change/en/5b8717b2209b784940f388864d5520f3.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:23:39.258243", "instruction_following": {"ranking_text": "23|145"}, "overall": {"ranking_text": "3|2|145"}} +{"task_type": "subject-replace", "data_source": "GEdit-Bench", "instruction": "Replace the eagle with a parrot.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-replace/en/469267bfc120943d28e93b6ecefe14af_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-replace/en/469267bfc120943d28e93b6ecefe14af.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/subject-replace/en/469267bfc120943d28e93b6ecefe14af.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/subject-replace/en/469267bfc120943d28e93b6ecefe14af.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/subject-replace/en/469267bfc120943d28e93b6ecefe14af.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/subject-replace/en/469267bfc120943d28e93b6ecefe14af.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:26:06.352498", "instruction_following": {"ranking_text": "13|5|2|4"}, "consistency": {"ranking_text": "135|24"}} +{"task_type": "subject-replace", "data_source": "GEdit-Bench", "instruction": "Turn the tree branches in the image into a witch’s magic wand.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-replace/en/70fa37d482c2e708435366323262de90_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/subject-replace/en/70fa37d482c2e708435366323262de90.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/subject-replace/en/70fa37d482c2e708435366323262de90.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/subject-replace/en/70fa37d482c2e708435366323262de90.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-replace/en/70fa37d482c2e708435366323262de90.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/subject-replace/en/70fa37d482c2e708435366323262de90.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:27:30.523499", "instruction_following": {"ranking_text": "234|15"}, "consistency": {"ranking_text": "1245|3"}, "overall": {"ranking_text": "24|3|15"}} +{"task_type": "subject-replace", "data_source": "GEdit-Bench", "instruction": "Replace the bracelet with a jade bangle.", "input_images": ["/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/subject-replace/en/038b15558b1082c59f0a92e4853554aa_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/subject-replace/en/038b15558b1082c59f0a92e4853554aa.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/subject-replace/en/038b15558b1082c59f0a92e4853554aa.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/subject-replace/en/038b15558b1082c59f0a92e4853554aa.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/subject-replace/en/038b15558b1082c59f0a92e4853554aa.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/subject-replace/en/038b15558b1082c59f0a92e4853554aa.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:30:25.168757", "instruction_following": {"ranking_text": "12|345"}, "consistency": {"ranking_text": "12|345"}, "overall": {"ranking_text": "12|45|3"}} +{"task_type": "subject-replace", "data_source": "GEdit-Bench", "instruction": "Change the clothes of the girl on the left side of the image to a black-and-white striped dress.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-replace/en/619ba60ef621caf9f1412bfa7a3eb5c1_SRCIMG.png"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image/fullset/subject-replace/en/619ba60ef621caf9f1412bfa7a3eb5c1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/subject-replace/en/619ba60ef621caf9f1412bfa7a3eb5c1.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/subject-replace/en/619ba60ef621caf9f1412bfa7a3eb5c1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/subject-replace/en/619ba60ef621caf9f1412bfa7a3eb5c1.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/subject-replace/en/619ba60ef621caf9f1412bfa7a3eb5c1.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:33:14.724067", "instruction_following": {"ranking_text": "15|3|2|4"}, "consistency": {"ranking_text": "15|3|2|4"}, "overall": {"ranking_text": "15|3|2|4"}} +{"task_type": "subject-replace", "data_source": "imgedit", "instruction": "Replace the brown suitcase in the image with a large potted plant.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/daily object/000117413.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/624.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/624.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/624.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/624.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/624.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:35:08.393307", "instruction_following": {"ranking_text": "145|23"}, "overall": {"ranking_text": "45|1|23"}} +{"task_type": "subject-replace", "data_source": "imgedit", "instruction": "Replace the girl in the image with a large crystal chandelier hanging from an invisible support, blending naturally into the forest environment.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/human/000264849.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/550.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/550.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/550.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/550.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/550.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:38:35.225370", "instruction_following": {"ranking_text": "25|4|3|1"}, "consistency": {"ranking_text": "25|4|13"}, "overall": {"ranking_text": "2|45|3|1"}} +{"task_type": "subject-replace", "data_source": "imgedit", "instruction": "Replace the bartender in the image with a person holding a large bouquet of flowers.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/human/000281608.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/552.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/552.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/552.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/552.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/552.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:40:51.649814", "overall": {"ranking_text": "3|2|1|45"}} +{"task_type": "subject-replace", "data_source": "imgedit", "instruction": "Replace the bird in the image with a squirrel.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/animal/000250140.jpg"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/573.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/573.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/573.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/573.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/573.png"], "label": "bad", "timestamp": "2025-08-29T15:41:53.965140"} +{"task_type": "ps_human", "data_source": "GEdit-Bench", "instruction": "Make him laugh heartily", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/6022c9e5401a53028e3b0690cce7a9db_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/ps_human/en/6022c9e5401a53028e3b0690cce7a9db.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/ps_human/en/6022c9e5401a53028e3b0690cce7a9db.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/ps_human/en/6022c9e5401a53028e3b0690cce7a9db.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/ps_human/en/6022c9e5401a53028e3b0690cce7a9db.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/6022c9e5401a53028e3b0690cce7a9db.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:44:02.961860", "consistency": {"ranking_text": "3|14|25"}, "overall": {"ranking_text": "3|14|5|2"}} +{"task_type": "ps_human", "data_source": "GEdit-Bench", "instruction": "Make him grow long hair", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/f21e2f3585f8cddeab9d472375e92bac_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/ps_human/en/f21e2f3585f8cddeab9d472375e92bac.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/ps_human/en/f21e2f3585f8cddeab9d472375e92bac.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/ps_human/en/f21e2f3585f8cddeab9d472375e92bac.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/f21e2f3585f8cddeab9d472375e92bac.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/ps_human/en/f21e2f3585f8cddeab9d472375e92bac.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:45:16.334424", "instruction_following": {"ranking_text": "245|1|3"}, "consistency": {"ranking_text": "245|1|3"}, "overall": {"ranking_text": "245|1|3"}} +{"task_type": "ps_human", "data_source": "GEdit-Bench", "instruction": "Make me look handsome.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/4b7e3f9099377e3823c1c3e0d924883a_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/ps_human/en/4b7e3f9099377e3823c1c3e0d924883a.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/ps_human/en/4b7e3f9099377e3823c1c3e0d924883a.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/4b7e3f9099377e3823c1c3e0d924883a.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/ps_human/en/4b7e3f9099377e3823c1c3e0d924883a.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/ps_human/en/4b7e3f9099377e3823c1c3e0d924883a.png"], "label": "bad", "timestamp": "2025-08-29T15:46:22.544129"} +{"task_type": "ps_human", "data_source": "GEdit-Bench", "instruction": "Make him look less angry", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/41d961b14b637889947080f1891f85ff_SRCIMG.png"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image/fullset/ps_human/en/41d961b14b637889947080f1891f85ff.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/ps_human/en/41d961b14b637889947080f1891f85ff.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/ps_human/en/41d961b14b637889947080f1891f85ff.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/ps_human/en/41d961b14b637889947080f1891f85ff.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/ps_human/en/41d961b14b637889947080f1891f85ff.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:49:14.549191", "instruction_following": {"ranking_text": "23|1|45"}, "consistency": {"ranking_text": "25|134"}, "overall": {"ranking_text": "2|3|1|5|4"}} +{"task_type": "ps_human", "data_source": "GEdit-Bench", "instruction": "Make him have more muscles and a stronger vibe", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/5968a24147a8564f74bd09104c4c032e_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/ps_human/en/5968a24147a8564f74bd09104c4c032e.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/ps_human/en/5968a24147a8564f74bd09104c4c032e.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/ps_human/en/5968a24147a8564f74bd09104c4c032e.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/ps_human/en/5968a24147a8564f74bd09104c4c032e.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/ps_human/en/5968a24147a8564f74bd09104c4c032e.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:51:56.768979", "instruction_following": {"ranking_text": "2|5|4|3|1"}, "consistency": {"ranking_text": "2|45|3|1"}, "overall": {"ranking_text": "2|5|4|3|1"}} +{"task_type": "ps_human", "data_source": "GEdit-Bench", "instruction": "Feminize my appearance.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/3b0f6ca611bfa2f2416bf7ade7f60811_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/3b0f6ca611bfa2f2416bf7ade7f60811.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/ps_human/en/3b0f6ca611bfa2f2416bf7ade7f60811.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/ps_human/en/3b0f6ca611bfa2f2416bf7ade7f60811.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/ps_human/en/3b0f6ca611bfa2f2416bf7ade7f60811.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/ps_human/en/3b0f6ca611bfa2f2416bf7ade7f60811.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:53:52.709546", "instruction_following": {"ranking_text": "145|23"}, "overall": {"ranking_text": "1|5|4|2|3"}} +{"task_type": "ps_human", "data_source": "GEdit-Bench", "instruction": "Edit this photo—slim the waist and lift the butt.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/2dd8b5fb8e22905ed49d87660eb82ee0_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/2dd8b5fb8e22905ed49d87660eb82ee0.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/ps_human/en/2dd8b5fb8e22905ed49d87660eb82ee0.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/ps_human/en/2dd8b5fb8e22905ed49d87660eb82ee0.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/ps_human/en/2dd8b5fb8e22905ed49d87660eb82ee0.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/ps_human/en/2dd8b5fb8e22905ed49d87660eb82ee0.png"], "label": "bad", "timestamp": "2025-08-29T15:54:53.658956"} +{"task_type": "ps_human", "data_source": "GEdit-Bench", "instruction": "Make him look like he is crying a lot", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/fe220565cb0f22a44d1f0a81a132ce9f_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/ps_human/en/fe220565cb0f22a44d1f0a81a132ce9f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/ps_human/en/fe220565cb0f22a44d1f0a81a132ce9f.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/ps_human/en/fe220565cb0f22a44d1f0a81a132ce9f.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/ps_human/en/fe220565cb0f22a44d1f0a81a132ce9f.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/ps_human/en/fe220565cb0f22a44d1f0a81a132ce9f.png"], "label": "triple_ranking", "timestamp": "2025-08-29T15:56:05.161497", "overall": {"ranking_text": "2|1345"}} +{"task_type": "subject-remove", "data_source": "GEdit-Bench", "instruction": "Remove the snake pattern from the image.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-remove/en/1110b6bc43aa5a3037467f5833ece3c5_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/subject-remove/en/1110b6bc43aa5a3037467f5833ece3c5.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/subject-remove/en/1110b6bc43aa5a3037467f5833ece3c5.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-remove/en/1110b6bc43aa5a3037467f5833ece3c5.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/subject-remove/en/1110b6bc43aa5a3037467f5833ece3c5.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/subject-remove/en/1110b6bc43aa5a3037467f5833ece3c5.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:00:05.060023", "instruction_following": {"ranking_text": "125|34"}, "consistency": {"ranking_text": "125|3|4"}, "overall": {"ranking_text": "125|3|4"}} +{"task_type": "subject-remove", "data_source": "GEdit-Bench", "instruction": "Remove the person in the distance wearing red clothes and a green backpack.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-remove/en/2ebb31e51f0de8bba7005352df7150b8_SRCIMG.png"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image/fullset/subject-remove/en/2ebb31e51f0de8bba7005352df7150b8.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/subject-remove/en/2ebb31e51f0de8bba7005352df7150b8.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-remove/en/2ebb31e51f0de8bba7005352df7150b8.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/subject-remove/en/2ebb31e51f0de8bba7005352df7150b8.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/subject-remove/en/2ebb31e51f0de8bba7005352df7150b8.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:04:13.839450", "instruction_following": {"ranking_text": "123|4|5"}, "consistency": {"ranking_text": "12|3|4|5"}, "overall": {"ranking_text": "123|4|5"}} +{"task_type": "subject-remove", "data_source": "GEdit-Bench", "instruction": "Delete the standing person wearing glasses.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-remove/en/16e59a89f37b5603a41fc60e2912a325_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/subject-remove/en/16e59a89f37b5603a41fc60e2912a325.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/subject-remove/en/16e59a89f37b5603a41fc60e2912a325.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/subject-remove/en/16e59a89f37b5603a41fc60e2912a325.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/subject-remove/en/16e59a89f37b5603a41fc60e2912a325.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/subject-remove/en/16e59a89f37b5603a41fc60e2912a325.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:06:11.231766", "instruction_following": {"ranking_text": "25|14|3"}, "consistency": {"ranking_text": "235|14"}, "overall": {"ranking_text": "2|5|1|4|3"}} +{"task_type": "subject-remove", "data_source": "GEdit-Bench", "instruction": "Remove the bag near the character’s hand.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-remove/en/bc8d567ee91ca1521adaa8d4a486851f_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/subject-remove/en/bc8d567ee91ca1521adaa8d4a486851f.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/subject-remove/en/bc8d567ee91ca1521adaa8d4a486851f.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/subject-remove/en/bc8d567ee91ca1521adaa8d4a486851f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/subject-remove/en/bc8d567ee91ca1521adaa8d4a486851f.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/subject-remove/en/bc8d567ee91ca1521adaa8d4a486851f.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:07:34.699255", "instruction_following": {"ranking_text": "2345|1"}, "consistency": {"ranking_text": "345|2|1"}, "overall": {"ranking_text": "345|2|1"}} +{"task_type": "subject-remove", "data_source": "imgedit", "instruction": "Remove the white rabbit sitting on the cabbage.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/animal/000000265.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/476.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/476.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/476.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/476.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/476.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:08:58.470589", "instruction_following": {"ranking_text": "345|1|2"}, "overall": {"ranking_text": "345|12"}} +{"task_type": "subject-remove", "data_source": "imgedit", "instruction": "Remove the helicopter in the foreground.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/transport/000270866.jpg"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/505.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/505.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/505.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/505.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/505.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:12:24.649173", "instruction_following": {"ranking_text": "2345|1"}, "consistency": {"ranking_text": "1245|3"}, "overall": {"ranking_text": "4|2|5|3|1"}} +{"task_type": "subject-remove", "data_source": "imgedit", "instruction": "Remove the person in the foreground of the image.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/human/000270466.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/465.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/465.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/465.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/465.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/465.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:14:42.796781", "instruction_following": {"ranking_text": "123|5|4"}, "consistency": {"ranking_text": "23|15|4"}, "overall": {"ranking_text": "23|1|5|4"}} +{"task_type": "subject-remove", "data_source": "imgedit", "instruction": "Remove the yacht in the image.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/transport/000156391.jpg"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/500.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/500.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/500.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/500.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/500.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:19:00.577699", "instruction_following": {"ranking_text": "4|35|1|2"}, "consistency": {"ranking_text": "2345|1"}, "overall": {"ranking_text": "4|35|1|2"}} +{"task_type": "extract", "data_source": "imgedit", "instruction": "Extract the black T-shirt and the black sheer skirt worn by the person in the image", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/clothes/00000024.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/441.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/441.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/441.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/441.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/441.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:20:06.177259", "instruction_following": {"ranking_text": "5|1234"}, "consistency": {"ranking_text": "25|3|14"}, "overall": {"ranking_text": "5|1234"}} +{"task_type": "extract", "data_source": "imgedit", "instruction": "Extract the grey Adidas T-shirt with white trim and stripes worn by the person in the image.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/clothes/00000033.jpg"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image_imgedit/444.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/444.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/444.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/444.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/444.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:21:17.532089", "instruction_following": {"ranking_text": "23|145"}, "consistency": {"ranking_text": "2|34|5|1"}, "overall": {"ranking_text": "2|3|145"}} +{"task_type": "extract", "data_source": "imgedit", "instruction": "Extract the coffee cup placed on the table in the image", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/daily object/000363394.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/430.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/430.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/430.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/430.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/430.png"], "label": "bad", "timestamp": "2025-08-29T16:22:08.599179"} +{"task_type": "extract", "data_source": "imgedit", "instruction": "Extract the dress in the image.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/daily object/000340873.jpg"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image_imgedit/429.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/429.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/429.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/429.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/429.png"], "label": "bad", "timestamp": "2025-08-29T16:23:07.229340"} +{"task_type": "extract", "data_source": "imgedit", "instruction": "Extract the wooden serving tray holding the dessert bars in the image.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/daily object/000273675.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/424.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/424.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/424.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/424.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/424.png"], "label": "bad", "timestamp": "2025-08-29T16:24:43.139936"} +{"task_type": "extract", "data_source": "imgedit", "instruction": "Extract the daily object(s) visible in the image.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/daily object/000158674.jpg"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/422.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/422.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/422.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/422.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/422.png"], "label": "bad", "timestamp": "2025-08-29T16:25:28.690350"} +{"task_type": "extract", "data_source": "imgedit", "instruction": "Extract the architecture from the image.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/architecture/000023390.jpg"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/394.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/394.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/394.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/394.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/394.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:27:07.814708", "instruction_following": {"ranking_text": "5|2|134"}, "consistency": {"ranking_text": "5|123|4"}, "overall": {"ranking_text": "5|2|134"}} +{"task_type": "extract", "data_source": "imgedit", "instruction": "Extract the human standing in the foreground wearing a grey zip-up hoodie with arms crossed.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/human/000270466.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/342.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/342.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/342.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/342.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/342.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:29:06.526541", "instruction_following": {"ranking_text": "35|124"}, "consistency": {"ranking_text": "3|24|5|1"}, "overall": {"ranking_text": "3|5|24|1"}} +{"task_type": "background_change", "data_source": "GEdit-Bench", "instruction": "Change the background to a nighttime cityscape.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/background_change/en/218747d7f3c9ce2eaef0ea3083362626_SRCIMG.png"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image/fullset/background_change/en/218747d7f3c9ce2eaef0ea3083362626.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/background_change/en/218747d7f3c9ce2eaef0ea3083362626.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/background_change/en/218747d7f3c9ce2eaef0ea3083362626.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/background_change/en/218747d7f3c9ce2eaef0ea3083362626.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/background_change/en/218747d7f3c9ce2eaef0ea3083362626.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:31:09.193323", "instruction_following": {"ranking_text": "1234|5"}, "consistency": {"ranking_text": "1234|5"}, "overall": {"ranking_text": "1234|5"}} +{"task_type": "background_change", "data_source": "GEdit-Bench", "instruction": "Remove the background for me.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/background_change/en/806f4a1d864636f48a994032447bb5a8_SRCIMG.png"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image/fullset/background_change/en/806f4a1d864636f48a994032447bb5a8.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/background_change/en/806f4a1d864636f48a994032447bb5a8.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/background_change/en/806f4a1d864636f48a994032447bb5a8.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/background_change/en/806f4a1d864636f48a994032447bb5a8.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/background_change/en/806f4a1d864636f48a994032447bb5a8.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:32:13.032352", "instruction_following": {"ranking_text": "14|3|25"}, "consistency": {"ranking_text": "145|3|2"}, "overall": {"ranking_text": "14|3|25"}} +{"task_type": "background_change", "data_source": "GEdit-Bench", "instruction": "Adjust the background to the ocean.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/background_change/en/ba8c75293f0f60353f6afb4b76e7eda0_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/background_change/en/ba8c75293f0f60353f6afb4b76e7eda0.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/background_change/en/ba8c75293f0f60353f6afb4b76e7eda0.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/background_change/en/ba8c75293f0f60353f6afb4b76e7eda0.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/background_change/en/ba8c75293f0f60353f6afb4b76e7eda0.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/background_change/en/ba8c75293f0f60353f6afb4b76e7eda0.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:34:00.561775", "instruction_following": {"ranking_text": "12|4|35"}, "consistency": {"ranking_text": "12|45|3"}, "overall": {"ranking_text": "2|1|4|5|3"}} +{"task_type": "background_change", "data_source": "GEdit-Bench", "instruction": "Add some snow to the background.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/background_change/en/09e1f235d3d395c3aff0fd36ec3dd034_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/background_change/en/09e1f235d3d395c3aff0fd36ec3dd034.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/background_change/en/09e1f235d3d395c3aff0fd36ec3dd034.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/background_change/en/09e1f235d3d395c3aff0fd36ec3dd034.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/background_change/en/09e1f235d3d395c3aff0fd36ec3dd034.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/background_change/en/09e1f235d3d395c3aff0fd36ec3dd034.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:36:52.768336", "instruction_following": {"ranking_text": "1234|5"}, "consistency": {"ranking_text": "25|14|3"}, "overall": {"ranking_text": "2|1|4|3|5"}} +{"task_type": "background_change", "data_source": "imgedit", "instruction": "Change the garden environment in the picture to a snowy landscape.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/animal/000000265.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/imgedit/224.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/224.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/224.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/224.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/224.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:38:25.923741", "instruction_following": {"ranking_text": "345|12"}, "consistency": {"ranking_text": "1345|2"}, "overall": {"ranking_text": "5|3|4|1|2"}} +{"task_type": "background_change", "data_source": "imgedit", "instruction": "Change the gravel ground in the foreground to a wooden deck setting.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/daily object/000363394.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/imgedit/301.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/301.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/301.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/301.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/301.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:41:17.937299", "instruction_following": {"ranking_text": "1234|5"}, "consistency": {"ranking_text": "4|13|2|5"}, "overall": {"ranking_text": "4|13|2|5"}} +{"task_type": "background_change", "data_source": "imgedit", "instruction": "Change the winding road and mountainous landscape in the background to a cityscape with skyscrapers and busy streets.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/transport/000276156.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/256.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/256.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/256.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/256.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/256.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:43:42.019412", "consistency": {"ranking_text": "1234|5"}, "overall": {"ranking_text": "14|3|2|5"}} +{"task_type": "background_change", "data_source": "imgedit", "instruction": "Change the windmill and houses in the picture from the countryside to a bustling city skyline.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/architecture/000403575.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/283.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/283.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/283.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/283.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/283.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:46:07.222152", "instruction_following": {"ranking_text": "2345|1"}, "consistency": {"ranking_text": "235|4|1"}, "overall": {"ranking_text": "235|4|1"}} +{"task_type": "tone_transfer", "data_source": "GEdit-Bench", "instruction": "change the weather to snowy", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/d83dad4db56f5c6c1270708a74311725_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/tone_transfer/en/d83dad4db56f5c6c1270708a74311725.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/tone_transfer/en/d83dad4db56f5c6c1270708a74311725.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/tone_transfer/en/d83dad4db56f5c6c1270708a74311725.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/tone_transfer/en/d83dad4db56f5c6c1270708a74311725.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/tone_transfer/en/d83dad4db56f5c6c1270708a74311725.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:48:49.464216", "instruction_following": {"ranking_text": "15|4|23"}, "consistency": {"ranking_text": "15|2|3|4"}, "overall": {"ranking_text": "15|2|3|4"}} +{"task_type": "tone_transfer", "data_source": "GEdit-Bench", "instruction": "Apply a filter to make the image brighter.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/d231513192c28e8f14d79a41fd648e9a_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/tone_transfer/en/d231513192c28e8f14d79a41fd648e9a.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/tone_transfer/en/d231513192c28e8f14d79a41fd648e9a.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/tone_transfer/en/d231513192c28e8f14d79a41fd648e9a.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/tone_transfer/en/d231513192c28e8f14d79a41fd648e9a.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/d231513192c28e8f14d79a41fd648e9a.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:50:39.521316", "overall": {"ranking_text": "2|5|34|1"}} +{"task_type": "tone_transfer", "data_source": "GEdit-Bench", "instruction": "Apply a filter to make the image brighter.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/d231513192c28e8f14d79a41fd648e9a_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/tone_transfer/en/d231513192c28e8f14d79a41fd648e9a.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/tone_transfer/en/d231513192c28e8f14d79a41fd648e9a.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/tone_transfer/en/d231513192c28e8f14d79a41fd648e9a.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/tone_transfer/en/d231513192c28e8f14d79a41fd648e9a.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/tone_transfer/en/d231513192c28e8f14d79a41fd648e9a.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:51:39.009140", "instruction_following": {"ranking_text": "345|1|2"}, "overall": {"ranking_text": "34|5|1|2"}} +{"task_type": "tone_transfer", "data_source": "GEdit-Bench", "instruction": "Change the nighttime scene in the image to daytime.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/94ab9306a1ea70be534d9ef36f3a19b0_SRCIMG.png"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image/fullset/tone_transfer/en/94ab9306a1ea70be534d9ef36f3a19b0.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/tone_transfer/en/94ab9306a1ea70be534d9ef36f3a19b0.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/tone_transfer/en/94ab9306a1ea70be534d9ef36f3a19b0.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/tone_transfer/en/94ab9306a1ea70be534d9ef36f3a19b0.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/tone_transfer/en/94ab9306a1ea70be534d9ef36f3a19b0.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:56:53.133787", "instruction_following": {"ranking_text": "124|3|5"}, "consistency": {"ranking_text": "134|2|5"}, "overall": {"ranking_text": "24|1|3|5"}} +{"task_type": "tone_transfer", "data_source": "GEdit-Bench", "instruction": "Enhance the clarity of this photo.", "input_images": ["/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/tone_transfer/en/c09e9a0c550da145d9afe12c543b0048_SRCIMG.png"], "output_images": ["/share/project/jiahao/Bagel/outputs/gen_image/fullset/tone_transfer/en/c09e9a0c550da145d9afe12c543b0048.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/tone_transfer/en/c09e9a0c550da145d9afe12c543b0048.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/tone_transfer/en/c09e9a0c550da145d9afe12c543b0048.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/tone_transfer/en/c09e9a0c550da145d9afe12c543b0048.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/tone_transfer/en/c09e9a0c550da145d9afe12c543b0048.png"], "label": "bad", "timestamp": "2025-08-29T16:57:49.064353"} +{"task_type": "tone_transfer", "data_source": "GEdit-Bench", "instruction": "Enhance and colorize this photo to make the subject more vivid.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/277a863acdd110cc9550f16da754a93d_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/tone_transfer/en/277a863acdd110cc9550f16da754a93d.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/tone_transfer/en/277a863acdd110cc9550f16da754a93d.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/277a863acdd110cc9550f16da754a93d.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/tone_transfer/en/277a863acdd110cc9550f16da754a93d.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/tone_transfer/en/277a863acdd110cc9550f16da754a93d.png"], "label": "triple_ranking", "timestamp": "2025-08-29T16:59:13.782450", "instruction_following": {"ranking_text": "14|5|23"}, "consistency": {"ranking_text": "1234|5"}, "overall": {"ranking_text": "14|5|23"}} +{"task_type": "tone_transfer", "data_source": "GEdit-Bench", "instruction": "Colorize this photo without altering the facial structure.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/8907d6eacd7b91ee4cf8a157802a53a5_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/tone_transfer/en/8907d6eacd7b91ee4cf8a157802a53a5.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/tone_transfer/en/8907d6eacd7b91ee4cf8a157802a53a5.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/tone_transfer/en/8907d6eacd7b91ee4cf8a157802a53a5.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/tone_transfer/en/8907d6eacd7b91ee4cf8a157802a53a5.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/8907d6eacd7b91ee4cf8a157802a53a5.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:01:35.827471", "instruction_following": {"ranking_text": "34|12|5"}, "overall": {"ranking_text": "3|4|12|5"}} +{"task_type": "tone_transfer", "data_source": "GEdit-Bench", "instruction": "change the weather to heavy rain", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/tone_transfer/en/2c7d8b151daa5920c523e40d1dda0d5e_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/tone_transfer/en/2c7d8b151daa5920c523e40d1dda0d5e.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/tone_transfer/en/2c7d8b151daa5920c523e40d1dda0d5e.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/tone_transfer/en/2c7d8b151daa5920c523e40d1dda0d5e.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/tone_transfer/en/2c7d8b151daa5920c523e40d1dda0d5e.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/tone_transfer/en/2c7d8b151daa5920c523e40d1dda0d5e.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:03:31.714768", "instruction_following": {"ranking_text": "45|2|1|3"}, "consistency": {"ranking_text": "123|4|5"}, "overall": {"ranking_text": "4|5|2|1|3"}} +{"task_type": "material_alter", "data_source": "GEdit-Bench", "instruction": "Replace the computer's casing with bamboo fiber composite.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/611ae6fbc57a2b364325650954b21510_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/material_alter/en/611ae6fbc57a2b364325650954b21510.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/material_alter/en/611ae6fbc57a2b364325650954b21510.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/material_alter/en/611ae6fbc57a2b364325650954b21510.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/material_alter/en/611ae6fbc57a2b364325650954b21510.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/material_alter/en/611ae6fbc57a2b364325650954b21510.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:09:07.604399", "instruction_following": {"ranking_text": "35|24|1"}, "consistency": {"ranking_text": "5|3|24|1"}, "overall": {"ranking_text": "5|3|24|1"}} +{"task_type": "material_alter", "data_source": "GEdit-Bench", "instruction": "Replace the sword in the image with a diamond sword.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/f93b1dd57b6a8791c872be6221c66dd0_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/material_alter/en/f93b1dd57b6a8791c872be6221c66dd0.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/material_alter/en/f93b1dd57b6a8791c872be6221c66dd0.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/material_alter/en/f93b1dd57b6a8791c872be6221c66dd0.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/material_alter/en/f93b1dd57b6a8791c872be6221c66dd0.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/material_alter/en/f93b1dd57b6a8791c872be6221c66dd0.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:12:41.837684", "instruction_following": {"ranking_text": "125|4|3"}, "overall": {"ranking_text": "1|5|2|4|3"}} +{"task_type": "material_alter", "data_source": "GEdit-Bench", "instruction": "Replace the bench’s material with marble.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/da8a0c7926b0c53a2c01c3a28e79a2ef_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/material_alter/en/da8a0c7926b0c53a2c01c3a28e79a2ef.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/material_alter/en/da8a0c7926b0c53a2c01c3a28e79a2ef.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/material_alter/en/da8a0c7926b0c53a2c01c3a28e79a2ef.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/material_alter/en/da8a0c7926b0c53a2c01c3a28e79a2ef.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/da8a0c7926b0c53a2c01c3a28e79a2ef.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:14:27.929191", "instruction_following": {"ranking_text": "134|25"}, "overall": {"ranking_text": "14|3|5|2"}} +{"task_type": "material_alter", "data_source": "GEdit-Bench", "instruction": "Transform the donut’s material into aluminum foil.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/625a9a448c17aecb16dce5b0da3075a6_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/material_alter/en/625a9a448c17aecb16dce5b0da3075a6.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/material_alter/en/625a9a448c17aecb16dce5b0da3075a6.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/material_alter/en/625a9a448c17aecb16dce5b0da3075a6.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/material_alter/en/625a9a448c17aecb16dce5b0da3075a6.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/material_alter/en/625a9a448c17aecb16dce5b0da3075a6.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:16:18.655810", "consistency": {"ranking_text": "123|4|5"}, "overall": {"ranking_text": "123|4|5"}} +{"task_type": "material_alter", "data_source": "GEdit-Bench", "instruction": "Change the clothing’s material to foam.", "input_images": ["/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/material_alter/en/43e8fede0b26141d75c64c1f03bfc96e_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/material_alter/en/43e8fede0b26141d75c64c1f03bfc96e.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/material_alter/en/43e8fede0b26141d75c64c1f03bfc96e.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/material_alter/en/43e8fede0b26141d75c64c1f03bfc96e.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/material_alter/en/43e8fede0b26141d75c64c1f03bfc96e.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/material_alter/en/43e8fede0b26141d75c64c1f03bfc96e.png"], "label": "bad", "timestamp": "2025-08-29T17:16:55.748171"} +{"task_type": "material_alter", "data_source": "GEdit-Bench", "instruction": "Replace the doctor's coat with a Merino wool sweater.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/db87dca6363b0c1afd3246ab8fcfe5d7_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/material_alter/en/db87dca6363b0c1afd3246ab8fcfe5d7.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/material_alter/en/db87dca6363b0c1afd3246ab8fcfe5d7.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/material_alter/en/db87dca6363b0c1afd3246ab8fcfe5d7.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/material_alter/en/db87dca6363b0c1afd3246ab8fcfe5d7.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/material_alter/en/db87dca6363b0c1afd3246ab8fcfe5d7.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:19:29.323843", "consistency": {"ranking_text": "34|5|2|1"}, "overall": {"ranking_text": "34|5|2|1"}} +{"task_type": "material_alter", "data_source": "GEdit-Bench", "instruction": "Cutlery is made of food-grade stainless steel.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/db8e18433b727737610cb3d8b71f4690_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/material_alter/en/db8e18433b727737610cb3d8b71f4690.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/material_alter/en/db8e18433b727737610cb3d8b71f4690.png", "/share/project/luoxin/projects/OmniGen/outputs/gedit/fullset/material_alter/en/db8e18433b727737610cb3d8b71f4690.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/material_alter/en/db8e18433b727737610cb3d8b71f4690.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/material_alter/en/db8e18433b727737610cb3d8b71f4690.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:21:58.457896", "instruction_following": {"ranking_text": "245|1|3"}, "consistency": {"ranking_text": "35|2|4|1"}, "overall": {"ranking_text": "5|2|4|1|3"}} +{"task_type": "material_alter", "data_source": "GEdit-Bench", "instruction": "Change the hat’s material to foam plastic.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/f521449fb89e5ded1f4ff725785d01b8_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/material_alter/en/f521449fb89e5ded1f4ff725785d01b8.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/material_alter/en/f521449fb89e5ded1f4ff725785d01b8.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_2/results/fullset/material_alter/en/f521449fb89e5ded1f4ff725785d01b8.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/material_alter/en/f521449fb89e5ded1f4ff725785d01b8.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/material_alter/en/f521449fb89e5ded1f4ff725785d01b8.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:24:37.699113", "instruction_following": {"ranking_text": "34|125"}, "consistency": {"ranking_text": "1345|2"}, "overall": {"ranking_text": "3|4|15|2"}} +{"task_type": "style_change", "data_source": "GEdit-Bench", "instruction": "Replace the image style with a 3D effect.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/style_change/en/74d28960183c3490877d0da05b4ced6b_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/style_change/en/74d28960183c3490877d0da05b4ced6b.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/style_change/en/74d28960183c3490877d0da05b4ced6b.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/style_change/en/74d28960183c3490877d0da05b4ced6b.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/style_change/en/74d28960183c3490877d0da05b4ced6b.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/style_change/en/74d28960183c3490877d0da05b4ced6b.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:25:50.365907", "instruction_following": {"ranking_text": "5|2|134"}, "overall": {"ranking_text": "5|2|134"}} +{"task_type": "style_change", "data_source": "GEdit-Bench", "instruction": "Adjust the image style to a watercolor effect.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/style_change/en/4e62777f17329aff2906ff86a217ccb6_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/style_change/en/4e62777f17329aff2906ff86a217ccb6.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/Qwen-Image-Edit/results_gs4.0/fullset/style_change/en/4e62777f17329aff2906ff86a217ccb6.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench/fullset/style_change/en/4e62777f17329aff2906ff86a217ccb6.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/flux_edit_max/results/fullset/style_change/en/4e62777f17329aff2906ff86a217ccb6.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/style_change/en/4e62777f17329aff2906ff86a217ccb6.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:28:15.286296", "consistency": {"ranking_text": "1234|5"}, "overall": {"ranking_text": "123|4|5"}} +{"task_type": "style_change", "data_source": "GEdit-Bench", "instruction": "Generate a cyberpunk-style photo.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/style_change/en/1021ff6859a5be7b3955a1fc8d1a9431_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/style_change/en/1021ff6859a5be7b3955a1fc8d1a9431.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/style_change/en/1021ff6859a5be7b3955a1fc8d1a9431.png", "/share/project/jiahao/Bagel/outputs/gen_image/fullset/style_change/en/1021ff6859a5be7b3955a1fc8d1a9431.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/style_change/en/1021ff6859a5be7b3955a1fc8d1a9431.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/FLUX-Kontext-dev/results_gs2.5/fullset/style_change/en/1021ff6859a5be7b3955a1fc8d1a9431.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:30:23.524348", "instruction_following": {"ranking_text": "245|1|3"}, "consistency": {"ranking_text": "124|35"}, "overall": {"ranking_text": "4|5|2|1|3"}} +{"task_type": "style_change", "data_source": "GEdit-Bench", "instruction": "Edit this image into a bright and sunny style for use as an avatar.", "input_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/style_change/en/f50278651439a107a2ff7e1b6f76ff08_SRCIMG.png"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gemini_2_5_image_preview/fullset/style_change/en/f50278651439a107a2ff7e1b6f76ff08.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/gpt_image_1/results/fullset/style_change/en/f50278651439a107a2ff7e1b6f76ff08.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/magic_brush/results_gs7.0_ig1.5/fullset/style_change/en/f50278651439a107a2ff7e1b6f76ff08.png", "/share/project/luoxin/projects/Step1X-Edit/results/GEdit-Bench-v1.1/fullset/style_change/en/f50278651439a107a2ff7e1b6f76ff08.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/GEdit-Bench/results/OmniGen2/results_ts5.0_ig1.5/fullset/style_change/en/f50278651439a107a2ff7e1b6f76ff08.png"], "label": "bad", "timestamp": "2025-08-29T17:31:15.446926"} +{"task_type": "style_change", "data_source": "imgedit", "instruction": "Transfer the image into a vibrant graffiti street-mural style.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/style/000015317.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/649.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/649.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/649.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/649.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/649.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:35:13.058195", "instruction_following": {"ranking_text": "1234|5"}, "consistency": {"ranking_text": "123|4|5"}, "overall": {"ranking_text": "3|1|24|5"}} +{"task_type": "style_change", "data_source": "imgedit", "instruction": "Transfer the image into a vibrant graffiti street-mural style.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/style/000280642.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/689.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/689.png", "/share/project/luoxin/projects/OmniGen/outputs/imgedit/689.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/689.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/689.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:38:48.065589", "overall": {"ranking_text": "4|123|5"}} +{"task_type": "style_change", "data_source": "imgedit", "instruction": "Transfer the image into a Lego-brick stop-motion diorama style.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/style/000278574.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/662.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/magic_brush/results_gs7.0_ig1.5/662.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/662.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/gemini_2_5_image_preview/662.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Flux-Kontext-dev/results_gs2.5/662.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:41:30.113954", "instruction_following": {"ranking_text": "3|145|2"}, "consistency": {"ranking_text": "1245|3"}, "overall": {"ranking_text": "3|145|2"}} +{"task_type": "style_change", "data_source": "imgedit", "instruction": "Transfer the image into a colourful ceramic mosaic-tile style.", "input_images": ["/share/project/luoxin/datasets/ImgEdit_Bench/Benchmark/singleturn/style/000280642.jpg"], "output_images": ["/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/Qwen-Image-Edit/results_gs4.0/680.png", "/share/project/jiahao/Bagel/outputs/gen_image_imgedit/680.png", "/share/project/luoxin/projects/OmniGen2-RL/evaluation/imgedit/results/OmniGen2/results_ts5.0_ig1.5/680.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench/680.png", "/share/project/luoxin/projects/Step1X-Edit/results/ImgEdit-Bench-v1.1/680.png"], "label": "triple_ranking", "timestamp": "2025-08-29T17:45:01.772788", "instruction_following": {"ranking_text": "34|5|2|1"}, "consistency": {"ranking_text": "15|24|3"}, "overall": {"ranking_text": "4|3|5|2|1"}}