| import argparse |
| import glob |
| import json |
| import os |
|
|
| import torch |
| from accelerate import PartialState |
| from PIL import Image |
| from tqdm import tqdm |
| from transformers import AutoModel, AutoTokenizer |
|
|
|
|
| class caption_processor: |
| def __init__(self, vlm_name, device): |
| self.vlm = AutoModel.from_pretrained( |
| vlm_name, |
| trust_remote_code=True, |
| attn_implementation="flash_attention_2", |
| torch_dtype=torch.bfloat16, |
| ) |
| self.vlm_tokenizer = AutoTokenizer.from_pretrained( |
| vlm_name, trust_remote_code=True |
| ) |
|
|
| self.vlm = self.vlm.eval().to(device) |
| self.prompt = """ |
| 1. describe the image in brief, Avoid using phrases in [In the/The image/scene shows/contains/is a] in the captions, directly describe the contents. |
| 2. Imagine this picture is the first frame of a 5-second video. Please describe the video and add dynamics, including the movement of objects and themes, as well as the overall camera movement.Avoid using phrases in [In the/The video/scene shows/contains/is a] in the descriptions, directly describe the contents. |
| 3. Please output in JSON format.{"caption": "...","video_description": "..."} |
| """ |
|
|
| def str_2_json(self, str): |
| |
| start_idx = str.find("{") |
| if start_idx == -1: |
| return None |
|
|
| |
| end_idx = str.rfind("}") |
| if end_idx == -1 or end_idx <= start_idx: |
| return None |
|
|
| |
| json_str = str[start_idx : end_idx + 1] |
|
|
| |
| try: |
| import json |
|
|
| return json.loads(json_str) |
| except json.JSONDecodeError: |
| return None |
|
|
| def process(self, image): |
| msgs = [{"role": "user", "content": [image, self.prompt]}] |
|
|
| answer = self.vlm.chat( |
| msgs=msgs, tokenizer=self.vlm_tokenizer, enable_thinking=False, stream=False |
| ) |
|
|
| dict_answer = self.str_2_json(answer) |
| if dict_answer is None: |
| return {"response": answer} |
|
|
| return dict_answer |
|
|
|
|
| def get_images_from_path(path): |
| if os.path.isdir(path): |
| return glob.glob(os.path.join(path, "*.jpg")) + glob.glob( |
| os.path.join(path, "*.png") |
| ) |
| elif os.path.isfile(path) and (path.endswith(".jpg") or path.endswith(".png")): |
| return [path] |
| else: |
| return [] |
|
|
|
|
| def parse_args(): |
| parser = argparse.ArgumentParser(description="Caption processor") |
| parser.add_argument("--vlm_name", type=str, required=True) |
| parser.add_argument("--output_dir", type=str, required=True) |
| parser.add_argument("--paths", type=str, required=True, nargs="+") |
| return parser.parse_args() |
|
|
|
|
| if __name__ == "__main__": |
| distributed_state = PartialState() |
| args = parse_args() |
| output_dir = args.output_dir |
| os.makedirs(output_dir, exist_ok=True) |
|
|
| vlm_name = args.vlm_name |
| paths = args.paths |
| all_paths = [] |
| for path in paths: |
| images = get_images_from_path(path) |
| all_paths.extend(images) |
| print("found", len(all_paths), "images") |
|
|
| processor = caption_processor( |
| vlm_name, |
| distributed_state.device, |
| ) |
| with distributed_state.split_between_processes( |
| all_paths, apply_padding=False |
| ) as batched_paths: |
| print("GPU", distributed_state.device, "found", len(batched_paths), "images") |
|
|
| for path in tqdm(batched_paths, desc="Processing images"): |
| try: |
| json_path = os.path.join(output_dir, os.path.basename(path) + ".json") |
| if os.path.exists(json_path): |
| print(f"File {json_path} already exists, skipping.") |
| continue |
|
|
| image = Image.open(path) |
| output = None |
|
|
| for _ in range(3): |
| output = processor.process(image) |
| if output is not None: |
| break |
|
|
| if output is None: |
| raise Exception("Failed to process image after 3 attempts") |
| else: |
| with open( |
| json_path, |
| "w", |
| encoding="utf-8", |
| ) as f: |
| json.dump(output, f, ensure_ascii=False, indent=2) |
| except Exception as e: |
| print(f"Error processing {path}: {e}") |
|
|