| |
|
|
| import argparse |
| import multiprocessing as mp |
| import shutil |
| import subprocess |
| import sys |
| import tempfile |
| import time |
| from pathlib import Path |
| from typing import Dict, Iterator, List, Sequence, Tuple |
|
|
| import numpy as np |
| import onnxruntime as ort |
| import torch |
| import torch.nn.functional as F |
| import torchvision.ops as tv_ops |
| from easy_dwpose import DWposeDetector |
| from easy_dwpose.body_estimation import resize_image |
| from easy_dwpose.body_estimation.detector import inference_detector, preprocess as detector_preprocess, demo_postprocess as detector_demo_postprocess |
| from easy_dwpose.body_estimation.pose import ( |
| postprocess as pose_postprocess, |
| preprocess as pose_preprocess, |
| ) |
| from PIL import Image |
|
|
|
|
| REPO_ROOT = Path(__file__).resolve().parents[1] |
| if str(REPO_ROOT) not in sys.path: |
| sys.path.insert(0, str(REPO_ROOT)) |
|
|
| from utils.stats_npz import load_stats, update_video_stats |
|
|
|
|
| DEFAULT_RAW_VIDEO_DIR = REPO_ROOT / "raw_video" |
| DEFAULT_DATASET_DIR = REPO_ROOT / "dataset" |
| DEFAULT_STATS_NPZ = REPO_ROOT / "stats.npz" |
| VIDEO_EXTENSIONS = {".mp4", ".mkv", ".webm", ".mov"} |
| COMPLETE_MARKER_NAME = ".complete" |
|
|
|
|
| def build_optimized_providers(device: str, optimized_provider: str, cache_dir: Path): |
| device = str(device) |
| gpu_id = 0 |
| if ":" in device: |
| gpu_id = int(device.split(":", 1)[1]) |
| cache_dir.mkdir(parents=True, exist_ok=True) |
| if optimized_provider == "tensorrt": |
| providers = ["TensorrtExecutionProvider", "CUDAExecutionProvider", "CPUExecutionProvider"] |
| provider_options = [ |
| { |
| "device_id": str(gpu_id), |
| "trt_engine_cache_enable": "1", |
| "trt_engine_cache_path": str(cache_dir), |
| "trt_timing_cache_enable": "1", |
| "trt_fp16_enable": "1", |
| }, |
| {"device_id": str(gpu_id)}, |
| {}, |
| ] |
| elif optimized_provider == "cuda": |
| providers = ["CUDAExecutionProvider", "CPUExecutionProvider"] |
| provider_options = [ |
| {"device_id": str(gpu_id)}, |
| {}, |
| ] |
| else: |
| providers = ["CPUExecutionProvider"] |
| provider_options = [{}] |
| return providers, provider_options |
|
|
|
|
| def create_detector(device: str, optimized_mode: bool, optimized_provider: str, tmp_root: Path) -> DWposeDetector: |
| detector = DWposeDetector(device=device) |
| if not optimized_mode: |
| return detector |
| providers, provider_options = build_optimized_providers(device, optimized_provider, tmp_root / "ort_trt_cache") |
| detector.pose_estimation.session_det = ort.InferenceSession( |
| "checkpoints/yolox_l.onnx", |
| providers=providers, |
| provider_options=provider_options, |
| ) |
| detector.pose_estimation.session_pose = ort.InferenceSession( |
| "checkpoints/dw-ll_ucoco_384.onnx", |
| providers=providers, |
| provider_options=provider_options, |
| ) |
| return detector |
|
|
|
|
| def parse_args() -> argparse.Namespace: |
| parser = argparse.ArgumentParser( |
| description="Extract DWpose NPZ files from raw videos." |
| ) |
| parser.add_argument("--raw-video-dir", type=Path, default=DEFAULT_RAW_VIDEO_DIR) |
| parser.add_argument("--dataset-dir", type=Path, default=DEFAULT_DATASET_DIR) |
| parser.add_argument("--fps", type=int, default=24) |
| parser.add_argument("--limit", type=int, default=None) |
| parser.add_argument("--workers", type=int, default=None) |
| parser.add_argument("--video-ids", nargs="*", default=None) |
| parser.add_argument("--force", action="store_true") |
| parser.add_argument("--delete-source-on-success", action="store_true") |
| parser.add_argument("--tmp-root", type=Path, default=Path("/tmp")) |
| parser.add_argument("--stats-npz", type=Path, default=DEFAULT_STATS_NPZ) |
| parser.add_argument( |
| "--single-poses-npz", |
| dest="single_poses_npz", |
| action="store_true", |
| default=True, |
| help="Save one aggregated poses.npz per video (default).", |
| ) |
| parser.add_argument( |
| "--per-frame-npz", |
| dest="single_poses_npz", |
| action="store_false", |
| help="Save one NPZ file per frame under the npz directory.", |
| ) |
| parser.add_argument( |
| "--stream-frames", |
| dest="stream_frames", |
| action="store_true", |
| default=True, |
| help="Decode frames directly from ffmpeg stdout without JPG spill (default).", |
| ) |
| parser.add_argument( |
| "--spill-jpg-frames", |
| dest="stream_frames", |
| action="store_false", |
| help="Use legacy ffmpeg-to-JPG spill path for comparison/debugging.", |
| ) |
| parser.add_argument( |
| "--optimized-mode", |
| dest="optimized_mode", |
| action="store_true", |
| default=True, |
| help="Enable optimized ndarray + batched pose inference path (default).", |
| ) |
| parser.add_argument( |
| "--legacy-mode", |
| dest="optimized_mode", |
| action="store_false", |
| help="Disable optimized path and use legacy per-frame single-image inference.", |
| ) |
| parser.add_argument( |
| "--optimized-frame-batch-size", |
| type=int, |
| default=8, |
| help="Frame micro-batch size for optimized pose inference.", |
| ) |
| parser.add_argument( |
| "--optimized-detect-resolution", |
| type=int, |
| default=512, |
| help="Detect resolution used only in optimized mode.", |
| ) |
| parser.add_argument( |
| "--optimized-frame-stride", |
| type=int, |
| default=1, |
| help="Process every Nth decoded frame in optimized mode.", |
| ) |
| parser.add_argument( |
| "--optimized-provider", |
| choices=("tensorrt", "cuda", "cpu"), |
| default="cuda", |
| help="Execution provider used only in optimized mode.", |
| ) |
| parser.add_argument( |
| "--optimized-gpu-pose-preprocess", |
| action="store_true", |
| help="Experimental: move pose crop affine/normalize to GPU in optimized mode.", |
| ) |
| parser.add_argument( |
| "--optimized-gpu-detector-postprocess", |
| action="store_true", |
| help="Experimental: run detector postprocess and NMS on GPU in optimized mode.", |
| ) |
| parser.add_argument( |
| "--optimized-io-binding", |
| action="store_true", |
| help="Experimental: use ONNX Runtime IO binding in optimized mode.", |
| ) |
| return parser.parse_args() |
|
|
|
|
| def select_video_paths(args: argparse.Namespace) -> List[Path]: |
| args.raw_video_dir.mkdir(parents=True, exist_ok=True) |
| args.dataset_dir.mkdir(parents=True, exist_ok=True) |
| video_id_filter = set(args.video_ids or []) |
| stats = load_stats(args.stats_npz) |
|
|
| selected = [] |
| for path in sorted(args.raw_video_dir.iterdir()): |
| if not path.is_file() or path.suffix.lower() not in VIDEO_EXTENSIONS: |
| continue |
| video_id = path.stem |
| if video_id_filter and video_id not in video_id_filter: |
| continue |
| npz_dir = args.dataset_dir / video_id / "npz" |
| complete_marker = npz_dir / COMPLETE_MARKER_NAME |
| if ( |
| not args.force |
| and npz_dir.exists() |
| and complete_marker.exists() |
| and stats.get(video_id, {}).get("process_status") == "ok" |
| ): |
| continue |
| selected.append(path) |
| if args.limit is not None and len(selected) >= args.limit: |
| break |
| return selected |
|
|
|
|
| def extract_frames_to_jpg(video_path: Path, frame_dir: Path, fps: int) -> None: |
| command = [ |
| "ffmpeg", |
| "-hide_banner", |
| "-loglevel", |
| "error", |
| "-y", |
| "-i", |
| str(video_path), |
| "-vf", |
| f"fps={fps}", |
| str(frame_dir / "%08d.jpg"), |
| ] |
| subprocess.run(command, check=True) |
|
|
|
|
| def probe_video_dimensions(video_path: Path) -> Tuple[int, int]: |
| command = [ |
| "ffprobe", |
| "-v", |
| "error", |
| "-select_streams", |
| "v:0", |
| "-show_entries", |
| "stream=width,height", |
| "-of", |
| "csv=p=0:s=x", |
| str(video_path), |
| ] |
| proc = subprocess.run(command, check=True, capture_output=True, text=True) |
| dims = (proc.stdout or "").strip() |
| if "x" not in dims: |
| raise RuntimeError(f"Unable to parse ffprobe dimensions for {video_path.name}: {dims!r}") |
| width_s, height_s = dims.split("x", 1) |
| return int(width_s), int(height_s) |
|
|
|
|
| def iter_streamed_frames(video_path: Path, fps: int) -> Iterator[Tuple[int, np.ndarray, int, int]]: |
| width, height = probe_video_dimensions(video_path) |
| frame_bytes = width * height * 3 |
| command = [ |
| "ffmpeg", |
| "-hide_banner", |
| "-loglevel", |
| "error", |
| "-i", |
| str(video_path), |
| "-vf", |
| f"fps={fps}", |
| "-f", |
| "rawvideo", |
| "-pix_fmt", |
| "rgb24", |
| "pipe:1", |
| ] |
| proc = subprocess.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE) |
| assert proc.stdout is not None |
| try: |
| frame_index = 0 |
| while True: |
| chunk = proc.stdout.read(frame_bytes) |
| if not chunk: |
| break |
| if len(chunk) != frame_bytes: |
| raise RuntimeError( |
| f"Short raw frame read for {video_path.name}: expected {frame_bytes} bytes, got {len(chunk)}" |
| ) |
| frame_index += 1 |
| frame_array = np.frombuffer(chunk, dtype=np.uint8).reshape((height, width, 3)) |
| yield frame_index, frame_array, width, height |
| finally: |
| if proc.stdout: |
| proc.stdout.close() |
| stderr = proc.stderr.read().decode("utf-8", errors="replace") if proc.stderr else "" |
| if proc.stderr: |
| proc.stderr.close() |
| returncode = proc.wait() |
| if returncode != 0: |
| raise RuntimeError(f"ffmpeg raw frame stream failed for {video_path.name}: {stderr.strip()}") |
|
|
|
|
| def build_npz_payload(pose_data: Dict[str, np.ndarray], width: int, height: int) -> Dict[str, np.ndarray]: |
| num_persons = int(pose_data["faces"].shape[0]) if "faces" in pose_data else 0 |
| payload: Dict[str, np.ndarray] = { |
| "num_persons": np.asarray(num_persons, dtype=np.int32), |
| "frame_width": np.asarray(width, dtype=np.int32), |
| "frame_height": np.asarray(height, dtype=np.int32), |
| } |
| if num_persons == 0: |
| return payload |
|
|
| bodies = pose_data["bodies"].reshape(num_persons, 18, 2).astype(np.float32, copy=False) |
| body_scores = pose_data["body_scores"].astype(np.float32, copy=False) |
| faces = pose_data["faces"].astype(np.float32, copy=False) |
| face_scores = pose_data["faces_scores"].astype(np.float32, copy=False) |
| hands = pose_data["hands"].astype(np.float32, copy=False) |
| hand_scores = pose_data["hands_scores"].astype(np.float32, copy=False) |
|
|
| for person_idx in range(num_persons): |
| prefix = f"person_{person_idx:03d}" |
| payload[f"{prefix}_body_keypoints"] = bodies[person_idx] |
| payload[f"{prefix}_body_scores"] = body_scores[person_idx] |
| payload[f"{prefix}_face_keypoints"] = faces[person_idx] |
| payload[f"{prefix}_face_scores"] = face_scores[person_idx] |
| left_hand_idx = person_idx * 2 |
| right_hand_idx = left_hand_idx + 1 |
| if left_hand_idx < len(hands): |
| payload[f"{prefix}_left_hand_keypoints"] = hands[left_hand_idx] |
| payload[f"{prefix}_left_hand_scores"] = hand_scores[left_hand_idx] |
| if right_hand_idx < len(hands): |
| payload[f"{prefix}_right_hand_keypoints"] = hands[right_hand_idx] |
| payload[f"{prefix}_right_hand_scores"] = hand_scores[right_hand_idx] |
| return payload |
|
|
|
|
| def run_session_outputs( |
| session: ort.InferenceSession, |
| input_array: np.ndarray, |
| use_io_binding: bool, |
| device_id: int, |
| ): |
| input_name = session.get_inputs()[0].name |
| output_names = [out.name for out in session.get_outputs()] |
| if not use_io_binding: |
| return session.run(output_names, {input_name: input_array}) |
| io_binding = session.io_binding() |
| io_binding.bind_cpu_input(input_name, input_array) |
| for output_name in output_names: |
| io_binding.bind_output(output_name, device_type="cuda", device_id=device_id) |
| session.run_with_iobinding(io_binding) |
| return io_binding.copy_outputs_to_cpu() |
|
|
|
|
| def inference_detector_gpu_postprocess( |
| session: ort.InferenceSession, |
| ori_img: np.ndarray, |
| device: torch.device, |
| use_io_binding: bool, |
| device_id: int, |
| ) -> np.ndarray: |
| input_shape = (640, 640) |
| img, ratio = detector_preprocess(ori_img, input_shape) |
| outputs = run_session_outputs(session, img[None, :, :, :], use_io_binding, device_id) |
| predictions = detector_demo_postprocess(outputs[0], input_shape)[0] |
| pred = torch.from_numpy(np.ascontiguousarray(predictions)).to(device=device, dtype=torch.float32) |
| boxes = pred[:, :4] |
| score_obj = pred[:, 4] |
| cls_scores = pred[:, 5:] |
| if cls_scores.ndim == 1: |
| cls_scores = cls_scores.unsqueeze(1) |
| cls0 = score_obj * cls_scores[:, 0] |
| mask = cls0 > 0.1 |
| if not torch.any(mask): |
| return np.array([]) |
| boxes = boxes[mask] |
| cls0 = cls0[mask] |
| boxes_xyxy = torch.empty_like(boxes) |
| boxes_xyxy[:, 0] = boxes[:, 0] - boxes[:, 2] / 2.0 |
| boxes_xyxy[:, 1] = boxes[:, 1] - boxes[:, 3] / 2.0 |
| boxes_xyxy[:, 2] = boxes[:, 0] + boxes[:, 2] / 2.0 |
| boxes_xyxy[:, 3] = boxes[:, 1] + boxes[:, 3] / 2.0 |
| boxes_xyxy /= ratio |
| keep = tv_ops.nms(boxes_xyxy, cls0, 0.45) |
| if keep.numel() == 0: |
| return np.array([]) |
| final_boxes = boxes_xyxy[keep] |
| final_scores = cls0[keep] |
| final_boxes = final_boxes[final_scores > 0.3] |
| if final_boxes.numel() == 0: |
| return np.array([]) |
| return final_boxes.detach().cpu().numpy() |
|
|
|
|
| def optimized_detector_call( |
| detector: DWposeDetector, |
| frame: np.ndarray, |
| detect_resolution: int, |
| include_hands: bool = True, |
| include_face: bool = True, |
| ) -> Dict[str, np.ndarray]: |
| del include_hands, include_face |
| return optimized_process_frame_batch(detector, [(1, frame, 0, 0)], detect_resolution)[0][1] |
|
|
|
|
| def empty_pose_payload() -> Dict[str, np.ndarray]: |
| empty_f = np.zeros((0,), dtype=np.float32) |
| return { |
| "bodies": empty_f.reshape(0, 2), |
| "body_scores": empty_f.reshape(0, 18), |
| "hands": empty_f.reshape(0, 21, 2), |
| "hands_scores": empty_f.reshape(0, 21), |
| "faces": empty_f.reshape(0, 68, 2), |
| "faces_scores": empty_f.reshape(0, 68), |
| } |
|
|
|
|
| def format_pose_output( |
| detector: DWposeDetector, |
| keypoints: np.ndarray, |
| scores: np.ndarray, |
| width: int, |
| height: int, |
| ) -> Dict[str, np.ndarray]: |
| keypoints_info = np.concatenate((keypoints, scores[..., None]), axis=-1) |
| neck = np.mean(keypoints_info[:, [5, 6]], axis=1) |
| neck[:, 2:4] = np.logical_and( |
| keypoints_info[:, 5, 2:4] > 0.3, |
| keypoints_info[:, 6, 2:4] > 0.3, |
| ).astype(int) |
| new_keypoints_info = np.insert(keypoints_info, 17, neck, axis=1) |
| mmpose_idx = [17, 6, 8, 10, 7, 9, 12, 14, 16, 13, 15, 2, 1, 4, 3] |
| openpose_idx = [1, 2, 3, 4, 6, 7, 8, 9, 10, 12, 13, 14, 15, 16, 17] |
| new_keypoints_info[:, openpose_idx] = new_keypoints_info[:, mmpose_idx] |
| keypoints_info = new_keypoints_info |
|
|
| keypoints = keypoints_info[..., :2] |
| scores = keypoints_info[..., 2] |
| return detector._format_pose(keypoints, scores, width, height) |
|
|
|
|
| def prepare_optimized_frame( |
| detector: DWposeDetector, |
| frame: np.ndarray, |
| detect_resolution: int, |
| ) -> Tuple[np.ndarray, np.ndarray, int, int]: |
| image = frame |
| if not isinstance(image, np.ndarray): |
| image = np.asarray(image.convert("RGB")) |
| image = resize_image(np.ascontiguousarray(image), target_resolution=detect_resolution) |
| height, width = image.shape[:2] |
| if getattr(detector, "_optimized_gpu_detector_postprocess", False): |
| det_result = inference_detector_gpu_postprocess( |
| detector.pose_estimation.session_det, |
| image, |
| detector._optimized_torch_device, |
| getattr(detector, "_optimized_io_binding", False), |
| getattr(detector, "_optimized_device_id", 0), |
| ) |
| else: |
| det_result = inference_detector(detector.pose_estimation.session_det, image) |
| return image, det_result, width, height |
|
|
|
|
| def gpu_pose_preprocess( |
| image: np.ndarray, |
| out_bbox: np.ndarray, |
| input_size: Tuple[int, int], |
| device: torch.device, |
| ) -> Tuple[np.ndarray, List[np.ndarray], List[np.ndarray]]: |
| if len(out_bbox) == 0: |
| return np.empty((0, 3, input_size[1], input_size[0]), dtype=np.float32), [], [] |
| img_t = torch.from_numpy(np.ascontiguousarray(image)).to(device=device, dtype=torch.float32) |
| img_t = img_t.permute(2, 0, 1).unsqueeze(0) |
| H, W = image.shape[:2] |
| out_w, out_h = input_size |
| boxes = np.asarray(out_bbox, dtype=np.float32) |
| x0 = boxes[:, 0] |
| y0 = boxes[:, 1] |
| x1 = boxes[:, 2] |
| y1 = boxes[:, 3] |
| centers = np.stack([(x0 + x1) * 0.5, (y0 + y1) * 0.5], axis=1).astype(np.float32) |
| scales = np.stack([(x1 - x0) * 1.25, (y1 - y0) * 1.25], axis=1).astype(np.float32) |
| aspect = out_w / out_h |
| w = scales[:, 0:1] |
| h = scales[:, 1:2] |
| scales = np.where(w > h * aspect, np.concatenate([w, w / aspect], axis=1), np.concatenate([h * aspect, h], axis=1)).astype(np.float32) |
| centers_t = torch.from_numpy(centers).to(device=device, dtype=torch.float32) |
| scales_t = torch.from_numpy(scales).to(device=device, dtype=torch.float32) |
| theta = torch.zeros((len(boxes), 2, 3), device=device, dtype=torch.float32) |
| theta[:, 0, 0] = scales_t[:, 0] / max(W - 1, 1) |
| theta[:, 1, 1] = scales_t[:, 1] / max(H - 1, 1) |
| theta[:, 0, 2] = 2.0 * centers_t[:, 0] / max(W - 1, 1) - 1.0 |
| theta[:, 1, 2] = 2.0 * centers_t[:, 1] / max(H - 1, 1) - 1.0 |
| grid = F.affine_grid(theta, size=(len(boxes), 3, out_h, out_w), align_corners=True) |
| crops = F.grid_sample(img_t.expand(len(boxes), -1, -1, -1), grid, mode='bilinear', padding_mode='zeros', align_corners=True) |
| mean = torch.tensor([123.675, 116.28, 103.53], device=device, dtype=torch.float32).view(1, 3, 1, 1) |
| std = torch.tensor([58.395, 57.12, 57.375], device=device, dtype=torch.float32).view(1, 3, 1, 1) |
| crops = (crops - mean) / std |
| return crops.detach().cpu().numpy(), [c for c in centers], [s for s in scales] |
|
|
|
|
| def optimized_process_frame_batch( |
| detector: DWposeDetector, |
| frames: Sequence[Tuple[int, np.ndarray, int, int]], |
| detect_resolution: int, |
| ) -> List[Tuple[int, Dict[str, np.ndarray], int, int]]: |
| session_pose = detector.pose_estimation.session_pose |
| model_input = session_pose.get_inputs()[0].shape |
| model_input_size = (model_input[3], model_input[2]) |
|
|
| prepared = [] |
| pose_inputs = [] |
| all_centers = [] |
| all_scales = [] |
|
|
| for frame_index, frame, width, height in frames: |
| input_image, det_result, input_width, input_height = prepare_optimized_frame(detector, frame, detect_resolution) |
| if len(det_result) == 0: |
| prepared.append((frame_index, empty_pose_payload(), width, height, 0, input_width, input_height)) |
| continue |
| torch_device = getattr(detector, "_optimized_torch_device", None) |
| if torch_device is not None: |
| batch_imgs, centers, scales = gpu_pose_preprocess(input_image, det_result, model_input_size, torch_device) |
| count = int(batch_imgs.shape[0]) |
| prepared.append((frame_index, None, width, height, count, input_width, input_height)) |
| if count: |
| pose_inputs.extend(list(batch_imgs)) |
| all_centers.extend(centers) |
| all_scales.extend(scales) |
| continue |
| resized_imgs, centers, scales = pose_preprocess(input_image, det_result, model_input_size) |
| count = len(resized_imgs) |
| prepared.append((frame_index, None, width, height, count, input_width, input_height)) |
| pose_inputs.extend([img.transpose(2, 0, 1) for img in resized_imgs]) |
| all_centers.extend(centers) |
| all_scales.extend(scales) |
|
|
| if pose_inputs: |
| batch = np.stack(pose_inputs, axis=0).astype(np.float32, copy=False) |
| sess_input = {session_pose.get_inputs()[0].name: batch} |
| sess_output = [out.name for out in session_pose.get_outputs()] |
| simcc_x, simcc_y = session_pose.run(sess_output, sess_input) |
| batched_outputs = [(simcc_x[i : i + 1], simcc_y[i : i + 1]) for i in range(batch.shape[0])] |
| keypoints_all, scores_all = pose_postprocess( |
| batched_outputs, |
| model_input_size, |
| all_centers, |
| all_scales, |
| ) |
| else: |
| keypoints_all = scores_all = None |
|
|
| results = [] |
| offset = 0 |
| for frame_index, pose_data, width, height, count, input_width, input_height in prepared: |
| if count == 0: |
| results.append((frame_index, pose_data, width, height)) |
| continue |
| keypoints = keypoints_all[offset : offset + count] |
| scores = scores_all[offset : offset + count] |
| offset += count |
| results.append( |
| ( |
| frame_index, |
| format_pose_output(detector, keypoints, scores, input_width, input_height), |
| width, |
| height, |
| ) |
| ) |
| return results |
|
|
|
|
| def process_video( |
| video_path: Path, |
| dataset_dir: Path, |
| fps: int, |
| detector: DWposeDetector, |
| tmp_root: Path, |
| force: bool, |
| single_poses_npz: bool, |
| stream_frames: bool, |
| optimized_mode: bool, |
| optimized_frame_batch_size: int, |
| optimized_detect_resolution: int, |
| optimized_frame_stride: int, |
| ) -> None: |
| video_id = video_path.stem |
| output_npz_dir = dataset_dir / video_id / "npz" |
| complete_marker = output_npz_dir / COMPLETE_MARKER_NAME |
| poses_npz_path = output_npz_dir / "poses.npz" |
| if output_npz_dir.exists() and complete_marker.exists() and not force: |
| print(f"Skip {video_id}: NPZ files already exist") |
| return |
|
|
| if output_npz_dir.exists() and (force or not complete_marker.exists()): |
| shutil.rmtree(output_npz_dir) |
| output_npz_dir.mkdir(parents=True, exist_ok=True) |
|
|
| aggregated_payloads = [] |
| frame_widths = [] |
| frame_heights = [] |
| frame_indices = [] |
| total_frames = 0 |
| decode_start = time.perf_counter() |
| process_start = decode_start |
|
|
| if stream_frames: |
| print(f"{video_id}: decoding mode=stream fps={fps} optimized={optimized_mode}") |
| frame_batch = [] |
| batch_size = max(1, optimized_frame_batch_size) |
| frame_stride = max(1, optimized_frame_stride) |
| for frame_index, frame, width, height in iter_streamed_frames(video_path, fps): |
| total_frames = frame_index |
| if optimized_mode: |
| if ((frame_index - 1) % frame_stride) != 0: |
| continue |
| frame_batch.append((frame_index, frame, width, height)) |
| if len(frame_batch) < batch_size: |
| continue |
| batch_results = optimized_process_frame_batch(detector, frame_batch, optimized_detect_resolution) |
| frame_batch = [] |
| for result_index, pose_data, result_width, result_height in batch_results: |
| payload = build_npz_payload(pose_data, result_width, result_height) |
| if single_poses_npz: |
| aggregated_payloads.append(payload) |
| frame_widths.append(result_width) |
| frame_heights.append(result_height) |
| frame_indices.append(result_index) |
| else: |
| np.savez(output_npz_dir / f"{result_index:08d}.npz", **payload) |
| if result_index == 1 or result_index % 100 == 0: |
| print(f"{video_id}: processed {result_index} frames") |
| continue |
| pose_data = detector(frame, draw_pose=False, include_hands=True, include_face=True) |
| payload = build_npz_payload(pose_data, width, height) |
| if single_poses_npz: |
| aggregated_payloads.append(payload) |
| frame_widths.append(width) |
| frame_heights.append(height) |
| else: |
| np.savez(output_npz_dir / f"{frame_index:08d}.npz", **payload) |
| if frame_index == 1 or frame_index % 100 == 0: |
| print(f"{video_id}: processed {frame_index} frames") |
| if optimized_mode and frame_batch: |
| for result_index, pose_data, result_width, result_height in optimized_process_frame_batch(detector, frame_batch, optimized_detect_resolution): |
| payload = build_npz_payload(pose_data, result_width, result_height) |
| if single_poses_npz: |
| aggregated_payloads.append(payload) |
| frame_widths.append(result_width) |
| frame_heights.append(result_height) |
| frame_indices.append(result_index) |
| else: |
| np.savez(output_npz_dir / f"{result_index:08d}.npz", **payload) |
| if result_index == 1 or result_index % 100 == 0: |
| print(f"{video_id}: processed {result_index} frames") |
| else: |
| print(f"{video_id}: decoding mode=jpg-spill fps={fps} optimized={optimized_mode}") |
| tmp_root.mkdir(parents=True, exist_ok=True) |
| frame_dir = Path(tempfile.mkdtemp(prefix=f"sign_dwpose_{video_id}_", dir=str(tmp_root))) |
| try: |
| extract_frames_to_jpg(video_path, frame_dir, fps) |
| frame_paths = sorted(frame_dir.glob("*.jpg")) |
| total_frames = len(frame_paths) |
| print(f"{video_id}: extracted {total_frames} frames at {fps} fps") |
| process_start = time.perf_counter() |
| for frame_index, frame_path in enumerate(frame_paths, start=1): |
| with Image.open(frame_path) as image: |
| frame = np.asarray(image.convert("RGB")) |
| height, width = frame.shape[:2] |
| if optimized_mode: |
| pose_data = optimized_detector_call( |
| detector, |
| frame, |
| optimized_detect_resolution, |
| include_hands=True, |
| include_face=True, |
| ) |
| else: |
| pose_data = detector(frame, draw_pose=False, include_hands=True, include_face=True) |
| payload = build_npz_payload(pose_data, width, height) |
| if single_poses_npz: |
| aggregated_payloads.append(payload) |
| frame_widths.append(width) |
| frame_heights.append(height) |
| frame_indices.append(frame_index) |
| frame_indices.append(frame_index) |
| else: |
| np.savez(output_npz_dir / f"{frame_index:08d}.npz", **payload) |
| if frame_index == 1 or frame_index % 100 == 0 or frame_index == total_frames: |
| print(f"{video_id}: processed {frame_index}/{total_frames} frames") |
| finally: |
| shutil.rmtree(frame_dir, ignore_errors=True) |
|
|
| decode_process_elapsed = time.perf_counter() - decode_start |
| print(f"{video_id}: processed total_frames={total_frames} elapsed={decode_process_elapsed:.2f}s mode={'stream' if stream_frames else 'jpg-spill'} optimized={optimized_mode}") |
|
|
| if single_poses_npz: |
| np.savez( |
| poses_npz_path, |
| video_id=np.asarray(video_id), |
| fps=np.asarray(fps, dtype=np.int32), |
| total_frames=np.asarray(total_frames, dtype=np.int32), |
| frame_widths=np.asarray(frame_widths, dtype=np.int32), |
| frame_heights=np.asarray(frame_heights, dtype=np.int32), |
| frame_indices=np.asarray(frame_indices, dtype=np.int32), |
| frame_payloads=np.asarray(aggregated_payloads, dtype=object), |
| ) |
|
|
| complete_marker.write_text( |
| f"video_id={video_id}\nfps={fps}\nframes={total_frames}\noutput_mode={'single_poses_npy' if single_poses_npz else 'per_frame_npz'}\ndecode_mode={'stream' if stream_frames else 'jpg-spill'}\noptimized_mode={optimized_mode}\noptimized_detect_resolution={optimized_detect_resolution}\noptimized_frame_stride={optimized_frame_stride}\ncompleted_at={time.strftime('%Y-%m-%d %H:%M:%S')}\n", |
| encoding="utf-8", |
| ) |
|
|
|
|
| def worker(rank: int, worker_count: int, video_paths: Sequence[Path], args: argparse.Namespace) -> None: |
| if not torch.cuda.is_available(): |
| raise RuntimeError("CUDA is not available; refusing to run DWpose on CPU") |
| device_count = torch.cuda.device_count() |
| if device_count <= rank: |
| raise RuntimeError( |
| f"CUDA device rank {rank} is unavailable; visible device_count={device_count}" |
| ) |
| device = f"cuda:{rank}" |
| detector = create_detector( |
| device=device, |
| optimized_mode=args.optimized_mode, |
| optimized_provider=args.optimized_provider, |
| tmp_root=args.tmp_root, |
| ) |
| if args.optimized_mode: |
| detector._optimized_device_id = int(device.split(":", 1)[1]) if ":" in device else 0 |
| detector._optimized_io_binding = bool(args.optimized_io_binding) |
| detector._optimized_gpu_detector_postprocess = bool(args.optimized_gpu_detector_postprocess) |
| if args.optimized_mode and (args.optimized_gpu_pose_preprocess or args.optimized_gpu_detector_postprocess): |
| detector._optimized_torch_device = torch.device(device) |
| print(f"Worker {rank}: device={device}, cuda_device_count={device_count}", flush=True) |
|
|
| for index, video_path in enumerate(video_paths): |
| if index % worker_count != rank: |
| continue |
| try: |
| update_video_stats( |
| args.stats_npz, |
| video_path.stem, |
| process_status="running", |
| last_error="", |
| updated_at=time.strftime("%Y-%m-%d %H:%M:%S"), |
| ) |
| process_video( |
| video_path=video_path, |
| dataset_dir=args.dataset_dir, |
| fps=args.fps, |
| detector=detector, |
| tmp_root=args.tmp_root, |
| force=args.force, |
| single_poses_npz=args.single_poses_npz, |
| stream_frames=args.stream_frames, |
| optimized_mode=args.optimized_mode, |
| optimized_frame_batch_size=args.optimized_frame_batch_size, |
| optimized_detect_resolution=args.optimized_detect_resolution, |
| optimized_frame_stride=args.optimized_frame_stride, |
| ) |
| update_video_stats( |
| args.stats_npz, |
| video_path.stem, |
| process_status="ok", |
| last_error="", |
| updated_at=time.strftime("%Y-%m-%d %H:%M:%S"), |
| ) |
| if args.delete_source_on_success and video_path.exists(): |
| video_path.unlink() |
| print(f"Worker {rank}: deleted source video {video_path.name}") |
| except Exception as exc: |
| update_video_stats( |
| args.stats_npz, |
| video_path.stem, |
| process_status="failed", |
| last_error=str(exc), |
| updated_at=time.strftime("%Y-%m-%d %H:%M:%S"), |
| ) |
| print(f"Worker {rank}: failed on {video_path.name}: {exc}") |
|
|
|
|
| def main() -> None: |
| args = parse_args() |
| video_paths = select_video_paths(args) |
| if not video_paths: |
| print("No videos need DWpose extraction.") |
| return |
|
|
| if not torch.cuda.is_available(): |
| raise RuntimeError("CUDA is not available; refusing to run DWpose on CPU") |
|
|
| visible_gpu_count = torch.cuda.device_count() |
| if visible_gpu_count < 1: |
| raise RuntimeError("No CUDA devices are visible to the DWpose worker") |
|
|
| if args.workers is not None: |
| worker_count = max(1, args.workers) |
| else: |
| worker_count = visible_gpu_count |
| worker_count = max(1, worker_count) |
| if worker_count > visible_gpu_count: |
| raise RuntimeError( |
| f"Requested workers={worker_count}, but only {visible_gpu_count} CUDA device(s) are visible" |
| ) |
| worker_count = min(worker_count, len(video_paths)) |
| print(f"DWpose main: visible_cuda_devices={visible_gpu_count}, worker_count={worker_count}, stream_frames={args.stream_frames}, optimized_mode={args.optimized_mode}, optimized_frame_batch_size={args.optimized_frame_batch_size}, optimized_detect_resolution={args.optimized_detect_resolution}, optimized_frame_stride={args.optimized_frame_stride}, optimized_provider={args.optimized_provider}, optimized_gpu_pose_preprocess={args.optimized_gpu_pose_preprocess}, optimized_gpu_detector_postprocess={args.optimized_gpu_detector_postprocess}, optimized_io_binding={args.optimized_io_binding}", flush=True) |
|
|
| if worker_count == 1: |
| worker(0, 1, video_paths, args) |
| return |
|
|
| mp.set_start_method("spawn", force=True) |
| processes = [] |
| for rank in range(worker_count): |
| process = mp.Process(target=worker, args=(rank, worker_count, video_paths, args)) |
| process.start() |
| processes.append(process) |
|
|
| failed = False |
| for process in processes: |
| process.join() |
| failed = failed or process.exitcode != 0 |
| if failed: |
| raise SystemExit("One or more DWpose workers failed.") |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|