| --- |
| task_categories: |
| - translation |
| language: |
| - en |
| --- |
| # Information |
| * Language: English |
| * The dataset contains both RGB (frontal and side view) and keypoints (only frontal view) data. However, the translation text is only available for frontal-view RGB data. Therefore, this repo only support this type of data. |
| * Gloss is not currently available. |
| * Storage |
| * RGB |
| * Train: 30.7 GB |
| * Validation: 1.65 GB |
| * Test: 2.24 GB |
|
|
| # Structure |
| Each sample will have a structure as follows: |
| ``` |
| { |
| 'VIDEO_ID': Value(dtype='string', id=None), |
| 'VIDEO_NAME': Value(dtype='string', id=None), |
| 'SENTENCE_ID': Value(dtype='string', id=None), |
| 'SENTENCE_NAME': Value(dtype='string', id=None), |
| 'START_REALIGNED': Value(dtype='float64', id=None), |
| 'END_REALIGNED': Value(dtype='float64', id=None), |
| 'SENTENCE': Value(dtype='string', id=None), |
| 'VIDEO': Value(dtype='large_binary', id=None) |
| } |
| |
| { |
| 'VIDEO_ID': '--7E2sU6zP4', |
| 'VIDEO_NAME': '--7E2sU6zP4-5-rgb_front', |
| 'SENTENCE_ID': '--7E2sU6zP4_10', |
| 'SENTENCE_NAME': '--7E2sU6zP4_10-5-rgb_front', |
| 'START_REALIGNED': 129.06, |
| 'END_REALIGNED': 142.48, |
| 'SENTENCE': "And I call them decorative elements because basically all they're meant to do is to enrich and color the page.", |
| 'VIDEO': <video-bytes> |
| } |
| ``` |
|
|
| # How To Use |
| Because the returned video will be in bytes, here is a way to extract frames and fps: |
| ```python |
| # pip install av |
| |
| import av |
| import io |
| import numpy as np |
| import os |
| from datasets import load_dataset |
| |
| |
| def extract_frames(video_bytes): |
| # Create a memory-mapped file from the bytes |
| container = av.open(io.BytesIO(video_bytes)) |
| |
| # Find the video stream |
| visual_stream = next(iter(container.streams.video), None) |
| |
| # Extract video properties |
| video_fps = visual_stream.average_rate |
| |
| # Initialize arrays to store frames |
| frames_array = [] |
| |
| # Extract frames |
| for packet in container.demux([visual_stream]): |
| for frame in packet.decode(): |
| img_array = np.array(frame.to_image()) |
| frames_array.append(img_array) |
| |
| return frames_array, video_fps |
| |
| |
| dataset = load_dataset("VieSignLang/how2sign-clips", split="test", streaming=True) |
| sample = next(iter(dataset))["video"] |
| frames, video_fps = extract_frames(sample) |
| print(f"Number of frames: {frames.shape[0]}") |
| print(f"Video FPS: {video_fps}") |
| ``` |