# coding: utf-8 """ Functions for processing video ATTENTION: you need to install ffmpeg and ffprobe in your env! """ import os.path as osp import numpy as np import subprocess import imageio import cv2 from rich.progress import track from .rprint import rlog as log from .rprint import rprint as print from .helper import prefix def exec_cmd(cmd): subprocess.run(cmd, shell=True, check=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) def images2video(images, wfp, **kwargs): fps = kwargs.get('fps', 30) video_format = kwargs.get('format', 'mp4') # default is mp4 format codec = kwargs.get('codec', 'libx264') # default is libx264 encoding quality = kwargs.get('quality') # video quality pixelformat = kwargs.get('pixelformat', 'yuv420p') # video pixel format image_mode = kwargs.get('image_mode', 'rgb') macro_block_size = kwargs.get('macro_block_size', 2) ffmpeg_params = ['-crf', str(kwargs.get('crf', 18))] writer = imageio.get_writer( wfp, fps=fps, format=video_format, codec=codec, quality=quality, ffmpeg_params=ffmpeg_params, pixelformat=pixelformat, macro_block_size=macro_block_size ) n = len(images) for i in track(range(n), description='Writing', transient=True): if image_mode.lower() == 'bgr': writer.append_data(images[i][..., ::-1]) else: writer.append_data(images[i]) writer.close() def video2gif(video_fp, fps=30, size=256): if osp.exists(video_fp): d = osp.split(video_fp)[0] fn = prefix(osp.basename(video_fp)) palette_wfp = osp.join(d, 'palette.png') gif_wfp = osp.join(d, f'{fn}.gif') # generate the palette cmd = f'ffmpeg -i {video_fp} -vf "fps={fps},scale={size}:-1:flags=lanczos,palettegen" {palette_wfp} -y' exec_cmd(cmd) # use the palette to generate the gif cmd = f'ffmpeg -i {video_fp} -i {palette_wfp} -filter_complex "fps={fps},scale={size}:-1:flags=lanczos[x];[x][1:v]paletteuse" {gif_wfp} -y' exec_cmd(cmd) else: print(f'video_fp: {video_fp} not exists!') def merge_audio_video(video_fp, audio_fp, wfp): if osp.exists(video_fp) and osp.exists(audio_fp): cmd = f'ffmpeg -i {video_fp} -i {audio_fp} -c:v copy -c:a aac {wfp} -y' exec_cmd(cmd) print(f'merge {video_fp} and {audio_fp} to {wfp}') else: print(f'video_fp: {video_fp} or audio_fp: {audio_fp} not exists!') def blend(img: np.ndarray, mask: np.ndarray, background_color=(255, 255, 255)): mask_float = mask.astype(np.float32) / 255. background_color = np.array(background_color).reshape([1, 1, 3]) bg = np.ones_like(img) * background_color img = np.clip(mask_float * img + (1 - mask_float) * bg, 0, 255).astype(np.uint8) return img def concat_frames(driving_image_lst, source_image, I_p_lst): # TODO: add more concat style, e.g., left-down corner driving out_lst = [] h, w, _ = I_p_lst[0].shape for idx, _ in track(enumerate(I_p_lst), total=len(I_p_lst), description='Concatenating result...'): I_p = I_p_lst[idx] source_image_resized = cv2.resize(source_image, (w, h)) if driving_image_lst is None: out = np.hstack((source_image_resized, I_p)) else: driving_image = driving_image_lst[idx] driving_image_resized = cv2.resize(driving_image, (w, h)) out = np.hstack((driving_image_resized, source_image_resized, I_p)) out_lst.append(out) return out_lst class VideoWriter: def __init__(self, **kwargs): self.fps = kwargs.get('fps', 30) self.wfp = kwargs.get('wfp', 'video.mp4') self.video_format = kwargs.get('format', 'mp4') self.codec = kwargs.get('codec', 'libx264') self.quality = kwargs.get('quality') self.pixelformat = kwargs.get('pixelformat', 'yuv420p') self.image_mode = kwargs.get('image_mode', 'rgb') self.ffmpeg_params = kwargs.get('ffmpeg_params') self.writer = imageio.get_writer( self.wfp, fps=self.fps, format=self.video_format, codec=self.codec, quality=self.quality, ffmpeg_params=self.ffmpeg_params, pixelformat=self.pixelformat ) def write(self, image): if self.image_mode.lower() == 'bgr': self.writer.append_data(image[..., ::-1]) else: self.writer.append_data(image) def close(self): if self.writer is not None: self.writer.close() def change_video_fps(input_file, output_file, fps=20, codec='libx264', crf=5): cmd = f"ffmpeg -i {input_file} -c:v {codec} -crf {crf} -r {fps} {output_file} -y" exec_cmd(cmd) def get_fps(filepath, default_fps=25): try: fps = cv2.VideoCapture(filepath).get(cv2.CAP_PROP_FPS) if fps in (0, None): fps = default_fps except Exception as e: log(e) fps = default_fps return fps def has_audio_stream(video_path: str) -> bool: """ Check if the video file contains an audio stream. :param video_path: Path to the video file :return: True if the video contains an audio stream, False otherwise """ if osp.isdir(video_path): return False cmd = [ 'ffprobe', '-v', 'error', '-select_streams', 'a', '-show_entries', 'stream=codec_type', '-of', 'default=noprint_wrappers=1:nokey=1', video_path ] try: result = subprocess.run(cmd, capture_output=True, text=True) if result.returncode != 0: log(f"Error occurred while probing video: {result.stderr}") return False # Check if there is any output from ffprobe command return bool(result.stdout.strip()) except Exception as e: log(f"Error occurred while probing video: {video_path}, you may need to install ffprobe! Now set audio to false!", style="bold red") return False def add_audio_to_video(silent_video_path: str, audio_video_path: str, output_video_path: str): cmd = [ 'ffmpeg', '-y', '-i', silent_video_path, '-i', audio_video_path, '-map', '0:v', '-map', '1:a', '-c:v', 'copy', '-shortest', output_video_path ] try: exec_cmd(' '.join(cmd)) log(f"Video with audio generated successfully: {output_video_path}") except subprocess.CalledProcessError as e: log(f"Error occurred: {e}") def bb_intersection_over_union(boxA, boxB): xA = max(boxA[0], boxB[0]) yA = max(boxA[1], boxB[1]) xB = min(boxA[2], boxB[2]) yB = min(boxA[3], boxB[3]) interArea = max(0, xB - xA + 1) * max(0, yB - yA + 1) boxAArea = (boxA[2] - boxA[0] + 1) * (boxA[3] - boxA[1] + 1) boxBArea = (boxB[2] - boxB[0] + 1) * (boxB[3] - boxB[1] + 1) iou = interArea / float(boxAArea + boxBArea - interArea) return iou