# coding: utf-8 """ functions for processing video """ import os.path as osp import numpy as np import subprocess import imageio import cv2 from rich.progress import track from .helper import prefix from .rprint import rprint as print def exec_cmd(cmd): subprocess.run(cmd, shell=True, check=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) def images2video(images, wfp, **kwargs): fps = kwargs.get('fps', 25) # MuseTalk prefers 25 fps by default video_format = kwargs.get('format', 'mp4') # default is mp4 format codec = kwargs.get('codec', 'libx264') # default is libx264 encoding quality = kwargs.get('quality') # video quality pixelformat = kwargs.get('pixelformat', 'yuv420p') # video pixel format image_mode = kwargs.get('image_mode', 'rgb') macro_block_size = kwargs.get('macro_block_size', 2) ffmpeg_params = ['-crf', str(kwargs.get('crf', 18))] writer = imageio.get_writer( wfp, fps=fps, format=video_format, codec=codec, quality=quality, ffmpeg_params=ffmpeg_params, pixelformat=pixelformat, macro_block_size=macro_block_size ) n = len(images) for i in track(range(n), description='writing', transient=True): if image_mode.lower() == 'bgr': writer.append_data(images[i][..., ::-1]) else: writer.append_data(images[i]) writer.close() # print(f':smiley: Dump to {wfp}\n', style="bold green") print(f'Dump to {wfp}\n') # useTalk prefers 25 fps by default def video2gif(video_fp, fps=25, size=256): if osp.exists(video_fp): d = osp.split(video_fp)[0] fn = prefix(osp.basename(video_fp)) palette_wfp = osp.join(d, 'palette.png') gif_wfp = osp.join(d, f'{fn}.gif') # generate the palette cmd = f'ffmpeg -i {video_fp} -vf "fps={fps},scale={size}:-1:flags=lanczos,palettegen" {palette_wfp} -y' exec_cmd(cmd) # use the palette to generate the gif cmd = f'ffmpeg -i {video_fp} -i {palette_wfp} -filter_complex "fps={fps},scale={size}:-1:flags=lanczos[x];[x][1:v]paletteuse" {gif_wfp} -y' exec_cmd(cmd) else: print(f'video_fp: {video_fp} not exists!') def merge_audio_video(video_fp, audio_fp, wfp): if osp.exists(video_fp) and osp.exists(audio_fp): cmd = f'ffmpeg -i {video_fp} -i {audio_fp} -c:v copy -c:a aac {wfp} -y' exec_cmd(cmd) print(f'merge {video_fp} and {audio_fp} to {wfp}') else: print(f'video_fp: {video_fp} or audio_fp: {audio_fp} not exists!') def blend(img: np.ndarray, mask: np.ndarray, background_color=(255, 255, 255)): mask_float = mask.astype(np.float32) / 255. background_color = np.array(background_color).reshape([1, 1, 3]) bg = np.ones_like(img) * background_color img = np.clip(mask_float * img + (1 - mask_float) * bg, 0, 255).astype(np.uint8) return img def concat_frames(I_p_lst, driving_rgb_lst, img_rgb): # TODO: add more concat style, e.g., left-down corner driving out_lst = [] for idx, _ in track(enumerate(I_p_lst), total=len(I_p_lst), description='Concatenating result...'): source_image_drived = I_p_lst[idx] image_drive = driving_rgb_lst[idx] # resize images to match source_image_drived shape h, w, _ = source_image_drived.shape image_drive_resized = cv2.resize(image_drive, (w, h)) img_rgb_resized = cv2.resize(img_rgb, (w, h)) # concatenate images horizontally frame = np.concatenate((image_drive_resized, img_rgb_resized, source_image_drived), axis=1) out_lst.append(frame) return out_lst class VideoWriter: def __init__(self, **kwargs): self.fps = kwargs.get('fps', 25) # MuseTalk prefers 25 fps by default self.wfp = kwargs.get('wfp', 'video.mp4') self.video_format = kwargs.get('format', 'mp4') self.codec = kwargs.get('codec', 'libx264') self.quality = kwargs.get('quality') self.pixelformat = kwargs.get('pixelformat', 'yuv420p') self.image_mode = kwargs.get('image_mode', 'rgb') self.ffmpeg_params = kwargs.get('ffmpeg_params') self.writer = imageio.get_writer( self.wfp, fps=self.fps, format=self.video_format, codec=self.codec, quality=self.quality, ffmpeg_params=self.ffmpeg_params, pixelformat=self.pixelformat ) def write(self, image): if self.image_mode.lower() == 'bgr': self.writer.append_data(image[..., ::-1]) else: self.writer.append_data(image) def close(self): if self.writer is not None: self.writer.close() def change_video_fps(input_file, output_file, fps=25, codec='libx264', crf=5): cmd = f"ffmpeg -i {input_file} -c:v {codec} -crf {crf} -r {fps} {output_file} -y" exec_cmd(cmd) def get_fps(filepath): import ffmpeg probe = ffmpeg.probe(filepath) video_stream = next((stream for stream in probe['streams'] if stream['codec_type'] == 'video'), None) fps = eval(video_stream['avg_frame_rate']) return fps