StableVideo / stablevideo /atlas_data.py
wchai's picture
convert to CPU
d515943
raw
history blame
11.4 kB
import random
import json
import torch
import torch.nn.functional as F
from torch.utils.data import Dataset
from torchvision import transforms
from torchvision.transforms.functional import crop
from stablevideo.atlas_utils import (
load_neural_atlases_models,
get_frames_data,
get_high_res_atlas,
get_atlas_crops,
reconstruct_video_layer,
create_uv_mask,
get_masks_boundaries,
get_random_crop_params,
get_atlas_bounding_box,
load_video
)
class AtlasData():
def __init__(self, video_name):
with open(f"data/{video_name}/config.json", "r") as f:
json_dict = json.load(f)
try:
maximum_number_of_frames = json_dict["number_of_frames"]
except:
maximum_number_of_frames = json_dict["maximum_number_of_frames"]
config = {
"device": "cpu",
"checkpoint_path": f"data/{video_name}/checkpoint.ckpt",
"resx": json_dict["resx"],
"resy": json_dict["resy"],
"maximum_number_of_frames": maximum_number_of_frames,
"return_atlas_alpha": False,
"grid_atlas_resolution": 2000,
"num_scales": 7,
"masks_border_expansion": 30,
"mask_alpha_threshold": 0.99, # 0.95
"align_corners": False
}
self.config = config
self.device = config["device"]
self.min_size = min(self.config["resx"], self.config["resy"])
self.max_size = max(self.config["resx"], self.config["resy"])
data_folder = f"data/{video_name}/{video_name}"
self.original_video = load_video(
data_folder,
resize=(self.config["resy"], self.config["resx"]),
num_frames=self.config["maximum_number_of_frames"],
)
self.original_video = self.original_video.to(self.device) # tensor
(
foreground_mapping,
background_mapping,
foreground_atlas_model,
background_atlas_model,
alpha_model,
) = load_neural_atlases_models(config)
(
original_background_all_uvs,
original_foreground_all_uvs,
self.all_alpha,
foreground_atlas_alpha,
) = get_frames_data(
config,
foreground_mapping,
background_mapping,
alpha_model,
)
self.background_reconstruction = reconstruct_video_layer(original_background_all_uvs, background_atlas_model)
# using original video for the foreground layer
self.foreground_reconstruction = self.original_video * self.all_alpha
(
self.background_all_uvs,
self.scaled_background_uvs,
self.background_min_u,
self.background_min_v,
self.background_max_u,
self.background_max_v,
) = self.preprocess_uv_values(
original_background_all_uvs, config["grid_atlas_resolution"], device=self.device, layer="background"
)
(
self.foreground_all_uvs,
self.scaled_foreground_uvs,
self.foreground_min_u,
self.foreground_min_v,
self.foreground_max_u,
self.foreground_max_v,
) = self.preprocess_uv_values(
original_foreground_all_uvs, config["grid_atlas_resolution"], device=self.device, layer="foreground"
)
self.background_uv_mask = create_uv_mask(
config,
background_mapping,
self.background_min_u,
self.background_min_v,
self.background_max_u,
self.background_max_v,
uv_shift=-0.5,
resolution_shift=1,
)
self.foreground_uv_mask = create_uv_mask(
config,
foreground_mapping,
self.foreground_min_u,
self.foreground_min_v,
self.foreground_max_u,
self.foreground_max_v,
uv_shift=0.5,
resolution_shift=0,
)
self.background_grid_atlas = get_high_res_atlas(
background_atlas_model,
self.background_min_v,
self.background_min_u,
self.background_max_v,
self.background_max_u,
config["grid_atlas_resolution"],
device=config["device"],
layer="background",
)
self.foreground_grid_atlas = get_high_res_atlas(
foreground_atlas_model,
self.foreground_min_v,
self.foreground_min_u,
self.foreground_max_v,
self.foreground_max_u,
config["grid_atlas_resolution"],
device=config["device"],
layer="foreground",
)
if config["return_atlas_alpha"]:
self.foreground_atlas_alpha = foreground_atlas_alpha # used for visualizations
self.cnn_min_crop_size = 2 ** self.config["num_scales"] + 1
self.mask_boundaries = get_masks_boundaries(
alpha_video=self.all_alpha.cpu(),
border=self.config["masks_border_expansion"],
threshold=self.config["mask_alpha_threshold"],
min_crop_size=self.cnn_min_crop_size,
)
self.cropped_foreground_atlas, self.foreground_atlas_bbox = get_atlas_bounding_box(
self.mask_boundaries, self.foreground_grid_atlas, self.foreground_all_uvs
)
self.step = -1
self.edited_atlas_dict, self.edit_dict, self.uv_mask = {}, {}, {}
@staticmethod
def preprocess_uv_values(layer_uv_values, resolution, device="cuda", layer="background"):
if layer == "background":
shift = 1
else:
shift = 0
uv_values = (layer_uv_values + shift) * resolution
min_u, min_v = uv_values.reshape(-1, 2).min(dim=0).values.long()
uv_values -= torch.tensor([min_u, min_v], device=device)
max_u, max_v = uv_values.reshape(-1, 2).max(dim=0).values.ceil().long()
edge_size = torch.tensor([max_u, max_v], device=device)
scaled_uv_values = ((uv_values.reshape(-1, 2) / edge_size) * 2 - 1).unsqueeze(1).unsqueeze(0)
return uv_values, scaled_uv_values, min_u, min_v, max_u, max_v
def get_random_crop_data(self, crop_size):
t = random.randint(0, self.config["maximum_number_of_frames"] - 1)
y_start, x_start, h_crop, w_crop = get_random_crop_params((self.config["resx"], self.config["resy"]), crop_size)
return y_start, x_start, h_crop, w_crop, t
def get_global_crops_multi(self, keyframes, res):
foreground_atlas_crops = []
background_atlas_crops = []
foreground_uvs = []
background_uvs = []
background_alpha_crops = []
foreground_alpha_crops = []
original_background_crops = []
original_foreground_crops = []
output_dict = {}
self.config["crops_min_cover"] = 0.95
self.config["grid_atlas_resolution"] = res
for cur_frame in keyframes:
y_start, x_start, frame_h, frame_w = self.mask_boundaries[cur_frame].tolist()
crop_size = (
max(
random.randint(round(self.config["crops_min_cover"] * frame_h), frame_h),
self.cnn_min_crop_size,
),
max(
random.randint(round(self.config["crops_min_cover"] * frame_w), frame_w),
self.cnn_min_crop_size,
),
)
y_crop, x_crop, h_crop, w_crop = get_random_crop_params((frame_w, frame_h), crop_size)
foreground_uv = self.foreground_all_uvs[
cur_frame,
y_start + y_crop : y_start + y_crop + h_crop,
x_start + x_crop : x_start + x_crop + w_crop,
]
alpha = self.all_alpha[
[cur_frame],
:,
y_start + y_crop : y_start + y_crop + h_crop,
x_start + x_crop : x_start + x_crop + w_crop,
]
original_foreground_crop = self.foreground_reconstruction[
[cur_frame],
:,
y_start + y_crop : y_start + y_crop + h_crop,
x_start + x_crop : x_start + x_crop + w_crop,
]
foreground_alpha_crops.append(alpha)
foreground_uvs.append(foreground_uv) # not scaled
original_foreground_crops.append(original_foreground_crop)
foreground_max_vals = torch.tensor(
[self.config["grid_atlas_resolution"]] * 2, device=self.device, dtype=torch.long
)
foreground_min_vals = torch.tensor([0] * 2, device=self.device, dtype=torch.long)
for uv_values in foreground_uvs:
min_uv = uv_values.amin(dim=[0, 1]).long()
max_uv = uv_values.amax(dim=[0, 1]).ceil().long()
foreground_min_vals = torch.minimum(foreground_min_vals, min_uv)
foreground_max_vals = torch.maximum(foreground_max_vals, max_uv)
h_v = foreground_max_vals[1] - foreground_min_vals[1]
w_u = foreground_max_vals[0] - foreground_min_vals[0]
foreground_atlas_crop = crop(
self.foreground_grid_atlas,
foreground_min_vals[1],
foreground_min_vals[0],
h_v,
w_u,
)
foreground_atlas_crops.append(foreground_atlas_crop)
for i, uv_values in enumerate(foreground_uvs):
foreground_uvs[i] = (
2 * (uv_values - foreground_min_vals) / (foreground_max_vals - foreground_min_vals) - 1
).unsqueeze(0)
crop_size = (
random.randint(round(self.config["crops_min_cover"] * self.min_size), self.min_size),
random.randint(round(self.config["crops_min_cover"] * self.max_size), self.max_size),
)
crop_data = self.get_random_crop_data(crop_size)
y, x, h, w, _ = crop_data
background_uv = self.background_all_uvs[keyframes, y : y + h, x : x + w]
original_background_crop = self.background_reconstruction[
keyframes, :, y : y + h, x : x + w
]
alpha = self.all_alpha[keyframes, :, y : y + h, x : x + w]
original_background_crops = [el.unsqueeze(0) for el in original_background_crop]
background_alpha_crops = [el.unsqueeze(0) for el in alpha]
background_atlas_crop, background_min_vals, background_max_vals = get_atlas_crops(
background_uv,
self.background_grid_atlas,
)
background_uv = 2 * (background_uv - background_min_vals) / (background_max_vals - background_min_vals) - 1
background_atlas_crops = [el.unsqueeze(0) for el in background_atlas_crop]
background_uvs = [el.unsqueeze(0) for el in background_uv]
output_dict["foreground_alpha"] = foreground_alpha_crops
output_dict["foreground_uvs"] = foreground_uvs
output_dict["original_foreground_crops"] = original_foreground_crops
output_dict["foreground_atlas_crops"] = foreground_atlas_crops
output_dict["background_alpha"] = background_alpha_crops
output_dict["background_uvs"] = background_uvs
output_dict["original_background_crops"] = original_background_crops
output_dict["background_atlas_crops"] = background_atlas_crops
return output_dict