|
import logging |
|
import random |
|
import torch |
|
import webvtt |
|
import os |
|
import cv2 |
|
from torchvision import transforms |
|
import soundfile as sf |
|
import moviepy.editor as mp |
|
from PIL import Image |
|
from moviepy.editor import VideoFileClip |
|
import torch |
|
import random |
|
import torch.backends.cudnn as cudnn |
|
import torch |
|
from torch.cuda.amp import autocast as autocast |
|
import torch.nn as nn |
|
|
|
from .registry import registry |
|
from .blip2 import Blip2Base, disabled_train |
|
from .conversation import Conversation, SeparatorStyle, StoppingCriteriaList, StoppingCriteriaSub |
|
from transformers import LlamaTokenizer |
|
from transformers import BitsAndBytesConfig |
|
from transformers import AutoConfig, AutoTokenizer |
|
from peft import ( |
|
LoraConfig, |
|
get_peft_model, |
|
get_peft_model_state_dict, |
|
prepare_model_for_int8_training, |
|
set_peft_model_state_dict, |
|
) |
|
import time |
|
import numpy as np |
|
import os |
|
from transformers import PretrainedConfig |
|
from transformers import PreTrainedModel |
|
from .conversation import CONV_VISION |
|
import cv2 |
|
def extract_audio(video_path, audio_path): |
|
video_clip = mp.VideoFileClip(video_path) |
|
audio_clip = video_clip.audio |
|
audio_clip.write_audiofile(audio_path, codec="libmp3lame", bitrate="320k") |
|
|
|
def generate_subtitles(video_path): |
|
video_id=video_path.split('/')[-1].split('.')[0] |
|
audio_path = f"workspace/inference_subtitles/mp3/{video_id}"+'.mp3' |
|
os.makedirs("workspace/inference_subtitles/mp3",exist_ok=True) |
|
try: |
|
extract_audio(video_path,audio_path) |
|
print("successfully extracted") |
|
os.system(f"whisper {audio_path} --language English --model large --output_format vtt --output_dir workspace/inference_subtitles/") |
|
|
|
os.system(f"rm {audio_path}") |
|
print("subtitle successfully generated") |
|
return f"workspace/inference_subtitles/{video_id}"+'.vtt' |
|
except Exception as e: |
|
print("error",e) |
|
print("error",video_path) |
|
return None |
|
|
|
class minigpt4_video_config(PretrainedConfig): |
|
model_type="minigpt4_video" |
|
PRETRAINED_MODEL_CONFIG_DICT = { |
|
"minigpt4_video": "configs/models/minigpt4.yaml", |
|
} |
|
def __init__( |
|
self, |
|
omg_config:dict = {}, |
|
**kwargs, |
|
): |
|
for key, value in omg_config.items(): |
|
setattr(self, key, value) |
|
super().__init__(**kwargs) |
|
|
|
|
|
@registry.register_model("mini_gpt4_llama_v2") |
|
class MiniGPT4_Video(Blip2Base, PreTrainedModel): |
|
""" |
|
BLIP2 GPT-LLAMA model. |
|
""" |
|
PRETRAINED_MODEL_CONFIG_DICT = { |
|
"minigpt4_video": "minigpt4/configs/models/minigpt4.yaml", |
|
} |
|
config_class=minigpt4_video_config |
|
|
|
def __init__( |
|
self, |
|
cfg={}, |
|
): |
|
|
|
|
|
try: |
|
cfg = cfg.to_dict() |
|
except: |
|
pass |
|
for key, value in cfg.items(): |
|
try: |
|
setattr(self, key, value) |
|
except: |
|
print(f"Error setting attribute {key} with value {value}") |
|
PreTrainedModel.__init__(self, minigpt4_video_config(cfg)) |
|
Blip2Base.__init__(self) |
|
|
|
vis_processor_cfg = {"name": "blip2_image_train","image_size": 224} |
|
self.vis_processor = registry.get_processor_class(vis_processor_cfg["name"]) |
|
self.vis_processor = self.vis_processor.from_config(vis_processor_cfg) |
|
self.CONV_VISION = CONV_VISION |
|
if "Mistral" in self.llama_model: |
|
from .modeling_mistral import MistralForCausalLM as llm_model |
|
print("Mistral model") |
|
self.model_type = "Mistral" |
|
else: |
|
from .modeling_llama_v2 import LlamaForCausalLM as llm_model |
|
print("Llama model") |
|
self.model_type = "Llama" |
|
self.tokenizer = self.init_tokenizer() |
|
|
|
print("token pooling", self.token_pooling) |
|
if self.freeze_vit: |
|
|
|
print("vit precision", self.vit_precision) |
|
self.visual_encoder, self.ln_vision = self.init_vision_encoder( |
|
self.vit_model, self.img_size, self.drop_path_rate, self.use_grad_checkpoint, self.vit_precision |
|
) |
|
for name, param in self.visual_encoder.named_parameters(): |
|
param.requires_grad = False |
|
self.visual_encoder = self.visual_encoder.eval() |
|
self.visual_encoder.train = disabled_train |
|
for name, param in self.ln_vision.named_parameters(): |
|
param.requires_grad = False |
|
self.ln_vision = self.ln_vision.eval() |
|
self.ln_vision.train = disabled_train |
|
logging.info("freeze vision encoder") |
|
print("freeze the vision encoder") |
|
|
|
else: |
|
self.vit_precision="fp32" |
|
self.visual_encoder, self.ln_vision = self.init_vision_encoder( |
|
self.vit_model, self.img_size, self.drop_path_rate, self.use_grad_checkpoint, self.vit_precision |
|
) |
|
|
|
print("unfreeze the vision encoder") |
|
print('Loading VIT Done') |
|
|
|
print('Loading LLAMA') |
|
|
|
self.B_SYS, self.E_SYS = "<<SYS>>\n", "\n<</SYS>>\n\n" |
|
token=os.environ.get("HF_TKN") |
|
self.llama_tokenizer = LlamaTokenizer.from_pretrained(self.llama_model,use_fast=False,token=token) |
|
self.llama_tokenizer.pad_token = "$$" |
|
|
|
self.use_fastv = False |
|
print("self.low_resource",self.low_resource) |
|
if self.low_resource: |
|
self.llama_model = llm_model.from_pretrained( |
|
self.llama_model, |
|
torch_dtype=torch.float16, |
|
|
|
load_in_8bit=True, |
|
|
|
|
|
device_map={'':torch.cuda.current_device()},token=token |
|
|
|
|
|
) |
|
else: |
|
self.llama_model = llm_model.from_pretrained( |
|
self.llama_model, |
|
torch_dtype=torch.float16,token=token |
|
) |
|
|
|
|
|
self.llama_model = prepare_model_for_int8_training(self.llama_model) |
|
loraconfig = LoraConfig( |
|
r=self.lora_r, |
|
lora_alpha=self.lora_alpha, |
|
target_modules=self.lora_target_modules, |
|
lora_dropout=self.lora_dropout, |
|
bias="none", |
|
task_type="CAUSAL_LM" |
|
) |
|
self.llama_model = get_peft_model(self.llama_model, loraconfig) |
|
|
|
self.llama_model.print_trainable_parameters() |
|
|
|
if self.use_grad_checkpoint_llm: |
|
self.llama_model.gradient_checkpointing_enable() |
|
|
|
print('Loading LLAMA Done') |
|
|
|
|
|
if self.token_pooling: |
|
self.llama_proj = nn.Linear( |
|
1408*4, self.llama_model.config.hidden_size |
|
) |
|
else: |
|
self.llama_proj = nn.Linear( |
|
1408, self.llama_model.config.hidden_size |
|
) |
|
if self.prompt_path: |
|
with open(self.prompt_path, 'r') as f: |
|
raw_prompts = f.read().splitlines() |
|
filted_prompts = [raw_prompt for raw_prompt in raw_prompts if "<ImageHere>" in raw_prompt] |
|
self.prompt_list = [self.prompt_template.format(p) for p in filted_prompts] |
|
print('Load {} training prompts'.format(len(self.prompt_list))) |
|
print('Prompt Example \n{}'.format(random.choice(self.prompt_list))) |
|
else: |
|
self.prompt_list = [] |
|
def prepare_input(self,video_path,subtitle_path,instruction): |
|
cap = cv2.VideoCapture(video_path) |
|
if subtitle_path is not None: |
|
|
|
vtt_file = webvtt.read(subtitle_path) |
|
print("subtitle loaded successfully") |
|
clip = VideoFileClip(video_path) |
|
total_num_frames = int(clip.duration * clip.fps) |
|
|
|
clip.close() |
|
else : |
|
|
|
total_num_frames = int(cap.get(cv2.CAP_PROP_FRAME_COUNT)) |
|
if self.model_type == "Mistral": |
|
max_images_length = 90 |
|
max_sub_len = 800 |
|
else: |
|
max_images_length = 45 |
|
max_sub_len = 400 |
|
images = [] |
|
frame_count = 0 |
|
sampling_interval = int(total_num_frames / max_images_length) |
|
if sampling_interval == 0: |
|
sampling_interval = 1 |
|
img_placeholder = "" |
|
subtitle_text_in_interval = "" |
|
history_subtitles = {} |
|
raw_frames=[] |
|
number_of_words=0 |
|
transform=transforms.Compose([ |
|
transforms.ToPILImage(), |
|
]) |
|
while cap.isOpened(): |
|
ret, frame = cap.read() |
|
if not ret: |
|
break |
|
|
|
|
|
if subtitle_path is not None: |
|
for subtitle in vtt_file: |
|
sub=subtitle.text.replace('\n',' ') |
|
if (subtitle.start_in_seconds <= (frame_count / int(clip.fps)) <= subtitle.end_in_seconds) and sub not in subtitle_text_in_interval: |
|
if not history_subtitles.get(sub,False): |
|
subtitle_text_in_interval+=sub+" " |
|
history_subtitles[sub]=True |
|
break |
|
if frame_count % sampling_interval == 0: |
|
raw_frames.append(Image.fromarray(cv2.cvtColor(frame.copy(), cv2.COLOR_BGR2RGB))) |
|
frame = transform(frame[:,:,::-1]) |
|
frame = self.vis_processor(frame) |
|
images.append(frame) |
|
img_placeholder += '<Img><ImageHere>' |
|
if subtitle_path is not None and subtitle_text_in_interval != "" and number_of_words< max_sub_len: |
|
img_placeholder+=f'<Cap>{subtitle_text_in_interval}' |
|
number_of_words+=len(subtitle_text_in_interval.split(' ')) |
|
subtitle_text_in_interval = "" |
|
frame_count += 1 |
|
|
|
if len(images) >= max_images_length: |
|
break |
|
|
|
while len(images) < max_images_length: |
|
images.append(images[-1]) |
|
img_placeholder += '<Img><ImageHere>' |
|
|
|
cap.release() |
|
cv2.destroyAllWindows() |
|
if len(images) == 0: |
|
|
|
return None,None |
|
images = torch.stack(images) |
|
instruction = img_placeholder + '\n' + instruction |
|
return images,instruction |
|
def encode_img(self, image): |
|
device = image.device |
|
if len(image.shape) > 4: |
|
image = image.reshape(-1, *image.shape[-3:]) |
|
with self.maybe_autocast(): |
|
image_embeds = self.ln_vision(self.visual_encoder(image)).to(device) |
|
image_embeds = image_embeds[:,1:,:] |
|
bs, pn, hs = image_embeds.shape |
|
if self.token_pooling: |
|
image_embeds = image_embeds.view(bs, int(pn/4), int(hs*4)) |
|
|
|
inputs_llama = self.llama_proj(image_embeds) |
|
atts_llama = torch.ones(inputs_llama.size()[:-1], dtype=torch.long).to(image.device) |
|
return inputs_llama, atts_llama |
|
|
|
def get_context_emb(self, prompt, img_list): |
|
img_device = img_list[0].device |
|
prompt_segs = prompt.split('<ImageHere>') |
|
assert len(prompt_segs) == len(img_list) + 1, "Unmatched numbers of image placeholders and images." |
|
seg_tokens = [ |
|
self.llama_tokenizer( |
|
seg, return_tensors="pt", add_special_tokens=i==0).to(img_device).input_ids |
|
for i, seg in enumerate(prompt_segs) |
|
] |
|
|
|
seg_embs = [self.embed_tokens(seg_t) for seg_t in seg_tokens] |
|
|
|
mixed_embs = [emb for pair in zip(seg_embs[:-1], img_list) for emb in pair] + [seg_embs[-1]] |
|
|
|
mixed_embs = torch.cat(mixed_embs, dim=1) |
|
|
|
return mixed_embs |
|
|
|
def prompt_wrap(self, img_embeds, atts_img, prompts, lengths=None): |
|
if prompts is None or len(prompts) == 0: |
|
|
|
return img_embeds, atts_img |
|
elif img_embeds is None: |
|
|
|
self.llama_tokenizer.padding_side = "right" |
|
prompt_tokens = self.llama_tokenizer( |
|
prompts, |
|
return_tensors="pt", |
|
padding="max_length", |
|
add_special_tokens=False |
|
).to(self.device) |
|
prompt_embeds = self.embed_tokens(prompt_tokens.input_ids) |
|
atts_prompt = prompt_tokens.attention_mask |
|
return prompt_embeds, atts_prompt |
|
|
|
else: |
|
|
|
emb_lists = [] |
|
if type(prompts) == str: |
|
prompts = [prompts] * len(img_embeds) |
|
for idx, (each_img_embed, each_prompt) in enumerate(zip(img_embeds, prompts)): |
|
pn = each_img_embed.shape[-2] |
|
if lengths is not None: |
|
each_img_embed = each_img_embed.reshape(-1, each_img_embed.shape[-1]) |
|
each_img_embed = each_img_embed[:lengths[idx] * pn] |
|
|
|
p_segs = each_prompt.split('<ImageHere>') |
|
interleave_emb = [] |
|
for idx, seg in enumerate(p_segs[:-1]): |
|
p_tokens = self.llama_tokenizer(seg, return_tensors="pt", add_special_tokens=False).to(img_embeds.device) |
|
p_embed = self.embed_tokens(p_tokens.input_ids) |
|
|
|
interleave_emb.append(torch.cat([p_embed, each_img_embed[None][:, idx*pn:(idx+1)*pn]], dim=1)) |
|
|
|
wrapped_emb = torch.cat(interleave_emb, dim=1) |
|
p_tokens = self.llama_tokenizer(p_segs[-1], return_tensors="pt", add_special_tokens=False).to(img_embeds.device) |
|
p_embed = self.embed_tokens(p_tokens.input_ids) |
|
wrapped_emb = torch.cat([wrapped_emb,p_embed], dim=1) |
|
emb_lists.append(wrapped_emb) |
|
|
|
emb_lens = [emb.shape[1] for emb in emb_lists] |
|
pad_emb = self.embed_tokens(torch.tensor(self.llama_tokenizer.pad_token_id, device=img_embeds.device)) |
|
|
|
|
|
max_length = self.max_context_len |
|
wrapped_embs = pad_emb.expand(len(emb_lens), max_length, -1).clone() |
|
wrapped_atts = torch.zeros([len(emb_lens), max_length], dtype=torch.int, device=img_embeds.device) |
|
|
|
for i, emb in enumerate(emb_lists): |
|
length = emb_lens[i] if emb_lens[i] < self.max_context_len else self.max_context_len |
|
wrapped_embs[i, :length] = emb[:, :length] |
|
wrapped_atts[i, :length] = 1 |
|
|
|
return wrapped_embs, wrapped_atts |
|
|
|
def concat_emb_input_output(self, input_embs, input_atts, output_embs, output_atts): |
|
""" |
|
Concatenate the batched input embedding and batched output embedding together. |
|
Both the input and the output embedding should be right padded. |
|
""" |
|
|
|
input_lens = [] |
|
cat_embs = [] |
|
cat_atts = [] |
|
|
|
for i in range(input_embs.size(0)): |
|
input_len = input_atts[i].sum() |
|
input_lens.append(input_len) |
|
|
|
cat_embs.append( |
|
torch.cat([ |
|
input_embs[i][:input_len], |
|
output_embs[i], |
|
input_embs[i][input_len:] |
|
]) |
|
) |
|
cat_atts.append( |
|
torch.cat([ |
|
input_atts[i][:input_len], |
|
output_atts[i], |
|
input_atts[i][input_len:] |
|
]) |
|
) |
|
|
|
cat_embs = torch.stack(cat_embs) |
|
cat_atts = torch.stack(cat_atts) |
|
return cat_embs, cat_atts, input_lens |
|
|
|
def get_conv_emb(self, conv_q, conv_a, conv_img): |
|
"""concatenate conversation and make sure the model is only trained to regress the answer""" |
|
|
|
regress_embs_list = [] |
|
targets_list = [] |
|
|
|
batch_size = len(conv_q) |
|
for batch_idx in range(batch_size): |
|
questions, answers = conv_q[batch_idx], conv_a[batch_idx] |
|
assigned_imgs = conv_img[batch_idx] |
|
questions = [self.prompt_wrap( |
|
img_embeds=img, |
|
atts_img=None, |
|
prompts=[q], |
|
lengths=[img.shape[1]] if img is not None else None) for q, img in zip(questions, assigned_imgs)] |
|
q_embs = [emb for emb, _ in questions] |
|
|
|
answers = [self.llama_tokenizer(a, return_tensors="pt", add_special_tokens=False).to(self.device) for a in answers] |
|
cur_emb = [] |
|
cur_target = [] |
|
for i in range(len(questions)): |
|
cur_emb.append(q_embs[i]) |
|
cur_target.append(torch.ones_like(q_embs[i][..., 0], dtype=torch.int) * -100) |
|
|
|
cur_emb.append(self.embed_tokens(answers[i].input_ids)) |
|
cur_target.append(answers[i].input_ids) |
|
|
|
cur_emb = torch.cat(cur_emb, dim=1) |
|
cur_target = torch.cat(cur_target, dim=1) |
|
|
|
regress_embs_list.append(cur_emb) |
|
targets_list.append(cur_target) |
|
|
|
max_len = min(max([target.shape[1] for target in targets_list]), self.max_txt_len) |
|
|
|
regress_embeds = torch.zeros([batch_size, max_len, cur_emb.shape[-1]], device=self.device) |
|
regress_attn = torch.zeros([batch_size, max_len], dtype=torch.int, device=self.device) |
|
targets = torch.ones([batch_size, max_len], dtype=torch.long, device=self.device) * -100 |
|
|
|
for batch_idx in range(batch_size): |
|
cur_len = regress_embs_list[batch_idx].shape[1] |
|
regress_embeds[batch_idx, :cur_len] = regress_embs_list[batch_idx][0, :max_len] |
|
regress_attn[batch_idx, :cur_len] = 1 |
|
targets[batch_idx, :cur_len] = targets_list[batch_idx][0, :max_len] |
|
|
|
return regress_embeds, regress_attn, targets |
|
|
|
def preparing_embedding(self, samples): |
|
def remove_special_tokens(data): |
|
|
|
|
|
data = [instruct.replace(" [caption]","") for instruct in data] |
|
data = [instruct.replace(" [vqa]","") for instruct in data] |
|
data = [instruct.replace(" [grounding]","") for instruct in data] |
|
data = [instruct.replace(" [identify]","") for instruct in data] |
|
data = [instruct.replace(" [refer]","") for instruct in data] |
|
return data |
|
|
|
|
|
if 'image' in samples: |
|
img_embeds, img_atts = self.encode_img(samples["image"]) |
|
else: |
|
img_embeds = img_atts = None |
|
|
|
if 'conv_q' in samples: |
|
|
|
conv_q, conv_a = samples['conv_q'], samples['conv_a'] |
|
|
|
connect_sym = samples['connect_sym'][0] |
|
conv_q = [q.split(connect_sym)for q in conv_q] |
|
conv_a = [a.split(connect_sym) for a in conv_a] |
|
conv_img = assign_imgs(conv_q, img_embeds) |
|
|
|
if self.chat_template: |
|
conv_q = [["[INST] " + item + "[/INST]" for item in items] for items in conv_q] |
|
|
|
regress_embeds, regress_atts, part_targets = self.get_conv_emb(conv_q, conv_a, conv_img) |
|
cond_embeds, cond_atts = regress_embeds[:, :0], regress_atts[:, :0] |
|
|
|
else: |
|
if "instruction_input" in samples: |
|
instruction = samples["instruction_input"] |
|
elif len(self.prompt_list) > 1: |
|
instruction = random.choice(self.prompt_list) |
|
else: |
|
instruction = None |
|
|
|
if self.remove_template: |
|
instruction = remove_special_tokens(instruction) |
|
|
|
if self.chat_template: |
|
instruction = ["[INST] " + instruct + "[/INST]" for instruct in instruction] |
|
|
|
if 'length' in samples: |
|
|
|
bsz, pn, hs = img_embeds.shape |
|
img_embeds = img_embeds.reshape(len(samples['image']), -1, pn, hs) |
|
cond_embeds, cond_atts = self.prompt_wrap(img_embeds, img_atts, instruction, samples['length']) |
|
else: |
|
cond_embeds, cond_atts = self.prompt_wrap(img_embeds, img_atts, instruction) |
|
|
|
|
|
self.llama_tokenizer.padding_side = "right" |
|
text = [t + self.end_sym for t in samples["answer"]] |
|
|
|
regress_tokens = self.llama_tokenizer( |
|
text, |
|
return_tensors="pt", |
|
padding="max_length", |
|
truncation=True, |
|
max_length=self.max_txt_len, |
|
add_special_tokens=False |
|
).to(self.device) |
|
|
|
regress_token_ids = regress_tokens.input_ids |
|
regress_atts = regress_tokens.attention_mask |
|
part_targets = regress_token_ids.masked_fill( |
|
regress_token_ids == self.llama_tokenizer.pad_token_id, -100 |
|
) |
|
|
|
regress_embeds = self.embed_tokens(regress_token_ids) |
|
|
|
return cond_embeds, cond_atts, regress_embeds, regress_atts, part_targets |
|
|
|
def forward(self, samples, reduction="mean"): |
|
|
|
cond_embeds, cond_atts, regress_embeds, regress_atts, part_targets = \ |
|
self.preparing_embedding(samples) |
|
|
|
|
|
inputs_embeds, attention_mask, input_lens = \ |
|
self.concat_emb_input_output(cond_embeds, cond_atts, regress_embeds, regress_atts) |
|
|
|
bos = torch.ones_like(part_targets[:, :1]) * self.llama_tokenizer.bos_token_id |
|
bos_embeds = self.embed_tokens(bos) |
|
bos_atts = attention_mask[:, :1] |
|
|
|
|
|
inputs_embeds = torch.cat([bos_embeds, inputs_embeds], dim=1) |
|
attention_mask = torch.cat([bos_atts, attention_mask], dim=1) |
|
|
|
targets = torch.ones([inputs_embeds.shape[0], inputs_embeds.shape[1]], |
|
dtype=torch.long).to(self.device).fill_(-100) |
|
for i, target in enumerate(part_targets): |
|
targets[i, input_lens[i]+1:input_lens[i]+len(target)+1] = target |
|
|
|
with self.maybe_autocast(): |
|
outputs = self.llama_model( |
|
inputs_embeds=inputs_embeds, |
|
attention_mask=attention_mask, |
|
return_dict=True, |
|
labels=targets, |
|
reduction=reduction, |
|
use_fastv=self.use_fastv |
|
) |
|
loss = outputs.loss |
|
|
|
return {"loss": loss} |
|
|
|
@torch.no_grad() |
|
def generate( |
|
self, |
|
images, |
|
texts, |
|
use_nucleus_sampling=False, |
|
num_beams=1, |
|
max_new_tokens=20, |
|
min_length=1, |
|
top_p=0.9, |
|
repetition_penalty=1.5, |
|
length_penalty=1, |
|
temperature=1, |
|
do_sample=False, |
|
stop_words_ids=[2], |
|
lengths=None, |
|
return_video_temporal_features=False, |
|
img_embeds=None, |
|
): |
|
''' |
|
function for generate test use |
|
''' |
|
|
|
stopping_criteria = StoppingCriteriaList([StoppingCriteriaSub( |
|
stops=[torch.tensor([i]).to(self.device) for i in stop_words_ids])]) |
|
if img_embeds is None: |
|
img_embeds, atts_img = self.encode_img(images.to(self.device)) |
|
else: |
|
|
|
img_embeds = img_embeds.reshape(-1, *img_embeds.shape[-2:]) |
|
img_embeds= img_embeds.to(self.device) |
|
img_embeds = self.llama_proj(img_embeds) |
|
atts_img = torch.ones(img_embeds.size()[:-1], dtype=torch.long).to(self.device) |
|
|
|
if lengths is not None: |
|
image_lists = [] |
|
img_embeds = img_embeds.reshape(len(lengths), -1, img_embeds.shape[-2], img_embeds.shape[-1]) |
|
for idx, img_embed in enumerate(img_embeds): |
|
image_lists.append([img_embed[i][None] for i in range(lengths[idx])]) |
|
else: |
|
image_lists = [[image_emb[None]] for image_emb in img_embeds] |
|
assert len(texts) == len(image_lists) |
|
batch_embs = [self.get_context_emb(text, img_list) for text, img_list in zip(texts, image_lists)] |
|
|
|
batch_size = len(batch_embs) |
|
max_len = max([emb.shape[1] for emb in batch_embs]) |
|
emb_dim = batch_embs[0].shape[2] |
|
dtype = batch_embs[0].dtype |
|
device = batch_embs[0].device |
|
|
|
embs = torch.zeros([batch_size, max_len, emb_dim], dtype=dtype, device=device) |
|
attn_mask = torch.zeros([batch_size, max_len], dtype=torch.int, device=device) |
|
for i, emb in enumerate(batch_embs): |
|
emb_len = emb.shape[1] |
|
embs[i, -emb_len:] = emb[0] |
|
attn_mask[i, -emb_len:] = 1 |
|
|
|
if self.model_type == "Llama": |
|
context_window = 3700 |
|
else: |
|
context_window = 7500 |
|
if embs.shape[1] > context_window: |
|
embs = embs[:, -context_window:] |
|
attn_mask = attn_mask[:, -context_window:] |
|
with self.maybe_autocast(): |
|
if return_video_temporal_features: |
|
last_hidden_state = self.llama_model( |
|
inputs_embeds=embs, |
|
attention_mask=attn_mask, |
|
output_hidden_states=True, |
|
).hidden_states[-1] |
|
video_temporal_features = last_hidden_state.mean(dim=1) |
|
|
|
|
|
outputs = self.llama_model.generate( |
|
inputs_embeds=embs, |
|
attention_mask=attn_mask, |
|
max_new_tokens=max_new_tokens, |
|
num_beams=num_beams, |
|
do_sample=do_sample, |
|
temperature=temperature, |
|
repetition_penalty=repetition_penalty, |
|
|
|
use_fastv=False, |
|
) |
|
|
|
answers = [] |
|
for output_token in outputs: |
|
if output_token[0] == 0: |
|
output_token = output_token[1:] |
|
output_texts = self.llama_tokenizer.decode(output_token, skip_special_tokens=True) |
|
output_texts = output_texts.split('</s>')[0] |
|
output_texts = output_texts.replace("<s>", "") |
|
output_texts = output_texts.split(r'[/INST]')[-1].strip() |
|
answers.append(output_texts) |
|
if return_video_temporal_features: |
|
return answers, video_temporal_features |
|
else: |
|
return answers |
|
def inference_fun (self,video_path,instruction,gen_subtitles=True): |
|
if gen_subtitles: |
|
subtitle_path=generate_subtitles(video_path) |
|
else : |
|
subtitle_path=None |
|
prepared_images,prepared_instruction=self.prepare_input(video_path,subtitle_path,instruction) |
|
if prepared_images is None: |
|
return "Video cann't be open ,check the video path again" |
|
length=len(prepared_images) |
|
prepared_images=prepared_images.unsqueeze(0) |
|
conv = self.CONV_VISION.copy() |
|
conv.system = "" |
|
|
|
conv.append_message(conv.roles[0], prepared_instruction) |
|
conv.append_message(conv.roles[1], None) |
|
prompt = [conv.get_prompt()] |
|
answers = self.generate(prepared_images, prompt, max_new_tokens=512, do_sample=True, lengths=[length],num_beams=1) |
|
return answers[0] |
|
@torch.no_grad() |
|
def generate_text_only( |
|
self, |
|
images, |
|
seg_tokens, |
|
use_nucleus_sampling=False, |
|
num_beams=1, |
|
max_new_tokens=20, |
|
min_length=1, |
|
top_p=0.9, |
|
repetition_penalty=1.5, |
|
length_penalty=1, |
|
temperature=1, |
|
do_sample=False, |
|
stop_words_ids=[2], |
|
lengths=None, |
|
return_video_temporal_features=False, |
|
img_embeds=None, |
|
): |
|
''' |
|
function for generate test use |
|
''' |
|
|
|
stopping_criteria = StoppingCriteriaList([StoppingCriteriaSub( |
|
stops=[torch.tensor([i]).to(self.device) for i in stop_words_ids])]) |
|
|
|
batch_embs = [torch.cat([self.embed_tokens(seg_t)]) for seg_t in seg_tokens] |
|
|
|
batch_size = len(batch_embs) |
|
max_len = max([emb.shape[1] for emb in batch_embs]) |
|
emb_dim = batch_embs[0].shape[2] |
|
dtype = batch_embs[0].dtype |
|
device = batch_embs[0].device |
|
|
|
embs = torch.zeros([batch_size, max_len, emb_dim], dtype=dtype, device=device) |
|
attn_mask = torch.zeros([batch_size, max_len], dtype=torch.int, device=device) |
|
for i, emb in enumerate(batch_embs): |
|
emb_len = emb.shape[1] |
|
embs[i, -emb_len:] = emb[0] |
|
attn_mask[i, -emb_len:] = 1 |
|
|
|
with self.maybe_autocast(): |
|
outputs = self.llama_model.generate( |
|
inputs_embeds=embs, |
|
attention_mask=attn_mask, |
|
max_new_tokens=max_new_tokens, |
|
num_beams=num_beams, |
|
do_sample=do_sample, |
|
temperature=temperature, |
|
repetition_penalty=repetition_penalty, |
|
|
|
) |
|
|
|
answers = [] |
|
for output_token in outputs: |
|
if output_token[0] == 0: |
|
output_token = output_token[1:] |
|
output_texts = self.llama_tokenizer.decode(output_token, skip_special_tokens=True) |
|
output_texts = output_texts.split('</s>')[0] |
|
output_texts = output_texts.replace("<s>", "") |
|
output_texts = output_texts.split(r'[/INST]')[-1].strip() |
|
answers.append(output_texts) |
|
return answers |
|
|
|
|
|
|
|
@torch.no_grad() |
|
def multi_select(self, images, texts, answers, num_cand=None): |
|
all_losses = [] |
|
for answer in answers: |
|
choice_samples = { |
|
'image': images, |
|
'instruction_input': texts, |
|
'answer': answer |
|
} |
|
loss = self.forward(choice_samples, reduction='none')['loss'].reshape(-1, 1) |
|
all_losses.append(loss) |
|
torch.cuda.empty_cache() |
|
all_losses = torch.cat(all_losses, dim=-1) |
|
if num_cand is not None: |
|
for i in range(all_losses.shape[0]): |
|
all_losses[i, num_cand[i]:] = 9999 |
|
output_class_ranks = torch.argsort(all_losses, dim=-1) |
|
return output_class_ranks.tolist() |
|
|
|
def predict_answers( |
|
self, |
|
samples, |
|
num_beams=5, |
|
inference_method="generate", |
|
max_len=10, |
|
min_len=1, |
|
num_ans_candidates=128, |
|
answer_list=None, |
|
prompt="", |
|
length_penalty=0, |
|
**kwargs |
|
): |
|
''' |
|
function for open-ended VQA |
|
''' |
|
images = samples["image"].cuda() |
|
texts = samples["instruction_input"] |
|
|
|
output_text = self.generate( |
|
images=images, |
|
texts=texts, |
|
num_beams=num_beams, |
|
max_new_tokens=max_len, |
|
min_length=min_len, |
|
length_penalty=length_penalty |
|
) |
|
|
|
if "apply_lemmatizer" in samples.keys() and samples["apply_lemmatizer"]: |
|
output_text = self._lemmatize(output_text) |
|
|
|
return output_text |
|
|
|
def predict_class( |
|
self, |
|
samples, |
|
num_beams=5, |
|
inference_method="generate", |
|
max_len=10, |
|
min_len=1, |
|
num_ans_candidates=5, |
|
answer_list=None, |
|
prompt="", |
|
length_penalty=0, |
|
**kwargs |
|
): |
|
''' |
|
function for multi-choice VQA |
|
''' |
|
|
|
image = samples["image"].cuda() |
|
instruction = samples['instruction_input'] |
|
answers = samples["choices"] |
|
num_cand = samples["num_choices"] |
|
|
|
ranks = self.multi_select(image, instruction, answers, num_cand) |
|
|
|
pred_ans = [] |
|
for i, rank in enumerate(ranks): |
|
pred = answers[rank[0]][i] |
|
pred_ans.append(pred) |
|
return pred_ans |
|
|
|
def embed_tokens(self, token_ids): |
|
try: |
|
embeds = self.llama_model.base_model.model.model.embed_tokens(token_ids) |
|
except AttributeError: |
|
embeds = self.llama_model.model.embed_tokens(token_ids) |
|
|
|
return embeds |
|
|
|
@classmethod |
|
def from_config(cls, cfg): |
|
model = cls( |
|
cfg=cfg, |
|
) |
|
ckpt_path = cfg.get("ckpt", "") |
|
if ckpt_path: |
|
print("Load Minigpt-4-LLM Checkpoint: {}".format(ckpt_path)) |
|
ckpt = torch.load(ckpt_path, map_location="cpu") |
|
msg = model.load_state_dict(ckpt['model'], strict=False) |
|
|
|
model.to('cuda') |
|
|
|
video_config = minigpt4_video_config(cfg) |
|
|
|
|
|
|
|
return model |
|
|
|
|
|
def assign_imgs(batched_instruct_list, batched_img_embeds): |
|
'''this function is used when the data is interleaved. |
|
the interlevaed data is separated, and this function assign |
|
corresponding image embeddings to each segment''' |
|
if len(batched_img_embeds.shape) == 3: |
|
batched_img_embeds = batched_img_embeds[:, None] |
|
|
|
batched_assigned = [] |
|
|
|
for instruct_list, img_embeds in zip(batched_instruct_list, batched_img_embeds): |
|
img_idx = 0 |
|
assigned_img = [] |
|
n_assigned = [] |
|
for instruct in instruct_list: |
|
n_img = instruct.count('<ImageHere>') |
|
if n_img > 0: |
|
assigned_img.append(img_embeds[None, img_idx:img_idx+n_img]) |
|
img_idx += n_img |
|
n_assigned.append(n_img) |
|
else: |
|
assigned_img.append(None) |
|
n_assigned.append(None) |
|
batched_assigned.append(assigned_img) |
|
|
|
return batched_assigned |
|
|