MST-MIXER/models/utils.py

242 lines
No EOL
10 KiB
Python

import torch
import torch.nn as nn
import torch.nn.functional as F
from transformers.utils import ModelOutput
from typing import Optional, Tuple
class ELBO(nn.Module):
def __init__(self):
super(ELBO, self).__init__()
def forward(self, QA, PA):
QA_flattened = QA.view(-1).unsqueeze(-1)
PA_flattened = PA.view(-1).unsqueeze(-1)
QA_flattened = torch.cat([torch.zeros_like(QA_flattened), QA_flattened], dim=-1)
PA_flattened = torch.cat([torch.zeros_like(PA_flattened), PA_flattened], dim=-1)
log_QA = F.log_softmax(QA_flattened, dim=1)
log_PA = F.log_softmax(PA_flattened, dim=1)
QA_dist = torch.exp(log_QA)
loss_QA = torch.mean(log_QA * QA_dist)
loss_PA = torch.mean(log_PA * QA_dist)
loss = loss_QA - loss_PA
return loss
def seperate_nextqa_input_modalities(
features, i3d_rgb_interval, i3d_flow_interval, question_intervals,
vis_state_vector_idx, question_state_vector_idx,
attention_values=None):
""" We separate the multimodal input hidden states. The state token embeddings are left out (+1 while indexing)
Args:
features (_type_): _description_
i3d_rgb_interval (_type_): _description_
i3d_flow_interval (_type_): _description_
sam_interval (_type_): _description_
audio_interval (_type_): _description_
history_intervals (_type_): _description_
question_intervals (_type_): _description_
Returns:
_type_: _description_
"""
features_copy = features.clone() # .detach()
i3d_rgb_hidden = features_copy[:, i3d_rgb_interval[0]+1:i3d_rgb_interval[1], :]
i3d_flow_hidden = features_copy[:, i3d_flow_interval[0]+1:i3d_flow_interval[1], :]
question_hidden = []
features_split = torch.split(features_copy, 1, dim=0)
for ques_inter, feat in zip(question_intervals, features_split):
ques_idx = torch.arange(ques_inter[0]+1, ques_inter[1]).unsqueeze(0).unsqueeze(-1).repeat(1, 1, feat.size(-1)).to(feat.device)
question_hidden.append(torch.gather(feat, 1, ques_idx))
if attention_values is None:
i3d_rgb_att = None
i3d_flow_att = None
question_att = None
else:
attention_values = attention_values.mean(1)
i3d_rgb_att = attention_values[:, vis_state_vector_idx[0], vis_state_vector_idx[0]+1:vis_state_vector_idx[1]]
i3d_flow_att = attention_values[:, vis_state_vector_idx[1], vis_state_vector_idx[1]+1:question_state_vector_idx[0]]
question_att = [attention_values[i, question_state_vector_idx[i], question_intervals[i][0] + 1: question_intervals[i][1]] for i in range(len(question_state_vector_idx))]
features_list = [i3d_rgb_hidden, i3d_flow_hidden, question_hidden]
att = [i3d_rgb_att, i3d_flow_att, question_att]
return features_list, att
def seperate_input_modalities(
features, i3d_rgb_interval, i3d_flow_interval, sam_interval, audio_interval, history_intervals, question_intervals,
vis_state_vector_idx, history_state_vector_idx, question_state_vector_idx,
attention_values=None):
""" We separate the multimodal input hidden states. The state token embeddings are left out (+1 while indexing)
Args:
features (_type_): _description_
i3d_rgb_interval (_type_): _description_
i3d_flow_interval (_type_): _description_
sam_interval (_type_): _description_
audio_interval (_type_): _description_
history_intervals (_type_): _description_
question_intervals (_type_): _description_
Returns:
_type_: _description_
"""
features_copy = features.clone() # .detach()
i3d_rgb_hidden = features_copy[:, i3d_rgb_interval[0]+1:i3d_rgb_interval[1], :]
i3d_flow_hidden = features_copy[:, i3d_flow_interval[0]+1:i3d_flow_interval[1], :]
sam_hidden = features_copy[:, sam_interval[0]+1:sam_interval[1], :]
audio_hidden = features_copy[:, audio_interval[0]+1:audio_interval[1], :]
history_hidden = []
question_hidden = []
features_split = torch.split(features_copy, 1, dim=0)
for hist_inter, ques_inter, feat in zip(history_intervals, question_intervals, features_split):
hist_idx = torch.arange(hist_inter[0]+1, hist_inter[1]).unsqueeze(0).unsqueeze(-1).repeat(1, 1, feat.size(-1)).to(feat.device)
history_hidden.append(torch.gather(feat, 1, hist_idx))
ques_idx = torch.arange(ques_inter[0]+1, ques_inter[1]).unsqueeze(0).unsqueeze(-1).repeat(1, 1, feat.size(-1)).to(feat.device)
question_hidden.append(torch.gather(feat, 1, ques_idx))
if attention_values is None:
i3d_rgb_att = None
i3d_flow_att = None
sam_att = None
audio_att = None
history_att = None
question_att = None
else:
attention_values = attention_values.mean(1)
i3d_rgb_att = attention_values[:, vis_state_vector_idx[0], vis_state_vector_idx[0]+1:vis_state_vector_idx[1]]
i3d_flow_att = attention_values[:, vis_state_vector_idx[1], vis_state_vector_idx[1]+1:vis_state_vector_idx[2]]
sam_att = attention_values[:, vis_state_vector_idx[2], vis_state_vector_idx[2]+1:vis_state_vector_idx[3]]
audio_att = attention_values[:, vis_state_vector_idx[3], vis_state_vector_idx[3]+1:history_state_vector_idx[0] - 1]
history_att = [attention_values[i, history_state_vector_idx[i], history_intervals[i][0] + 1 : history_intervals[i][1]] for i in range(len(history_state_vector_idx))]
question_att = [attention_values[i, question_state_vector_idx[i], question_intervals[i][0] + 1: question_intervals[i][1]] for i in range(len(question_state_vector_idx))]
features_list = [i3d_rgb_hidden, i3d_flow_hidden, sam_hidden, audio_hidden, history_hidden, question_hidden]
att = [i3d_rgb_att, i3d_flow_att, sam_att, audio_att, history_att, question_att]
return features_list, att
def get_knn_graph(features, num_nn, device):
features = features.permute((1, 2, 0))
cosine_sim_pairwise = F.cosine_similarity(features, features.unsqueeze(1), dim=-2)
cosine_sim_pairwise = cosine_sim_pairwise.permute((2, 0, 1))
num_nn = min(num_nn, cosine_sim_pairwise.size(-1))
adj_mat = torch.zeros_like(cosine_sim_pairwise).to(device)
_, to_keep = torch.topk(cosine_sim_pairwise, num_nn, dim=-1, sorted=False)
adj_mat = adj_mat.scatter(-1, to_keep, torch.ones_like(adj_mat).to(device))
return adj_mat
def track_features_vis(features, att, top_k, device, node_idx=None):
top_k = min(features.size(1), top_k)
if att is None:
node_idx = torch.randint(low=0, high=features.size(1), size=(features.size(0), top_k))
else:
_, node_idx = torch.topk(att, top_k, dim=-1, sorted=False)
node_idx = node_idx.unsqueeze(-1).repeat(1, 1, features.size(-1)).to(device)
selected_features = torch.gather(features, 1, node_idx)
return selected_features, node_idx
def track_features_text(features, att, top_k, device, node_idx=None):
hidden_dim = features[0].size(-1)
min_len = min([feat.size(1) for feat in features])
top_k = min(min_len, top_k)
if att is None:
node_idx = [torch.randint(low=0, high=feat.size(1), size=(feat.size(0), top_k)) for feat in features]
else:
node_idx = [torch.topk(a, top_k, dim=-1, sorted=False)[-1] for a in att]
node_idx = [idx.unsqueeze(-1).repeat(1, 1, hidden_dim).to(device) for idx in node_idx]
selected_features = [torch.gather(feat, 1, idx) for feat, idx in zip(features, node_idx)]
selected_features = torch.cat(selected_features, dim=0)
return selected_features, node_idx
def diag_tensor(tensors):
device = tensors[0].device
n = sum([t.size(-1) for t in tensors])
bsz = tensors[0].size(0)
diag_tensor = torch.zeros((bsz, n, n)).float().to(device)
delimiter = 0
delimiters = [0]
for t in tensors:
diag_tensor[:, delimiter:delimiter+t.size(-1), delimiter:delimiter+t.size(-1)] = t
delimiter += t.size(-1)
delimiters.append(delimiter)
return diag_tensor, delimiters
def embed_graphs(features, delimiters):
state_vectors = []
for i in range(len(delimiters) - 1):
state_vectors.append(features[:, delimiters[i]:delimiters[i+1], :].mean(dim=1))
return state_vectors
class AVSDEncoderOutput(ModelOutput):
last_hidden_state: torch.FloatTensor = None
hidden_states: Optional[Tuple[torch.FloatTensor]] = None
attentions: Optional[Tuple[torch.FloatTensor]] = None
QAs_local = None
PAs_local = None
QA_global = None
PA_global = None
state_vectors = None
class AVSDSeq2SeqModelOutput(ModelOutput):
last_hidden_state: torch.FloatTensor = None
past_key_values: Optional[Tuple[Tuple[torch.FloatTensor]]] = None
decoder_hidden_states: Optional[Tuple[torch.FloatTensor]] = None
decoder_attentions: Optional[Tuple[torch.FloatTensor]] = None
cross_attentions: Optional[Tuple[torch.FloatTensor]] = None
encoder_last_hidden_state: Optional[torch.FloatTensor] = None
encoder_hidden_states: Optional[Tuple[torch.FloatTensor]] = None
encoder_attentions: Optional[Tuple[torch.FloatTensor]] = None
QAs_local = None
PAs_local = None
QA_global = None
PA_global = None
state_vectors = None
class AVSDSeq2SeqLMOutput(ModelOutput):
gen_loss: Optional[torch.FloatTensor] = None
elbo_loss_global: Optional[torch.FloatTensor] = None
elbo_loss_local: Optional[torch.FloatTensor] = None
logits: torch.FloatTensor = None
past_key_values: Optional[Tuple[Tuple[torch.FloatTensor]]] = None
decoder_hidden_states: Optional[Tuple[torch.FloatTensor]] = None
decoder_attentions: Optional[Tuple[torch.FloatTensor]] = None
cross_attentions: Optional[Tuple[torch.FloatTensor]] = None
encoder_last_hidden_state: Optional[torch.FloatTensor] = None
encoder_hidden_states: Optional[Tuple[torch.FloatTensor]] = None
encoder_attentions: Optional[Tuple[torch.FloatTensor]] = None
encoder_QAs_local = None
encoder_PAs_local = None
encoder_QA_global = None
encoder_PA_global = None
encoder_state_vectors = None