Spaces:
Sleeping
Sleeping
import torch | |
import numpy as np | |
import scipy | |
from config.config import cfg | |
from torch.nn import functional as F | |
import torchgeometry as tgm | |
def cam2pixel(cam_coord, f, c): | |
x = cam_coord[:, 0] / cam_coord[:, 2] * f[0] + c[0] | |
y = cam_coord[:, 1] / cam_coord[:, 2] * f[1] + c[1] | |
z = cam_coord[:, 2] | |
return np.stack((x, y, z), 1) | |
def pixel2cam(pixel_coord, f, c): | |
x = (pixel_coord[:, 0] - c[0]) / f[0] * pixel_coord[:, 2] | |
y = (pixel_coord[:, 1] - c[1]) / f[1] * pixel_coord[:, 2] | |
z = pixel_coord[:, 2] | |
return np.stack((x, y, z), 1) | |
def world2cam(world_coord, R, t): | |
cam_coord = np.dot(R, world_coord.transpose(1, 0)).transpose( | |
1, 0) + t.reshape(1, 3) | |
return cam_coord | |
def cam2world(cam_coord, R, t): | |
world_coord = np.dot(np.linalg.inv(R), | |
(cam_coord - t.reshape(1, 3)).transpose(1, | |
0)).transpose( | |
1, 0) | |
return world_coord | |
def rigid_transform_3D(A, B): | |
n, dim = A.shape | |
centroid_A = np.mean(A, axis=0) | |
centroid_B = np.mean(B, axis=0) | |
H = np.dot(np.transpose(A - centroid_A), B - centroid_B) / n | |
U, s, V = np.linalg.svd(H) | |
R = np.dot(np.transpose(V), np.transpose(U)) | |
if np.linalg.det(R) < 0: | |
s[-1] = -s[-1] | |
V[2] = -V[2] | |
R = np.dot(np.transpose(V), np.transpose(U)) | |
varP = np.var(A, axis=0).sum() | |
c = 1 / varP * np.sum(s) | |
t = -np.dot(c * R, np.transpose(centroid_A)) + np.transpose(centroid_B) | |
return c, R, t | |
def rigid_transform_3D_batch(A, B): | |
n, dim = A.shape | |
centroid_A = np.mean(A, axis=0) | |
centroid_B = np.mean(B, axis=0) | |
H = np.dot(np.transpose(A - centroid_A), B - centroid_B) / n | |
U, s, V = np.linalg.svd(H) | |
R = np.dot(np.transpose(V), np.transpose(U)) | |
if np.linalg.det(R) < 0: | |
s[-1] = -s[-1] | |
V[2] = -V[2] | |
R = np.dot(np.transpose(V), np.transpose(U)) | |
varP = np.var(A, axis=0).sum() | |
c = 1 / varP * np.sum(s) | |
t = -np.dot(c * R, np.transpose(centroid_A)) + np.transpose(centroid_B) | |
A2 = np.transpose(np.dot(c * R, np.transpose(A))) + t | |
return A2 | |
def rigid_align(A, B): | |
c, R, t = rigid_transform_3D(A, B) | |
A2 = np.transpose(np.dot(c * R, np.transpose(A))) + t | |
return A2 | |
def rigid_align_batch(A, B): | |
A2 = np.stack([ | |
rigid_transform_3D_batch(a_i, b_i) | |
for a_i, b_i in zip(A, B) | |
]) | |
return A2 | |
def transform_joint_to_other_db(src_joint, src_name, dst_name): | |
src_joint_num = len(src_name) | |
dst_joint_num = len(dst_name) | |
new_joint = np.zeros(((dst_joint_num, ) + src_joint.shape[1:]), | |
dtype=np.float32) | |
for src_idx in range(len(src_name)): | |
name = src_name[src_idx] | |
if name in dst_name: | |
dst_idx = dst_name.index(name) | |
new_joint[dst_idx] = src_joint[src_idx] | |
return new_joint | |
def transform_joint_to_other_db_batch(src_joint, src_name, dst_name): | |
src_joint_num = len(src_name) | |
dst_joint_num = len(dst_name) | |
person_num = src_joint.shape[0] | |
new_joint = np.zeros((( | |
person_num, | |
dst_joint_num, | |
) + src_joint.shape[2:]), | |
dtype=np.float32) | |
for src_idx in range(len(src_name)): | |
name = src_name[src_idx] | |
if name in dst_name: | |
dst_idx = dst_name.index(name) | |
new_joint[:, dst_idx] = src_joint[:, src_idx] | |
return new_joint | |
def rot6d_to_axis_angle(x): | |
batch_size = x.shape[0] | |
x = x.view(-1, 3, 2) | |
a1 = x[:, :, 0] | |
a2 = x[:, :, 1] | |
b1 = F.normalize(a1) | |
b2 = F.normalize(a2 - torch.einsum('bi,bi->b', b1, a2).unsqueeze(-1) * b1) | |
b3 = torch.cross(b1, b2) | |
rot_mat = torch.stack((b1, b2, b3), dim=-1) # 3x3 rotation matrix | |
rot_mat = torch.cat( | |
[rot_mat, torch.zeros( | |
(batch_size, 3, 1)).cuda().float()], 2) # 3x4 rotation matrix | |
axis_angle = tgm.rotation_matrix_to_angle_axis(rot_mat).reshape( | |
-1, 3) # axis-angle | |
axis_angle[torch.isnan(axis_angle)] = 0.0 | |
return axis_angle | |
def sample_joint_features(img_feat, joint_xy): | |
height, width = img_feat.shape[2:] | |
x = joint_xy[:, :, 0] / (width - 1) * 2 - 1 | |
y = joint_xy[:, :, 1] / (height - 1) * 2 - 1 | |
grid = torch.stack((x, y), 2)[:, :, None, :] | |
img_feat = F.grid_sample( | |
img_feat, grid, | |
align_corners=True)[:, :, :, 0] # batch_size, channel_dim, joint_num | |
img_feat = img_feat.permute( | |
0, 2, 1).contiguous() # batch_size, joint_num, channel_dim | |
return img_feat | |
def soft_argmax_2d(heatmap2d): | |
batch_size = heatmap2d.shape[0] | |
height, width = heatmap2d.shape[2:] | |
heatmap2d = heatmap2d.reshape((batch_size, -1, height * width)) | |
heatmap2d = F.softmax(heatmap2d, 2) | |
heatmap2d = heatmap2d.reshape((batch_size, -1, height, width)) | |
accu_x = heatmap2d.sum(dim=(2)) | |
accu_y = heatmap2d.sum(dim=(3)) | |
accu_x = accu_x * torch.arange(width).float().cuda()[None, None, :] | |
accu_y = accu_y * torch.arange(height).float().cuda()[None, None, :] | |
accu_x = accu_x.sum(dim=2, keepdim=True) | |
accu_y = accu_y.sum(dim=2, keepdim=True) | |
coord_out = torch.cat((accu_x, accu_y), dim=2) | |
return coord_out | |
def soft_argmax_3d(heatmap3d): | |
batch_size = heatmap3d.shape[0] | |
depth, height, width = heatmap3d.shape[2:] | |
heatmap3d = heatmap3d.reshape((batch_size, -1, depth * height * width)) | |
heatmap3d = F.softmax(heatmap3d, 2) | |
heatmap3d = heatmap3d.reshape((batch_size, -1, depth, height, width)) | |
accu_x = heatmap3d.sum(dim=(2, 3)) | |
accu_y = heatmap3d.sum(dim=(2, 4)) | |
accu_z = heatmap3d.sum(dim=(3, 4)) | |
accu_x = accu_x * torch.arange(width).float().cuda()[None, None, :] | |
accu_y = accu_y * torch.arange(height).float().cuda()[None, None, :] | |
accu_z = accu_z * torch.arange(depth).float().cuda()[None, None, :] | |
accu_x = accu_x.sum(dim=2, keepdim=True) | |
accu_y = accu_y.sum(dim=2, keepdim=True) | |
accu_z = accu_z.sum(dim=2, keepdim=True) | |
coord_out = torch.cat((accu_x, accu_y, accu_z), dim=2) | |
return coord_out | |
def restore_bbox(bbox_center, bbox_size, aspect_ratio, extension_ratio): | |
bbox = bbox_center.view(-1, 1, 2) + torch.cat( | |
(-bbox_size.view(-1, 1, 2) / 2., bbox_size.view(-1, 1, 2) / 2.), | |
1) # xyxy in (cfg.output_hm_shape[2], cfg.output_hm_shape[1]) space | |
bbox[:, :, | |
0] = bbox[:, :, 0] / cfg.output_hm_shape[2] * cfg.input_body_shape[1] | |
bbox[:, :, | |
1] = bbox[:, :, 1] / cfg.output_hm_shape[1] * cfg.input_body_shape[0] | |
bbox = bbox.view(-1, 4) | |
# xyxy -> xywh | |
bbox[:, 2] = bbox[:, 2] - bbox[:, 0] | |
bbox[:, 3] = bbox[:, 3] - bbox[:, 1] | |
# aspect ratio preserving bbox | |
w = bbox[:, 2] | |
h = bbox[:, 3] | |
c_x = bbox[:, 0] + w / 2. | |
c_y = bbox[:, 1] + h / 2. | |
mask1 = w > (aspect_ratio * h) | |
mask2 = w < (aspect_ratio * h) | |
h[mask1] = w[mask1] / aspect_ratio | |
w[mask2] = h[mask2] * aspect_ratio | |
bbox[:, 2] = w * extension_ratio | |
bbox[:, 3] = h * extension_ratio | |
bbox[:, 0] = c_x - bbox[:, 2] / 2. | |
bbox[:, 1] = c_y - bbox[:, 3] / 2. | |
# xywh -> xyxy | |
bbox[:, 2] = bbox[:, 2] + bbox[:, 0] | |
bbox[:, 3] = bbox[:, 3] + bbox[:, 1] | |
return bbox | |