-
Notifications
You must be signed in to change notification settings - Fork 2
/
4.train_mlp_FreiHAND_bone_loss.py
125 lines (99 loc) · 4.07 KB
/
4.train_mlp_FreiHAND_bone_loss.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import argparse
import os.path as osp
import torch
import torch.optim as optim
import torch.nn as nn
import torch.nn.functional as F
from hand_shape_pose.config import cfg
from hand_shape_pose.model.pose_mlp_network import MLPPoseNetwork
from hand_shape_pose.model.net_hg import Net_HM_HG
from hand_shape_pose.data.build import build_dataset
from hand_shape_pose.util.logger import setup_logger, get_logger_filename
from hand_shape_pose.util.miscellaneous import mkdir
from hand_shape_pose.util.vis import save_batch_image_with_mesh_joints
#from hand_shape_pose.util import renderer
from hand_shape_pose.util.net_util import load_net_model
from hand_shape_pose.util.bone_constraint_loss_function import bone_constraint_loss
import time
def main():
parser = argparse.ArgumentParser(description="3D Hand Shape and Pose Inference")
parser.add_argument(
"--config-file",
default="configs/train_FreiHAND_dataset.yaml",
metavar="FILE",
help="path to config file",
)
parser.add_argument(
"opts",
help="Modify config options using the command-line",
default=None,
nargs=argparse.REMAINDER,
)
args = parser.parse_args()
cfg.merge_from_file(args.config_file)
cfg.merge_from_list(args.opts)
cfg.freeze()
# Load data
dataset_val = build_dataset(cfg.TRAIN.DATASET, cfg.TRAIN.BACKGROUND_SET, cfg.TRAIN.DATA_SIZE)
data_loader_val = torch.utils.data.DataLoader(
dataset_val,
batch_size=cfg.MODEL.BATCH_SIZE,
shuffle=True,
num_workers=cfg.MODEL.NUM_WORKERS
)
# Load network model
model = MLPPoseNetwork(cfg)
device = cfg.MODEL.DEVICE
model.to(device)
model.load_model(cfg)
model = model.train()
# fix hm model weight
for param in model.net_hm.parameters():
param.requires_grad = False
optimizer = optim.RMSprop(model.mlp.parameters(), lr=0.0001)
#optimizer = optim.Adam(model.mlp.parameters(), lr=0.00001)
pose_loss = nn.MSELoss(reduction='sum')
num_epoch = 400
for epoch in range(num_epoch):
# reduce learning rate every 50 epoch
if epoch % 50 == 0 and epoch != 0:
for param_group in optimizer.param_groups:
param_group['lr'] /= 10
tic = time.time()
total_loss_train = 0.0
tot_loss_len = 0.0
tot_loss_dir = 0.0
# model training per batcg
for i, batch in enumerate(data_loader_val):
images, cam_params, pose_roots, pose_scales, image_ids = batch
images, cam_params, pose_roots, pose_scales = \
images.to(device), cam_params.to(device), pose_roots.to(device), pose_scales.to(device)
# ground truth pose
gt_pose_cam_xyz = torch.Tensor().to(device)
for img_id in image_ids:
gt_pose_cam_xyz = torch.cat((gt_pose_cam_xyz, dataset_val.pose_gts[img_id].to(device)), 0)
gt_pose_cam_xyz = gt_pose_cam_xyz.view(-1, 21, 3)
# forward propagation
optimizer.zero_grad()
_, est_pose_uv, est_pose_cam_xyz = model(images, cam_params, pose_scales, pose_root=pose_roots)
# bone constraint loss
len_loss, dir_loss = bone_constraint_loss(est_pose_cam_xyz, gt_pose_cam_xyz, device)
loss = pose_loss(est_pose_cam_xyz, gt_pose_cam_xyz) + len_loss + dir_loss
# back propagation
loss.backward()
optimizer.step()
total_loss_train += loss.item()
tot_loss_len += len_loss.item()
tot_loss_dir += dir_loss.item()
# record time
toc = time.time()
print('loss of epoch %2d: %6.2f, L_len: %3.2f, L_dir: %5.2f, time: %0.4f s' %(int(epoch+1), total_loss_train, tot_loss_len, tot_loss_dir, toc-tic))
# save model weight every epoch
torch.save(model.mlp.state_dict(), "mlp.pth")
if __name__ == "__main__":
main()
#loss of epoch 90: 28.11, time: 2.4332 s