Skip to content

Commit

Permalink
fix pct_3D_points default value
Browse files Browse the repository at this point in the history
  • Loading branch information
Francesco Pittaluga committed Nov 6, 2019
1 parent 50dd44c commit ea2ef4e
Show file tree
Hide file tree
Showing 3 changed files with 6 additions and 6 deletions.
4 changes: 2 additions & 2 deletions train_coarse.py
Original file line number Diff line number Diff line change
Expand Up @@ -51,9 +51,9 @@
parser.add_argument("--crop_size", type=int, default=256, help="%(type)s: Size to crop images to (default: %(default)s)")
parser.add_argument("--scale_size", type=lambda s: [int(i) for i in s.split(',')], default=[296,394,512],
help="int,int,int: Sizes to randomly scale images to before cropping them (default: 296,394,512)")
parser.add_argument("--pct_3D_points", type=lambda s: [float(i) for i in s.split(',')][:2], default=[.05,1.0],
parser.add_argument("--pct_3D_points", type=lambda s: [float(i) for i in s.split(',')][:2], default=[5.,100.],
help="float,float: Min and max percent of 3D points to keep when performing random subsampling for data augmentation "+\
"(default: .05,1.0)")
"(default: 5.,100.)")
parser.add_argument("--per_loss_wt", type=float, default=1., help="%(type)s: Perceptual loss weight (default: %(default)s)")
parser.add_argument("--pix_loss_wt", type=float, default=1., help="%(type)s: Pixel loss weight (default: %(default)s)")
parser.add_argument("--max_iter", type=int, default=1e6, help="%(type)s: Stop training after MAX_ITER iterations (default: %(default)s)")
Expand Down
4 changes: 2 additions & 2 deletions train_refine.py
Original file line number Diff line number Diff line change
Expand Up @@ -54,9 +54,9 @@
parser.add_argument("--crop_size", type=int, default=256, help="%(type)s: Size to crop images to (default: %(default)s)")
parser.add_argument("--scale_size", type=lambda s: [int(i) for i in s.split(',')], default=[296,394,512],
help="int,int,int: Sizes to randomly scale images to before cropping them (default: 296,394,512)")
parser.add_argument("--pct_3D_points", type=lambda s: [float(i) for i in s.split(',')][:2], default=[.05,1.0],
parser.add_argument("--pct_3D_points", type=lambda s: [float(i) for i in s.split(',')][:2], default=[5.,100.],
help="float,float: Min and max percent of 3D points to keep when performing random subsampling for data augmentation "+\
"(default: .05,1.0)")
"(default: 5.,100.)")
parser.add_argument("--per_loss_wt", type=float, default=1., help="%(type)s: Perceptual loss weight (default: %(default)s)")
parser.add_argument("--pix_loss_wt", type=float, default=1., help="%(type)s: Pixel loss weight (default: %(default)s)")
parser.add_argument("--adv_loss_wt", type=float, default=1e3, help="%(type)s: Adversarial loss weight (default: %(default)s)")
Expand Down
4 changes: 2 additions & 2 deletions train_visib.py
Original file line number Diff line number Diff line change
Expand Up @@ -47,9 +47,9 @@
parser.add_argument("--crop_size", type=int, default=256, help="%(type)s: Size to crop images to (default: %(default)s)")
parser.add_argument("--scale_size", type=lambda s: [int(i) for i in s.split(',')], default=[296,394,512],
help="int,int,int: Sizes to randomly scale images to before cropping them (default: 296,394,512)")
parser.add_argument("--pct_3D_points", type=lambda s: [float(i) for i in s.split(',')][:2], default=[.05,1.0],
parser.add_argument("--pct_3D_points", type=lambda s: [float(i) for i in s.split(',')][:2], default=[5.,100.],
help="float,float: Min and max percent of 3D points to keep when performing random subsampling for data augmentation "+\
"(default: .05,1.0)")
"(default: 5.,100.)")
parser.add_argument("--vis_thresh", type=float, default=.05, help="%(type)s: Threshold used to compute ground truth visibility mask."+\
"i.e., gt_visibibility_mask = ((inp_depth-gt_depth)/gt_depth) > VISIB_THRESH. (default: %(default)s)")
parser.add_argument("--max_iter", type=int, default=1e6, help="%(type)s: Stop training after MAX_ITER iterations (default: %(default)s)")
Expand Down

0 comments on commit ea2ef4e

Please sign in to comment.