New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Config Day2Night | Split Datasets #96
Comments
Could you provide the link to the config file? |
@mingyuliutw yup # logger options
image_save_iter: 1000 # How often do you want to save output images during training
image_display_iter: 50000 # How often do you want to display output images during training
display_size: 8 # How many images do you want to display each time
snapshot_save_iter: 1000 # How often do you want to save trained models
snapshot_smart_override: True # Use to smart cleaning old snapshots while saving
log_iter: 10 # How often do you want to log the training stats
# optimization options
max_iter: 1000000 # maximum number of training iterations
batch_size: 1 # batch size
weight_decay: 0.0001 # weight decay
beta1: 0.5 # Adam parameter
beta2: 0.999 # Adam parameter
init: kaiming # initialization [gaussian/kaiming/xavier/orthogonal]
lr: 0.0001 # initial learning rate
lr_policy: step # learning rate scheduler
step_size: 100000 # how often to decay learning rate
gamma: 0.5 # how much to decay learning rate
gan_w: 1 # weight of adversarial loss
recon_x_w: 10 # weight of image reconstruction loss
recon_h_w: 0 # weight of hidden reconstruction loss
recon_kl_w: 0.01 # weight of KL loss for reconstruction
recon_x_cyc_w: 10 # weight of cycle consistency loss
recon_kl_cyc_w: 0.01 # weight of KL loss for cycle consistency
vgg_w: 0 # weight of domain-invariant perceptual loss
# model options
gen:
dim: 64 # number of filters in the bottommost layer
mlp_dim: 256 # number of filters in MLP
style_dim: 8 # length of style code
activ: relu # activation function [relu/lrelu/prelu/selu/tanh]
n_downsample: 2 # number of downsampling layers in content encoder
n_res: 4 # number of residual blocks in content encoder/decoder
pad_type: reflect # padding type [zero/reflect]
dis:
dim: 64 # number of filters in the bottommost layer
norm: none # normalization layer [none/bn/in/ln]
activ: lrelu # activation function [relu/lrelu/prelu/selu/tanh]
n_layer: 4 # number of layers in D
gan_type: lsgan # GAN loss [lsgan/nsgan]
num_scales: 3 # number of scales
pad_type: reflect # padding type [zero/reflect]
# data options
input_dim_a: 3 # number of image channels [1/3]
input_dim_b: 3 # number of image channels [1/3]
num_workers: 8 # number of data loading threads
new_size: 256 # first resize the shortest image side to this size
crop_image_height: 256 # random crop image of this height
crop_image_width: 256 # random crop image of this width
data_root: /mnt/w/prj/data/nexet/nexet_2017_1 # dataset folder location |
@mingyuliutw i think it is dataset problem, it is NEXET dataset.
Also i found some labels mismatches after processing dataset with simple python script detecting time of day by color intensity. here Twilight images be like Some doubtful image's labels:This is Day (label) (and it's correct, because it is tunnel) |
@mingyuliutw one more question. Can i add accuracy to tensorboardX somehow? I found you calculating loss in UNIT. But i didn't find anything about calculation accuracy between domain-translated images and source. In paper you describe how you done experiments: Do you have any code of this? It'll be cool if you share that. For example here:
but how can i determine that model trained better? |
The results look nice. I obtained my original results without using the domain invariant VGG loss too. As I mentioned, the VGG loss helps for some datasets, particularly for synthetic to real. Yes, you can add accuracy to tensorboard. |
|
Hi, there, i'm trying to train dataset and stucked with some learning problems
Example of 10000 iterations day 2 night (train)
Example of 10000 iterations night 2 day (train)
I think that it's
unit_day2night.yaml
configuration problem. Can you share yours config file?Another option, that it's dataset's split issue.
I used 50/50 for train and test. And both divided into two equal domains A and B.
The text was updated successfully, but these errors were encountered: