Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The performance of detection in VOC #6

Closed
jancylee opened this issue Mar 28, 2021 · 21 comments
Closed

The performance of detection in VOC #6

jancylee opened this issue Mar 28, 2021 · 21 comments

Comments

@jancylee
Copy link

(8gpus) When I use the pretrained network with coco-800ep-resnet50 to do the detection task with VOC, the "AP" is only 44.76, while you can achieve 56.7. I don't konw why the gap is so large. Note that I change the batchsize from 16 to 8, and as a result, the base lr is set from 0.02 to 0.01.

@WXinlong
Copy link
Owner

@jancylee Did you directly download the provided pre-trained weights? Please provide your experiment scripts and I will see what's the problem.

@jancylee
Copy link
Author

I pretrain the model by myself (coco-800ep-resnet50).

@jancylee
Copy link
Author

Compared to your codes, I didn't change any settings in the pretrained process, and I only change the batchsize and base lr in the detection process.

@WXinlong
Copy link
Owner

Please make sure you have followed the instructions in the readme: https://github.com/WXinlong/DenseCL#extracting-backbone-weights

You have to 1) extract the backbone weights and 2) convert it to detectron2 format.

@jancylee
Copy link
Author

I did transform it. By the way, when I train the detection process from scratch (without loading pretrained model), the AP is only 12.8.

@WXinlong
Copy link
Owner

It looks like the problems are in your detection experiments, not pre-trained weights. You are suggested to first reproduce the detection results using either random init. or supervised pretrained model, i.e., to make sure you can get the same results with the same settings.

@jancylee
Copy link
Author

I use 4 gpus, batchsize=4, base lr =0.005, iter=240004, steps = 180004,22000*4
In your settings, 8qpus, batchsize=16, base lr = 0.02, iter=24000, steps=18000,22000
And I download your pretrained model (coco-800ep-resnet50), the performance is AP=51.19, yours is 56.7, it's still a large gap........What should I change settings to achieve 56.7?

@zzzzzz0407
Copy link

@jancylee Please directly copy your training config.yaml instead of several parameters, we should make sure you set the correct parameters (e.g. input format: RGB, pixel mean/std and etc).

@jancylee
Copy link
Author

The config.yaml of VOC07&12 object dection:
CUDNN_BENCHMARK: false
DATALOADER:
ASPECT_RATIO_GROUPING: true
FILTER_EMPTY_ANNOTATIONS: true
NUM_WORKERS: 4
REPEAT_THRESHOLD: 0.0
SAMPLER_TRAIN: TrainingSampler
DATASETS:
PRECOMPUTED_PROPOSAL_TOPK_TEST: 1000
PRECOMPUTED_PROPOSAL_TOPK_TRAIN: 2000
PROPOSAL_FILES_TEST: []
PROPOSAL_FILES_TRAIN: []
TEST:

  • voc_2007_test
    TRAIN:
  • voc_2007_trainval
  • voc_2012_trainval
    GLOBAL:
    HACK: 1.0
    INPUT:
    CROP:
    ENABLED: false
    SIZE:
    • 0.9
    • 0.9
      TYPE: relative_range
      FORMAT: RGB
      MASK_FORMAT: polygon
      MAX_SIZE_TEST: 1333
      MAX_SIZE_TRAIN: 1333
      MIN_SIZE_TEST: 800
      MIN_SIZE_TRAIN:
  • 480
  • 512
  • 544
  • 576
  • 608
  • 640
  • 672
  • 704
  • 736
  • 768
  • 800
    MIN_SIZE_TRAIN_SAMPLING: choice
    RANDOM_FLIP: horizontal
    MODEL:
    ANCHOR_GENERATOR:
    ANGLES:
      • -90
      • 0
      • 90
        ASPECT_RATIOS:
      • 0.5
      • 1.0
      • 2.0
        NAME: DefaultAnchorGenerator
        OFFSET: 0.0
        SIZES:
      • 32
      • 64
      • 128
      • 256
      • 512
        BACKBONE:
        FREEZE_AT: 0
        NAME: build_resnet_backbone
        DEVICE: cuda
        FPN:
        FUSE_TYPE: sum
        IN_FEATURES: []
        NORM: ''
        OUT_CHANNELS: 256
        KEYPOINT_ON: false
        LOAD_PROPOSALS: false
        MASK_ON: false
        META_ARCHITECTURE: GeneralizedRCNN
        PANOPTIC_FPN:
        COMBINE:
        ENABLED: true
        INSTANCES_CONFIDENCE_THRESH: 0.5
        OVERLAP_THRESH: 0.5
        STUFF_AREA_LIMIT: 4096
        INSTANCE_LOSS_WEIGHT: 1.0
        PIXEL_MEAN:
  • 123.675
  • 116.28
  • 103.53
    PIXEL_STD:
  • 58.395
  • 57.12
  • 57.375
    PROPOSAL_GENERATOR:
    MIN_SIZE: 0
    NAME: RPN
    RESNETS:
    DEFORM_MODULATED: false
    DEFORM_NUM_GROUPS: 1
    DEFORM_ON_PER_STAGE:
    • false
    • false
    • false
    • false
      DEPTH: 50
      NORM: SyncBN
      NUM_GROUPS: 1
      OUT_FEATURES:
    • res4
      RES2_OUT_CHANNELS: 256
      RES5_DILATION: 1
      STEM_OUT_CHANNELS: 64
      STRIDE_IN_1X1: false
      WIDTH_PER_GROUP: 64
      RETINANET:
      BBOX_REG_LOSS_TYPE: smooth_l1
      BBOX_REG_WEIGHTS: &id001
    • 1.0
    • 1.0
    • 1.0
    • 1.0
      FOCAL_LOSS_ALPHA: 0.25
      FOCAL_LOSS_GAMMA: 2.0
      IN_FEATURES:
    • p3
    • p4
    • p5
    • p6
    • p7
      IOU_LABELS:
    • 0
    • -1
    • 1
      IOU_THRESHOLDS:
    • 0.4
    • 0.5
      NMS_THRESH_TEST: 0.5
      NORM: ''
      NUM_CLASSES: 80
      NUM_CONVS: 4
      PRIOR_PROB: 0.01
      SCORE_THRESH_TEST: 0.05
      SMOOTH_L1_LOSS_BETA: 0.1
      TOPK_CANDIDATES_TEST: 1000
      ROI_BOX_CASCADE_HEAD:
      BBOX_REG_WEIGHTS:
      • 10.0
      • 10.0
      • 5.0
      • 5.0
      • 20.0
      • 20.0
      • 10.0
      • 10.0
      • 30.0
      • 30.0
      • 15.0
      • 15.0
        IOUS:
    • 0.5
    • 0.6
    • 0.7
      ROI_BOX_HEAD:
      BBOX_REG_LOSS_TYPE: smooth_l1
      BBOX_REG_LOSS_WEIGHT: 1.0
      BBOX_REG_WEIGHTS:
    • 10.0
    • 10.0
    • 5.0
    • 5.0
      CLS_AGNOSTIC_BBOX_REG: false
      CONV_DIM: 256
      FC_DIM: 1024
      NAME: ''
      NORM: ''
      NUM_CONV: 0
      NUM_FC: 0
      POOLER_RESOLUTION: 14
      POOLER_SAMPLING_RATIO: 0
      POOLER_TYPE: ROIAlignV2
      SMOOTH_L1_BETA: 0.0
      TRAIN_ON_PRED_BOXES: false
      ROI_HEADS:
      BATCH_SIZE_PER_IMAGE: 512
      IN_FEATURES:
    • res4
      IOU_LABELS:
    • 0
    • 1
      IOU_THRESHOLDS:
    • 0.5
      NAME: Res5ROIHeadsExtraNorm
      NMS_THRESH_TEST: 0.5
      NUM_CLASSES: 20
      POSITIVE_FRACTION: 0.25
      PROPOSAL_APPEND_GT: true
      SCORE_THRESH_TEST: 0.05
      ROI_KEYPOINT_HEAD:
      CONV_DIMS:
    • 512
    • 512
    • 512
    • 512
    • 512
    • 512
    • 512
    • 512
      LOSS_WEIGHT: 1.0
      MIN_KEYPOINTS_PER_IMAGE: 1
      NAME: KRCNNConvDeconvUpsampleHead
      NORMALIZE_LOSS_BY_VISIBLE_KEYPOINTS: true
      NUM_KEYPOINTS: 17
      POOLER_RESOLUTION: 14
      POOLER_SAMPLING_RATIO: 0
      POOLER_TYPE: ROIAlignV2
      ROI_MASK_HEAD:
      CLS_AGNOSTIC_MASK: false
      CONV_DIM: 256
      NAME: MaskRCNNConvUpsampleHead
      NORM: ''
      NUM_CONV: 0
      POOLER_RESOLUTION: 14
      POOLER_SAMPLING_RATIO: 0
      POOLER_TYPE: ROIAlignV2
      RPN:
      BATCH_SIZE_PER_IMAGE: 256
      BBOX_REG_LOSS_TYPE: smooth_l1
      BBOX_REG_LOSS_WEIGHT: 1.0
      BBOX_REG_WEIGHTS: *id001
      BOUNDARY_THRESH: -1
      HEAD_NAME: StandardRPNHead
      IN_FEATURES:
    • res4
      IOU_LABELS:
    • 0
    • -1
    • 1
      IOU_THRESHOLDS:
    • 0.3
    • 0.7
      LOSS_WEIGHT: 1.0
      NMS_THRESH: 0.7
      POSITIVE_FRACTION: 0.5
      POST_NMS_TOPK_TEST: 1000
      POST_NMS_TOPK_TRAIN: 2000
      PRE_NMS_TOPK_TEST: 6000
      PRE_NMS_TOPK_TRAIN: 12000
      SMOOTH_L1_BETA: 0.0
      SEM_SEG_HEAD:
      COMMON_STRIDE: 4
      CONVS_DIM: 128
      IGNORE_VALUE: 255
      IN_FEATURES:
    • p2
    • p3
    • p4
    • p5
      LOSS_WEIGHT: 1.0
      NAME: SemSegFPNHead
      NORM: GN
      NUM_CLASSES: 54
      WEIGHTS: coco_800ep_base_given.pkl
      OUTPUT_DIR: output/coco_800ep_base_given
      SEED: -1
      SOLVER:
      AMP:
      ENABLED: false
      BASE_LR: 0.005
      BIAS_LR_FACTOR: 1.0
      CHECKPOINT_PERIOD: 5000
      CLIP_GRADIENTS:
      CLIP_TYPE: value
      CLIP_VALUE: 1.0
      ENABLED: false
      NORM_TYPE: 2.0
      GAMMA: 0.1
      IMS_PER_BATCH: 4
      LR_SCHEDULER_NAME: WarmupMultiStepLR
      MAX_ITER: 96000
      MOMENTUM: 0.9
      NESTEROV: false
      REFERENCE_WORLD_SIZE: 0
      STEPS:
  • 72000
  • 88000
    WARMUP_FACTOR: 0.001
    WARMUP_ITERS: 100
    WARMUP_METHOD: linear
    WEIGHT_DECAY: 0.0001
    WEIGHT_DECAY_BIAS: 0.0001
    WEIGHT_DECAY_NORM: 0.0
    TEST:
    AUG:
    ENABLED: false
    FLIP: true
    MAX_SIZE: 4000
    MIN_SIZES:
    • 400
    • 500
    • 600
    • 700
    • 800
    • 900
    • 1000
    • 1100
    • 1200
      DETECTIONS_PER_IMAGE: 100
      EVAL_PERIOD: 0
      EXPECTED_RESULTS: []
      KEYPOINT_OKS_SIGMAS: []
      PRECISE_BN:
      ENABLED: true
      NUM_ITER: 200
      VERSION: 2
      VIS_PERIOD: 0

@jancylee
Copy link
Author

And when I use my preteained model (coco-800ep-resnet50, and the training settings is same to yours) to fine tune the object dection in VOC(the training settings is as above), it's only AP=48.16, compared to AP=51.19(your pretrained model and my fine-tuned object detection) and AP=56.7(result of your paper ).

@zzzzzz0407
Copy link

@jancylee Can you try to use the official model (e.g. mocov2) to reproduce their voc detection performance? https://github.com/open-mmlab/OpenSelfSup/blob/master/docs/MODEL_ZOO.md
In my opinion, their is no issue in the config you provided except the batchsize, i mean, maybe too much small bring in a performance drop. You can evaluate it with the official model.

@jancylee
Copy link
Author

jancylee commented Apr 1, 2021

Thanks a lot. I know the reason, when I use the batchsize=16, it achieves 56.54(conmpared to the paper: 56.7), but this is based on your pretrained model provided by your github website. While when I train the pretrained model by myself, it only achieves 49.78. I completely use your code and follow your pretraining settings. I don't konw why. By the way, I used the moco-v2 pretrained model to fine tune, it can achieves 53.92, which is almost the same as the your paper results.

@jancylee
Copy link
Author

jancylee commented Apr 1, 2021

I can provide you the training settings later. But the only difference is the workers_per_gpu, I set 8 while you set 4, which is only influence the data loading speed.

@zzzzzz0407
Copy link

@jancylee So the batch size plays a important role to train detection, can you provide the config for training DenseCL?

@jancylee
Copy link
Author

jancylee commented Apr 1, 2021

2021-03-25 09:31:12,868 - openselfsup - INFO - Environment info:

sys.platform: linux
Python: 3.7.10 (default, Feb 26 2021, 18:47:35) [GCC 7.3.0]
CUDA available: True
CUDA_HOME: /usr/local/cuda
NVCC: Cuda compilation tools, release 10.1, V10.1.243
GPU 0,1,2,3,4,5,6,7: Tesla V100-SXM2-32GB
GCC: gcc (Ubuntu 5.4.0-6ubuntu1~16.04.12) 5.4.0 20160609
PyTorch: 1.4.0
PyTorch compiling details: PyTorch built with:

  • GCC 7.3
  • Intel(R) Math Kernel Library Version 2020.0.2 Product Build 20200624 for Intel(R) 64 architecture applications
  • Intel(R) MKL-DNN v0.21.1 (Git Hash 7d2fd500bc78936d1d648ca713b901012f470dbc)
  • OpenMP 201511 (a.k.a. OpenMP 4.5)
  • NNPACK is enabled
  • CUDA Runtime 10.0
  • NVCC architecture flags: -gencode;arch=compute_37,code=sm_37;-gencode;arch=compute_50,code=sm_50;-gencode;arch=compute_60,code=sm_60;-gencode;arch=compute_61,code=sm_61;-gencode;arch=compute_70,code=sm_70;-gencode;arch=compute_75,code=sm_75;-gencode;arch=compute_37,code=compute_37
  • CuDNN 7.6.3
  • Magma 2.5.1
  • Build settings: BLAS=MKL, BUILD_NAMEDTENSOR=OFF, BUILD_TYPE=Release, CXX_FLAGS= -Wno-deprecated -fvisibility-inlines-hidden -fopenmp -DUSE_FBGEMM -DUSE_QNNPACK -DUSE_PYTORCH_QNNPACK -O2 -fPIC -Wno-narrowing -Wall -Wextra -Wno-missing-field-initializers -Wno-type-limits -Wno-array-bounds -Wno-unknown-pragmas -Wno-sign-compare -Wno-unused-parameter -Wno-unused-variable -Wno-unused-function -Wno-unused-result -Wno-strict-overflow -Wno-strict-aliasing -Wno-error=deprecated-declarations -Wno-stringop-overflow -Wno-error=pedantic -Wno-error=redundant-decls -Wno-error=old-style-cast -fdiagnostics-color=always -faligned-new -Wno-unused-but-set-variable -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Wno-stringop-overflow, DISABLE_NUMA=1, PERF_WITH_AVX=1, PERF_WITH_AVX2=1, PERF_WITH_AVX512=1, USE_CUDA=ON, USE_EXCEPTION_PTR=1, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_MKL=ON, USE_MKLDNN=ON, USE_MPI=OFF, USE_NCCL=ON, USE_NNPACK=ON, USE_OPENMP=ON, USE_STATIC_DISPATCH=OFF,

TorchVision: 0.5.0
OpenCV: 4.5.1
MMCV: 1.0.3
OpenSelfSup: 0.2.0+9e827db

2021-03-25 09:31:12,869 - openselfsup - INFO - Distributed training: True
2021-03-25 09:31:12,869 - openselfsup - INFO - Config:
/home/codes/DenseCL/configs/base.py
train_cfg = {}
test_cfg = {}
optimizer_config = dict() # grad_clip, coalesce, bucket_size_mb

yapf:disable

log_config = dict(
interval=50,
hooks=[
dict(type='TextLoggerHook'),
dict(type='TensorboardLoggerHook')
])

yapf:enable

runtime settings

dist_params = dict(backend='nccl')
cudnn_benchmark = True
log_level = 'INFO'
load_from = None
resume_from = None
workflow = [('train', 1)]

/home/codes/DenseCL/configs/selfsup/densecl/densecl_coco_800ep.py
base = '../../base.py'

model settings

model = dict(
type='DenseCL',
pretrained=None,
queue_len=65536,
feat_dim=128,
momentum=0.999,
loss_lambda=0.5,
backbone=dict(
type='ResNet',
depth=50,
in_channels=3,
out_indices=[4], # 0: conv-1, x: stage-x
norm_cfg=dict(type='BN')),
neck=dict(
type='DenseCLNeck',
in_channels=2048,
hid_channels=2048,
out_channels=128,
num_grid=None),
head=dict(type='ContrastiveHead', temperature=0.2))
#head2=dict(type='TripleHead', margin=0.3),
#head3=dict(type='ContrastiveLXNHead', temperature=0.2))

dataset settings

data_source_cfg = dict(
type='COCO',
memcached=True,
mclient_path='/mnt/lustre/share/memcached_client')
data_train_list = ''

data_train_root = '/data2/ImageDataset/coco/train2017/train2017/'

data_train_root = '/home/data/train2017/'
dataset_type = 'ContrastiveDataset'
img_norm_cfg = dict(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])
train_pipeline = [
dict(type='RandomResizedCrop', size=224, scale=(0.2, 1.)),
dict(
type='RandomAppliedTrans',
transforms=[
dict(
type='ColorJitter',
brightness=0.4,
contrast=0.4,
saturation=0.4,
hue=0.1)
],
p=0.8),
dict(type='RandomGrayscale', p=0.2),
dict(
type='RandomAppliedTrans',
transforms=[
dict(
type='GaussianBlur',
sigma_min=0.1,
sigma_max=2.0)
],
p=0.5),
dict(type='RandomHorizontalFlip'),
dict(type='ToTensor'),
dict(type='Normalize', **img_norm_cfg),
]
data = dict(
imgs_per_gpu=32, # total 32*8=256
workers_per_gpu=8,
drop_last=True,
train=dict(
type=dataset_type,
data_source=dict(
list_file=data_train_list, root=data_train_root,
**data_source_cfg),
pipeline=train_pipeline))

optimizer

optimizer = dict(type='SGD', lr=0.3, weight_decay=0.0001, momentum=0.9)

learning policy

lr_config = dict(policy='CosineAnnealing', min_lr=0.)
checkpoint_config = dict(interval=40)

runtime settings

total_epochs = 800

2021-03-25 09:31:12,871 - openselfsup - INFO - Set random seed to 0, deterministic: False
2021-03-25 09:31:21,655 - openselfsup - INFO - Start running, host: root@c354f9782387, work_dir: /home/codes/DenseCL/work_dirs/selfsup/densecl/densecl_coco_800ep
2021-03-25 09:31:21,655 - openselfsup - INFO - workflow: [('train', 1)], max: 800 epochs
2021-03-25 09:32:07,330 - openselfsup - INFO - Epoch [1][50/462] lr: 3.000e-01, eta: 3 days, 21:37:36, time: 0.912, data_time: 0.492, memory: 12516, loss_contra_single: 4.2904, loss_contra_dense: 4.3897, loss: 8.6800
2021-03-25 09:32:17,519 - openselfsup - INFO - Epoch [1][100/462] lr: 3.000e-01, eta: 2 days, 9:16:22, time: 0.204, data_time: 0.001, memory: 12516, loss_contra_single: 4.8449, loss_contra_dense: 4.7468, loss: 9.5917
2021-03-25 09:32:27,725 - openselfsup - INFO - Epoch [1][150/462] lr: 3.000e-01, eta: 1 day, 21:09:31, time: 0.204, data_time: 0.000, memory: 12516, loss_contra_single: 4.9559, loss_contra_dense: 4.7926, loss: 9.7485
2021-03-25 09:32:37,942 - openselfsup - INFO - Epoch [1][200/462] lr: 3.000e-01, eta: 1 day, 15:06:21, time: 0.204, data_time: 0.000, memory: 12516, loss_contra_single: 4.9797, loss_contra_dense: 4.7833, loss: 9.7630
2021-03-25 09:32:48,175 - openselfsup - INFO - Epoch [1][250/462] lr: 3.000e-01, eta: 1 day, 11:28:46, time: 0.205, data_time: 0.000, memory: 12516, loss_contra_single: 4.9750, loss_contra_dense: 4.7667, loss: 9.7417
2021-03-25 09:32:58,344 - openselfsup - INFO - Epoch [1][300/462] lr: 3.000e-01, eta: 1 day, 9:02:21, time: 0.203, data_time: 0.000, memory: 12516, loss_contra_single: 4.9372, loss_contra_dense: 4.7216, loss: 9.6588
2021-03-25 09:33:08,456 - openselfsup - INFO - Epoch [1][350/462] lr: 3.000e-01, eta: 1 day, 7:16:39, time: 0.202, data_time: 0.000, memory: 12516, loss_contra_single: 4.9337, loss_contra_dense: 4.7003, loss: 9.6340
2021-03-25 09:33:18,506 - openselfsup - INFO - Epoch [1][400/462] lr: 3.000e-01, eta: 1 day, 5:56:29, time: 0.201, data_time: 0.000, memory: 12516, loss_contra_single: 4.9463, loss_contra_dense: 4.6928, loss: 9.6391
2021-03-25 09:33:28,615 - openselfsup - INFO - Epoch [1][450/462] lr: 3.000e-01, eta: 1 day, 4:54:52, time: 0.202, data_time: 0.000, memory: 12516, loss_contra_single: 4.9608, loss_contra_dense: 4.6925, loss: 9.6534
2021-03-25 09:34:07,235 - openselfsup - INFO - Epoch [2][50/462] lr: 3.000e-01, eta: 1 day, 8:24:45, time: 0.700, data_time: 0.453, memory: 12516, loss_contra_single: 4.9793, loss_contra_dense: 4.6959, loss: 9.6752
2021-03-25 09:34:17,127 - openselfsup - INFO - Epoch [2][100/462] lr: 3.000e-01, eta: 1 day, 7:19:46, time: 0.198, data_time: 0.000, memory: 12516, loss_contra_single: 4.9893, loss_contra_dense: 4.6976, loss: 9.6869
2021-03-25 09:34:26,972 - openselfsup - INFO - Epoch [2][150/462] lr: 3.000e-01, eta: 1 day, 6:24:51, time: 0.197, data_time: 0.000, memory: 12516, loss_contra_single: 4.9949, loss_contra_dense: 4.7060, loss: 9.7009
2021-03-25 09:34:36,778 - openselfsup - INFO - Epoch [2][200/462] lr: 3.000e-01, eta: 1 day, 5:37:53, time: 0.196, data_time: 0.001, memory: 12516, loss_contra_single: 4.9986, loss_contra_dense: 4.7151, loss: 9.7137
2021-03-25 09:34:46,731 - openselfsup - INFO - Epoch [2][250/462] lr: 3.000e-01, eta: 1 day, 4:58:45, time: 0.199, data_time: 0.000, memory: 12516, loss_contra_single: 5.0037, loss_contra_dense: 4.7257, loss: 9.7293
2021-03-25 09:34:56,564 - openselfsup - INFO - Epoch [2][300/462] lr: 3.000e-01, eta: 1 day, 4:23:43, time: 0.197, data_time: 0.000, memory: 12516, loss_contra_single: 4.9995, loss_contra_dense: 4.7437, loss: 9.7433
2021-03-25 09:35:06,499 - openselfsup - INFO - Epoch [2][350/462] lr: 3.000e-01, eta: 1 day, 3:53:44, time: 0.199, data_time: 0.001, memory: 12516, loss_contra_single: 4.9955, loss_contra_dense: 4.7586, loss: 9.7541
2021-03-25 09:35:16,331 - openselfsup - INFO - Epoch [2][400/462] lr: 3.000e-01, eta: 1 day, 3:26:37, time: 0.197, data_time: 0.001, memory: 12516, loss_contra_single: 4.9901, loss_contra_dense: 4.7736, loss: 9.7637
2021-03-25 09:35:26,245 - openselfsup - INFO - Epoch [2][450/462] lr: 3.000e-01, eta: 1 day, 3:02:55, time: 0.198, data_time: 0.000, memory: 12516, loss_contra_single: 4.9849, loss_contra_dense: 4.7901, loss: 9.7750
2021-03-25 09:36:05,476 - openselfsup - INFO - Epoch [3][50/462] lr: 3.000e-01, eta: 1 day, 5:02:47, time: 0.708, data_time: 0.487, memory: 12516, loss_contra_single: 4.9844, loss_contra_dense: 4.8135, loss: 9.7979
2021-03-25 09:36:15,384 - openselfsup - INFO - Epoch [3][100/462] lr: 3.000e-01, eta: 1 day, 4:36:54, time: 0.198, data_time: 0.000, memory: 12516, loss_contra_single: 4.9781, loss_contra_dense: 4.8317, loss: 9.8098
2021-03-25 09:36:25,287 - openselfsup - INFO - Epoch [3][150/462] lr: 3.000e-01, eta: 1 day, 4:13:22, time: 0.198, data_time: 0.000, memory: 12516, loss_contra_single: 4.9746, loss_contra_dense: 4.8472, loss: 9.8218
2021-03-25 09:36:35,163 - openselfsup - INFO - Epoch [3][200/462] lr: 3.000e-01, eta: 1 day, 3:51:45, time: 0.197, data_time: 0.000, memory: 12516, loss_contra_single: 4.9696, loss_contra_dense: 4.8614, loss: 9.8310
.........
2021-03-26 12:21:25,422 - openselfsup - INFO - Epoch [794][50/462] lr: 5.667e-05, eta: 0:13:32, time: 1.546, data_time: 1.022, memory: 12516, loss_contra_single: 3.3514, loss_contra_dense: 3.3958, loss: 6.7473
2021-03-26 12:21:52,574 - openselfsup - INFO - Epoch [794][100/462] lr: 5.667e-05, eta: 0:13:19, time: 0.543, data_time: 0.001, memory: 12516, loss_contra_single: 3.3540, loss_contra_dense: 3.3993, loss: 6.7533
2021-03-26 12:22:19,797 - openselfsup - INFO - Epoch [794][150/462] lr: 5.667e-05, eta: 0:13:07, time: 0.544, data_time: 0.001, memory: 12516, loss_contra_single: 3.3488, loss_contra_dense: 3.3952, loss: 6.7440
2021-03-26 12:22:47,042 - openselfsup - INFO - Epoch [794][200/462] lr: 5.667e-05, eta: 0:12:54, time: 0.545, data_time: 0.002, memory: 12516, loss_contra_single: 3.3503, loss_contra_dense: 3.3952, loss: 6.7456
2021-03-26 12:23:14,157 - openselfsup - INFO - Epoch [794][250/462] lr: 5.667e-05, eta: 0:12:41, time: 0.543, data_time: 0.001, memory: 12516, loss_contra_single: 3.3494, loss_contra_dense: 3.3933, loss: 6.7428
2021-03-26 12:23:41,462 - openselfsup - INFO - Epoch [794][300/462] lr: 5.667e-05, eta: 0:12:29, time: 0.546, data_time: 0.001, memory: 12516, loss_contra_single: 3.3535, loss_contra_dense: 3.3990, loss: 6.7526
2021-03-26 12:24:08,767 - openselfsup - INFO - Epoch [794][350/462] lr: 5.667e-05, eta: 0:12:16, time: 0.546, data_time: 0.001, memory: 12516, loss_contra_single: 3.3507, loss_contra_dense: 3.3964, loss: 6.7471
2021-03-26 12:24:35,944 - openselfsup - INFO - Epoch [794][400/462] lr: 5.667e-05, eta: 0:12:03, time: 0.544, data_time: 0.001, memory: 12516, loss_contra_single: 3.3532, loss_contra_dense: 3.3980, loss: 6.7512
2021-03-26 12:25:01,908 - openselfsup - INFO - Epoch [794][450/462] lr: 5.667e-05, eta: 0:11:51, time: 0.520, data_time: 0.001, memory: 12516, loss_contra_single: 3.3498, loss_contra_dense: 3.3968, loss: 6.7467
2021-03-26 12:26:21,693 - openselfsup - INFO - Epoch [795][50/462] lr: 4.164e-05, eta: 0:11:35, time: 1.507, data_time: 0.929, memory: 12516, loss_contra_single: 3.3507, loss_contra_dense: 3.3954, loss: 6.7461
2021-03-26 12:26:48,843 - openselfsup - INFO - Epoch [795][100/462] lr: 4.164e-05, eta: 0:11:23, time: 0.543, data_time: 0.001, memory: 12516, loss_contra_single: 3.3529, loss_contra_dense: 3.3998, loss: 6.7527
2021-03-26 12:27:16,140 - openselfsup - INFO - Epoch [795][150/462] lr: 4.164e-05, eta: 0:11:10, time: 0.546, data_time: 0.001, memory: 12516, loss_contra_single: 3.3472, loss_contra_dense: 3.3894, loss: 6.7365
2021-03-26 12:27:43,551 - openselfsup - INFO - Epoch [795][200/462] lr: 4.164e-05, eta: 0:10:57, time: 0.548, data_time: 0.001, memory: 12516, loss_contra_single: 3.3509, loss_contra_dense: 3.3920, loss: 6.7429
2021-03-26 12:28:10,263 - openselfsup - INFO - Epoch [795][250/462] lr: 4.164e-05, eta: 0:10:45, time: 0.534, data_time: 0.001, memory: 12516, loss_contra_single: 3.3505, loss_contra_dense: 3.3980, loss: 6.7485
2021-03-26 12:28:37,463 - openselfsup - INFO - Epoch [795][300/462] lr: 4.164e-05, eta: 0:10:32, time: 0.544, data_time: 0.002, memory: 12516, loss_contra_single: 3.3510, loss_contra_dense: 3.3978, loss: 6.7489
2021-03-26 12:29:03,851 - openselfsup - INFO - Epoch [795][350/462] lr: 4.164e-05, eta: 0:10:19, time: 0.529, data_time: 0.001, memory: 12516, loss_contra_single: 3.3467, loss_contra_dense: 3.3945, loss: 6.7412
2021-03-26 12:29:15,641 - openselfsup - INFO - Epoch [795][400/462] lr: 4.164e-05, eta: 0:10:06, time: 0.234, data_time: 0.000, memory: 12516, loss_contra_single: 3.3486, loss_contra_dense: 3.3935, loss: 6.7421
2021-03-26 12:29:41,146 - openselfsup - INFO - Epoch [795][450/462] lr: 4.164e-05, eta: 0:09:54, time: 0.511, data_time: 0.003, memory: 12516, loss_contra_single: 3.3532, loss_contra_dense: 3.3981, loss: 6.7513
2021-03-26 12:30:49,824 - openselfsup - INFO - Epoch [796][50/462] lr: 2.891e-05, eta: 0:09:38, time: 1.202, data_time: 0.589, memory: 12516, loss_contra_single: 3.3496, loss_contra_dense: 3.3921, loss: 6.7417
2021-03-26 12:31:17,298 - openselfsup - INFO - Epoch [796][100/462] lr: 2.891e-05, eta: 0:09:25, time: 0.549, data_time: 0.001, memory: 12516, loss_contra_single: 3.3510, loss_contra_dense: 3.3962, loss: 6.7472
2021-03-26 12:31:44,261 - openselfsup - INFO - Epoch [796][150/462] lr: 2.891e-05, eta: 0:09:13, time: 0.540, data_time: 0.002, memory: 12516, loss_contra_single: 3.3518, loss_contra_dense: 3.3961, loss: 6.7479
2021-03-26 12:32:11,650 - openselfsup - INFO - Epoch [796][200/462] lr: 2.891e-05, eta: 0:09:00, time: 0.548, data_time: 0.001, memory: 12516, loss_contra_single: 3.3503, loss_contra_dense: 3.3957, loss: 6.7461
2021-03-26 12:32:38,803 - openselfsup - INFO - Epoch [796][250/462] lr: 2.891e-05, eta: 0:08:47, time: 0.543, data_time: 0.001, memory: 12516, loss_contra_single: 3.3511, loss_contra_dense: 3.3947, loss: 6.7458
2021-03-26 12:32:52,589 - openselfsup - INFO - Epoch [796][300/462] lr: 2.891e-05, eta: 0:08:34, time: 0.277, data_time: 0.001, memory: 12516, loss_contra_single: 3.3544, loss_contra_dense: 3.4028, loss: 6.7572
2021-03-26 12:33:06,576 - openselfsup - INFO - Epoch [796][350/462] lr: 2.891e-05, eta: 0:08:22, time: 0.279, data_time: 0.000, memory: 12516, loss_contra_single: 3.3506, loss_contra_dense: 3.3926, loss: 6.7432
2021-03-26 12:33:33,965 - openselfsup - INFO - Epoch [796][400/462] lr: 2.891e-05, eta: 0:08:09, time: 0.548, data_time: 0.001, memory: 12516, loss_contra_single: 3.3493, loss_contra_dense: 3.3944, loss: 6.7437
2021-03-26 12:34:01,336 - openselfsup - INFO - Epoch [796][450/462] lr: 2.891e-05, eta: 0:07:56, time: 0.547, data_time: 0.001, memory: 12516, loss_contra_single: 3.3507, loss_contra_dense: 3.3954, loss: 6.7460
2021-03-26 12:35:08,833 - openselfsup - INFO - Epoch [797][50/462] lr: 1.851e-05, eta: 0:07:40, time: 1.176, data_time: 0.548, memory: 12516, loss_contra_single: 3.3529, loss_contra_dense: 3.3998, loss: 6.7527
2021-03-26 12:35:36,459 - openselfsup - INFO - Epoch [797][100/462] lr: 1.851e-05, eta: 0:07:28, time: 0.552, data_time: 0.001, memory: 12516, loss_contra_single: 3.3521, loss_contra_dense: 3.3962, loss: 6.7483
2021-03-26 12:36:03,984 - openselfsup - INFO - Epoch [797][150/462] lr: 1.851e-05, eta: 0:07:15, time: 0.551, data_time: 0.001, memory: 12516, loss_contra_single: 3.3460, loss_contra_dense: 3.3921, loss: 6.7380
2021-03-26 12:36:28,534 - openselfsup - INFO - Epoch [797][200/462] lr: 1.851e-05, eta: 0:07:02, time: 0.492, data_time: 0.001, memory: 12516, loss_contra_single: 3.3482, loss_contra_dense: 3.3928, loss: 6.7410
2021-03-26 12:36:43,668 - openselfsup - INFO - Epoch [797][250/462] lr: 1.851e-05, eta: 0:06:49, time: 0.300, data_time: 0.001, memory: 12516, loss_contra_single: 3.3477, loss_contra_dense: 3.3957, loss: 6.7434
2021-03-26 12:37:10,218 - openselfsup - INFO - Epoch [797][300/462] lr: 1.851e-05, eta: 0:06:37, time: 0.532, data_time: 0.003, memory: 12516, loss_contra_single: 3.3495, loss_contra_dense: 3.3969, loss: 6.7464
2021-03-26 12:37:37,531 - openselfsup - INFO - Epoch [797][350/462] lr: 1.851e-05, eta: 0:06:24, time: 0.546, data_time: 0.001, memory: 12516, loss_contra_single: 3.3499, loss_contra_dense: 3.3961, loss: 6.7461
2021-03-26 12:38:05,004 - openselfsup - INFO - Epoch [797][400/462] lr: 1.851e-05, eta: 0:06:11, time: 0.549, data_time: 0.001, memory: 12516, loss_contra_single: 3.3506, loss_contra_dense: 3.3961, loss: 6.7467
2021-03-26 12:38:32,088 - openselfsup - INFO - Epoch [797][450/462] lr: 1.851e-05, eta: 0:05:58, time: 0.542, data_time: 0.001, memory: 12516, loss_contra_single: 3.3519, loss_contra_dense: 3.3956, loss: 6.7475
2021-03-26 12:39:39,993 - openselfsup - INFO - Epoch [798][50/462] lr: 1.041e-05, eta: 0:05:43, time: 1.184, data_time: 0.589, memory: 12516, loss_contra_single: 3.3476, loss_contra_dense: 3.3926, loss: 6.7402
2021-03-26 12:40:06,950 - openselfsup - INFO - Epoch [798][100/462] lr: 1.041e-05, eta: 0:05:30, time: 0.539, data_time: 0.001, memory: 12516, loss_contra_single: 3.3554, loss_contra_dense: 3.4029, loss: 6.7583
2021-03-26 12:40:18,646 - openselfsup - INFO - Epoch [798][150/462] lr: 1.041e-05, eta: 0:05:17, time: 0.235, data_time: 0.001, memory: 12516, loss_contra_single: 3.3486, loss_contra_dense: 3.3930, loss: 6.7416
2021-03-26 12:40:39,766 - openselfsup - INFO - Epoch [798][200/462] lr: 1.041e-05, eta: 0:05:04, time: 0.422, data_time: 0.000, memory: 12516, loss_contra_single: 3.3549, loss_contra_dense: 3.3996, loss: 6.7545
2021-03-26 12:41:07,195 - openselfsup - INFO - Epoch [798][250/462] lr: 1.041e-05, eta: 0:04:51, time: 0.549, data_time: 0.001, memory: 12516, loss_contra_single: 3.3471, loss_contra_dense: 3.3931, loss: 6.7402
2021-03-26 12:41:34,295 - openselfsup - INFO - Epoch [798][300/462] lr: 1.041e-05, eta: 0:04:38, time: 0.542, data_time: 0.001, memory: 12516, loss_contra_single: 3.3496, loss_contra_dense: 3.3983, loss: 6.7478
2021-03-26 12:42:01,648 - openselfsup - INFO - Epoch [798][350/462] lr: 1.041e-05, eta: 0:04:26, time: 0.547, data_time: 0.001, memory: 12516, loss_contra_single: 3.3475, loss_contra_dense: 3.3934, loss: 6.7410
2021-03-26 12:42:28,720 - openselfsup - INFO - Epoch [798][400/462] lr: 1.041e-05, eta: 0:04:13, time: 0.542, data_time: 0.001, memory: 12516, loss_contra_single: 3.3493, loss_contra_dense: 3.3954, loss: 6.7447
2021-03-26 12:42:56,082 - openselfsup - INFO - Epoch [798][450/462] lr: 1.041e-05, eta: 0:04:00, time: 0.547, data_time: 0.001, memory: 12516, loss_contra_single: 3.3506, loss_contra_dense: 3.3960, loss: 6.7466
2021-03-26 12:43:56,636 - openselfsup - INFO - Epoch [799][50/462] lr: 4.626e-06, eta: 0:03:44, time: 1.040, data_time: 0.621, memory: 12516, loss_contra_single: 3.3519, loss_contra_dense: 3.3987, loss: 6.7506
2021-03-26 12:44:09,563 - openselfsup - INFO - Epoch [799][100/462] lr: 4.626e-06, eta: 0:03:31, time: 0.259, data_time: 0.001, memory: 12516, loss_contra_single: 3.3519, loss_contra_dense: 3.3982, loss: 6.7501
2021-03-26 12:44:36,807 - openselfsup - INFO - Epoch [799][150/462] lr: 4.626e-06, eta: 0:03:19, time: 0.544, data_time: 0.001, memory: 12516, loss_contra_single: 3.3463, loss_contra_dense: 3.3912, loss: 6.7374
2021-03-26 12:45:04,153 - openselfsup - INFO - Epoch [799][200/462] lr: 4.626e-06, eta: 0:03:06, time: 0.547, data_time: 0.001, memory: 12516, loss_contra_single: 3.3494, loss_contra_dense: 3.3964, loss: 6.7458
2021-03-26 12:45:31,267 - openselfsup - INFO - Epoch [799][250/462] lr: 4.626e-06, eta: 0:02:53, time: 0.542, data_time: 0.001, memory: 12516, loss_contra_single: 3.3496, loss_contra_dense: 3.3959, loss: 6.7455
2021-03-26 12:45:58,880 - openselfsup - INFO - Epoch [799][300/462] lr: 4.626e-06, eta: 0:02:40, time: 0.552, data_time: 0.002, memory: 12516, loss_contra_single: 3.3501, loss_contra_dense: 3.3959, loss: 6.7460
2021-03-26 12:46:26,155 - openselfsup - INFO - Epoch [799][350/462] lr: 4.626e-06, eta: 0:02:27, time: 0.545, data_time: 0.001, memory: 12516, loss_contra_single: 3.3497, loss_contra_dense: 3.3944, loss: 6.7441
2021-03-26 12:46:53,468 - openselfsup - INFO - Epoch [799][400/462] lr: 4.626e-06, eta: 0:02:14, time: 0.546, data_time: 0.001, memory: 12516, loss_contra_single: 3.3532, loss_contra_dense: 3.3977, loss: 6.7509
2021-03-26 12:47:20,703 - openselfsup - INFO - Epoch [799][450/462] lr: 4.626e-06, eta: 0:02:01, time: 0.545, data_time: 0.001, memory: 12516, loss_contra_single: 3.3537, loss_contra_dense: 3.3994, loss: 6.7531
2021-03-26 12:48:48,682 - openselfsup - INFO - Epoch [800][50/462] lr: 1.157e-06, eta: 0:01:46, time: 1.582, data_time: 0.965, memory: 12516, loss_contra_single: 3.3498, loss_contra_dense: 3.3974, loss: 6.7472
2021-03-26 12:49:16,110 - openselfsup - INFO - Epoch [800][100/462] lr: 1.157e-06, eta: 0:01:33, time: 0.549, data_time: 0.001, memory: 12516, loss_contra_single: 3.3481, loss_contra_dense: 3.3978, loss: 6.7460
2021-03-26 12:49:42,993 - openselfsup - INFO - Epoch [800][150/462] lr: 1.157e-06, eta: 0:01:20, time: 0.538, data_time: 0.001, memory: 12516, loss_contra_single: 3.3473, loss_contra_dense: 3.3921, loss: 6.7394
2021-03-26 12:50:10,210 - openselfsup - INFO - Epoch [800][200/462] lr: 1.157e-06, eta: 0:01:07, time: 0.544, data_time: 0.001, memory: 12516, loss_contra_single: 3.3516, loss_contra_dense: 3.3967, loss: 6.7482
2021-03-26 12:50:37,746 - openselfsup - INFO - Epoch [800][250/462] lr: 1.157e-06, eta: 0:00:54, time: 0.551, data_time: 0.001, memory: 12516, loss_contra_single: 3.3483, loss_contra_dense: 3.3919, loss: 6.7402
2021-03-26 12:51:05,142 - openselfsup - INFO - Epoch [800][300/462] lr: 1.157e-06, eta: 0:00:41, time: 0.548, data_time: 0.001, memory: 12516, loss_contra_single: 3.3506, loss_contra_dense: 3.3970, loss: 6.7477
2021-03-26 12:51:32,655 - openselfsup - INFO - Epoch [800][350/462] lr: 1.157e-06, eta: 0:00:28, time: 0.550, data_time: 0.001, memory: 12516, loss_contra_single: 3.3494, loss_contra_dense: 3.3969, loss: 6.7463
2021-03-26 12:51:59,865 - openselfsup - INFO - Epoch [800][400/462] lr: 1.157e-06, eta: 0:00:15, time: 0.544, data_time: 0.001, memory: 12516, loss_contra_single: 3.3500, loss_contra_dense: 3.3958, loss: 6.7458
2021-03-26 12:52:25,972 - openselfsup - INFO - Epoch [800][450/462] lr: 1.157e-06, eta: 0:00:03, time: 0.523, data_time: 0.001, memory: 12516, loss_contra_single: 3.3513, loss_contra_dense: 3.3977, loss: 6.7491
2021-03-26 12:52:29,731 - openselfsup - INFO - Saving checkpoint at 800 epochs

@zzzzzz0407
Copy link

@jancylee Could you please upload the model that you pretrained to google cloud / baidu cloud? We will check it for you.

@jancylee
Copy link
Author

jancylee commented Apr 1, 2021

链接:https://pan.baidu.com/s/1tUUzN7UPPKfOoSKHhIC1Bw
提取码:5zhc

@WXinlong
Copy link
Owner

WXinlong commented Apr 1, 2021

@jancylee Have you extracted the backbone weights using tools/extract_backbone_weights.py before fine-tuning object detection?
If so, please upload the converted model.

@jancylee
Copy link
Author

jancylee commented Apr 1, 2021

@zzzzzz0407
Copy link

zzzzzz0407 commented Apr 6, 2021

https://pan.baidu.com/s/1M8DOOsc3Yg_loJPxu-75Mw
4zj3

@jancylee We have tested the model that you train and there is no problem.

image

Please make sure you have followed the instructions in the readme:
You have to 1) extract the backbone weights (https://github.com/WXinlong/DenseCL#extracting-backbone-weights
) and 2) convert it to detectron2 format (https://github.com/WXinlong/DenseCL/blob/main/benchmarks/detection/README.md).

@jancylee
Copy link
Author

jancylee commented Apr 8, 2021

Thank you very much. It was the "extract the backbone weights" worked.

@jancylee jancylee closed this as completed Apr 8, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants