Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feature pyramids #555

Open
ghost opened this issue Mar 25, 2018 · 5 comments
Open

feature pyramids #555

ghost opened this issue Mar 25, 2018 · 5 comments

Comments

@ghost
Copy link

ghost commented Mar 25, 2018

in yolo3, you use concept from feature pyramid networks, can you explain why 'feature pyramid' likes dssd, retinanet outperform 'feature pyramid' likes ssd for object detection (not segmentation)? thanks!

@pjreddie
Copy link
Owner

Feature pyramids use backwards connections from later layers to merge information with earlier layers for predicting detections at higher resolution. This allows them to use both strong semantic signals from later layers and fine grained spatial features from earlier layers. Pretty cool design!

@ghost
Copy link
Author

ghost commented Mar 26, 2018

i have question, with same clustering method on same dataset, but yolo3's anchors' scales are larger than yolo2's many times, anyway to make them have same scale?

@pjreddie
Copy link
Owner

So YOLOv2 I made some design choice errors, I made the anchor box size be relative to the feature size in the last layer. Since the network was downsampling by 32 this means it was relative to 32 pixels so an anchor of 9x9 was actually 288px x 288px.

In YOLOv3 anchor sizes are actual pixel values. this simplifies a lot of stuff and was only a little bit harder to implement

@xiayq1
Copy link

xiayq1 commented Jan 23, 2019

Sry to bother u. In my training, I set random=0, using VOC data for training. Only use 10000 steps. The YOLO can detect the objects. But if I set random=1, also 10000 steps. The YOLO can't detect anything.
But I am more confused about why using different input size to train the YOLO. What kind of benefits will the different size of input bring to the YOLO?
Like if I use the fixed size input, I think it's much better than using multi dimension size.

@computervisionlearner
Copy link

So YOLOv2 I made some design choice errors, I made the anchor box size be relative to the feature size in the last layer. Since the network was downsampling by 32 this means it was relative to 32 pixels so an anchor of 9x9 was actually 288px x 288px.

In YOLOv3 anchor sizes are actual pixel values. this simplifies a lot of stuff and was only a little bit harder to implement
Hi, I guess yolov3 anchor sizes is relative to cfg file, as 608*608, not original image size, right?
During multi-scale training process (as),

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants