Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Spatial Transformer Layer #3114

Open
sontran opened this issue Sep 24, 2015 · 15 comments
Open

Spatial Transformer Layer #3114

sontran opened this issue Sep 24, 2015 · 15 comments

Comments

@sontran
Copy link

sontran commented Sep 24, 2015

This layer seems to help fine grain localization. Link to paper by Max Jaderberg et al
http://arxiv.org/abs/1506.02025
Torch implementation is here
https://github.com/qassemoquab/stnbhwd
Theano/Lasagne implementation/doc is here
https://lasagne.readthedocs.org/en/latest/modules/layers/special.html#lasagne.layers.TransformerLayer

@xygorn
Copy link

xygorn commented Sep 25, 2015

I am working on a initial version of this with an affine transformation and bilinear sampling kernel. In my initial design, I am making a layer for the grid generator and resampler together, and the localization net can be built separately for flexibility.
I will set up a pull request as soon as I port it over to the current master.

@kevinlin311tw
Copy link

I am also interested in Spatial Transformer Layer (SPL). Does it possible to embed SPL in Alexnet?

@ducha-aiki
Copy link
Contributor

@kevinlin311tw
Copy link

@ducha-aiki Thank you so much. I will take a look at his code. Did you try his transformer layer? Because my OS is linux, I cannot simply compile his caffe.

@ducha-aiki
Copy link
Contributor

@kevinlin311tw not yet. I suppose, you could just copy-paste transform_layer.cpp/cu, entry from caffe.proto and from header to your build.

@n3011
Copy link

n3011 commented Nov 23, 2015

@sergeyk are you guys planning to include spatial transformer network with caffe?

@siavashk
Copy link

With @XiaoxiaoGuo implementation, is it possible to perturb the transformation parameters (\theta) in a random manner during training? Similar to how the dropout layer turns some neurons off randomly.

If this is possible, you can generate spatial perturbations of data during the learning phase. This might be interesting for some people including myself.

@futurely
Copy link

Here is another implementation including complete examples. https://github.com/daerduoCarey/SpatialTransformerLayer

@matthieudelaro
Copy link

Here is a ready-to-compile caffe, including the implementation by @daerduoCarey:
https://github.com/matthieudelaro/caffeBVLCplus/tree/stn (stn branch!)
I put files where they belong, modified caffe.proto and filler.hpp, update files to the structural changes of Caffe, un-locked CPU implementation of STN, etc.

@rremani
Copy link

rremani commented Dec 26, 2016

@matthieudelaro I want to build the spatial transform layer with py-faster-rcnn. Can you list me the steps.
Thanks

@shelhamer shelhamer changed the title Implementing spatial transformer layer Spatial Transformer Layer Apr 14, 2017
@whuhxb
Copy link

whuhxb commented Jun 15, 2017

@matthieudelaro Have you successfully added the SpatialTransformerLayer with py-faster-rcnn?

@yanxp
Copy link

yanxp commented Aug 12, 2017

@matthieudelaro @whuhxb Have you successfully added the SpatialTransformerLayer with py-faster-rcnn?

@whuhxb
Copy link

whuhxb commented Aug 15, 2017 via email

@mamunir
Copy link

mamunir commented May 2, 2019

Any resource of spatial transformer network in py faster rcnn is appreciated. Thanks

@yanxp @whuhxb @matthieudelaro

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests