diff --git a/torchvision/csrc/cpu/PSROIPool_cpu.cpp b/torchvision/csrc/cpu/ps_roi_pool_kernel.cpp similarity index 93% rename from torchvision/csrc/cpu/PSROIPool_cpu.cpp rename to torchvision/csrc/cpu/ps_roi_pool_kernel.cpp index c6e0a64cac3..171de9edc6a 100644 --- a/torchvision/csrc/cpu/PSROIPool_cpu.cpp +++ b/torchvision/csrc/cpu/ps_roi_pool_kernel.cpp @@ -1,7 +1,6 @@ -#include -#include -#include -#include +#include "ps_roi_pool_kernel.h" + +namespace { template inline void add(T* address, const T& val) { @@ -9,7 +8,7 @@ inline void add(T* address, const T& val) { } template -void PSROIPoolForward( +void ps_roi_pool_forward_kernel_impl( const T* input, const T spatial_scale, int channels, @@ -79,7 +78,7 @@ void PSROIPoolForward( } template -void PSROIPoolBackward( +void ps_roi_pool_backward_kernel_impl( const T* grad_output, const int* channel_mapping, int num_rois, @@ -143,7 +142,9 @@ void PSROIPoolBackward( } } -std::tuple PSROIPool_forward_cpu( +} // namespace + +std::tuple ps_roi_pool_forward_cpu( const at::Tensor& input, const at::Tensor& rois, double spatial_scale, @@ -157,7 +158,7 @@ std::tuple PSROIPool_forward_cpu( at::TensorArg input_t{input, "input", 1}, rois_t{rois, "rois", 2}; - at::CheckedFrom c = "PSROIPool_forward_cpu"; + at::CheckedFrom c = "ps_roi_pool_forward_cpu"; at::checkAllSameType(c, {input_t, rois_t}); int num_rois = rois.size(0); @@ -182,8 +183,8 @@ std::tuple PSROIPool_forward_cpu( auto input_ = input.contiguous(), rois_ = rois.contiguous(); AT_DISPATCH_FLOATING_TYPES_AND_HALF( - input.scalar_type(), "PSROIPool_forward", [&] { - PSROIPoolForward( + input.scalar_type(), "ps_roi_pool_forward", [&] { + ps_roi_pool_forward_kernel_impl( input_.data_ptr(), spatial_scale, channels, @@ -200,7 +201,7 @@ std::tuple PSROIPool_forward_cpu( return std::make_tuple(output, channel_mapping); } -at::Tensor PSROIPool_backward_cpu( +at::Tensor ps_roi_pool_backward_cpu( const at::Tensor& grad, const at::Tensor& rois, const at::Tensor& channel_mapping, @@ -221,7 +222,7 @@ at::Tensor PSROIPool_backward_cpu( at::TensorArg grad_t{grad, "grad", 1}, rois_t{rois, "rois", 2}, channel_mapping_t{channel_mapping, "channel_mapping", 3}; - at::CheckedFrom c = "PSROIPool_backward_cpu"; + at::CheckedFrom c = "ps_roi_pool_backward_cpu"; at::checkAllSameType(c, {grad_t, rois_t}); auto num_rois = rois.size(0); @@ -237,8 +238,8 @@ at::Tensor PSROIPool_backward_cpu( auto grad_ = grad.contiguous(), rois_ = rois.contiguous(); AT_DISPATCH_FLOATING_TYPES_AND_HALF( - grad.scalar_type(), "PSROIPool_backward", [&] { - PSROIPoolBackward( + grad.scalar_type(), "ps_roi_pool_backward", [&] { + ps_roi_pool_backward_kernel_impl( grad_.data_ptr(), channel_mapping.data_ptr(), num_rois, diff --git a/torchvision/csrc/cpu/ps_roi_pool_kernel.h b/torchvision/csrc/cpu/ps_roi_pool_kernel.h new file mode 100644 index 00000000000..14a4e22681a --- /dev/null +++ b/torchvision/csrc/cpu/ps_roi_pool_kernel.h @@ -0,0 +1,23 @@ +#pragma once + +#include +#include "../macros.h" + +VISION_API std::tuple ps_roi_pool_forward_cpu( + const at::Tensor& input, + const at::Tensor& rois, + double spatial_scale, + int64_t pooled_height, + int64_t pooled_width); + +VISION_API at::Tensor ps_roi_pool_backward_cpu( + const at::Tensor& grad, + const at::Tensor& rois, + const at::Tensor& channel_mapping, + double spatial_scale, + int64_t pooled_height, + int64_t pooled_width, + int64_t batch_size, + int64_t channels, + int64_t height, + int64_t width); diff --git a/torchvision/csrc/cpu/vision_cpu.h b/torchvision/csrc/cpu/vision_cpu.h index 22119b5e292..baf64f89689 100644 --- a/torchvision/csrc/cpu/vision_cpu.h +++ b/torchvision/csrc/cpu/vision_cpu.h @@ -4,25 +4,6 @@ // TODO: Delete this file once all the methods are gone -VISION_API std::tuple PSROIPool_forward_cpu( - const at::Tensor& input, - const at::Tensor& rois, - double spatial_scale, - int64_t pooled_height, - int64_t pooled_width); - -VISION_API at::Tensor PSROIPool_backward_cpu( - const at::Tensor& grad, - const at::Tensor& rois, - const at::Tensor& channel_mapping, - double spatial_scale, - int64_t pooled_height, - int64_t pooled_width, - int64_t batch_size, - int64_t channels, - int64_t height, - int64_t width); - VISION_API at::Tensor ROIAlign_forward_cpu( const at::Tensor& input, const at::Tensor& rois, diff --git a/torchvision/csrc/cuda/PSROIPool_cuda.cu b/torchvision/csrc/cuda/ps_roi_pool_kernel.cu similarity index 93% rename from torchvision/csrc/cuda/PSROIPool_cuda.cu rename to torchvision/csrc/cuda/ps_roi_pool_kernel.cu index ab6a50b009c..aa1c834e059 100644 --- a/torchvision/csrc/cuda/PSROIPool_cuda.cu +++ b/torchvision/csrc/cuda/ps_roi_pool_kernel.cu @@ -1,13 +1,14 @@ -#include -#include #include #include #include #include "cuda_helpers.h" +#include "ps_roi_pool_kernel.h" + +namespace { template -__global__ void PSROIPoolForward( +__global__ void ps_roi_pool_forward_kernel_impl( int nthreads, const T* input, const T spatial_scale, @@ -73,7 +74,7 @@ __global__ void PSROIPoolForward( } template -__global__ void PSROIPoolBackward( +__global__ void ps_roi_pool_backward_kernel_impl( int nthreads, const T* grad_output, const int* channel_mapping, @@ -132,7 +133,9 @@ __global__ void PSROIPoolBackward( } } -std::tuple PSROIPool_forward_cuda( +} // namespace + +std::tuple ps_roi_pool_forward_cuda( const at::Tensor& input, const at::Tensor& rois, double spatial_scale, @@ -146,7 +149,7 @@ std::tuple PSROIPool_forward_cuda( at::TensorArg input_t{input, "input", 1}, rois_t{rois, "rois", 2}; - at::CheckedFrom c = "PSROIPool_forward_cuda"; + at::CheckedFrom c = "ps_roi_pool_forward_cuda"; at::checkAllSameGPU(c, {input_t, rois_t}); at::checkAllSameType(c, {input_t, rois_t}); @@ -183,8 +186,8 @@ std::tuple PSROIPool_forward_cuda( auto input_ = input.contiguous(), rois_ = rois.contiguous(); AT_DISPATCH_FLOATING_TYPES_AND_HALF( - input.scalar_type(), "PSROIPool_forward", [&] { - PSROIPoolForward<<>>( + input.scalar_type(), "ps_roi_pool_forward", [&] { + ps_roi_pool_forward_kernel_impl<<>>( output_size, input_.data_ptr(), spatial_scale, @@ -202,7 +205,7 @@ std::tuple PSROIPool_forward_cuda( return std::make_tuple(output, channel_mapping); } -at::Tensor PSROIPool_backward_cuda( +at::Tensor ps_roi_pool_backward_cuda( const at::Tensor& grad, const at::Tensor& rois, const at::Tensor& channel_mapping, @@ -223,7 +226,7 @@ at::Tensor PSROIPool_backward_cuda( at::TensorArg grad_t{grad, "grad", 1}, rois_t{rois, "rois", 2}, channel_mapping_t{channel_mapping, "channel_mapping", 3}; - at::CheckedFrom c = "PSROIPool_backward_cuda"; + at::CheckedFrom c = "ps_roi_pool_backward_cuda"; at::checkAllSameGPU(c, {grad_t, rois_t, channel_mapping_t}); at::checkAllSameType(c, {grad_t, rois_t}); @@ -251,8 +254,8 @@ at::Tensor PSROIPool_backward_cuda( auto grad_ = grad.contiguous(), rois_ = rois.contiguous(); AT_DISPATCH_FLOATING_TYPES_AND_HALF( - grad.scalar_type(), "PSROIPool_backward", [&] { - PSROIPoolBackward<<>>( + grad.scalar_type(), "ps_roi_pool_backward", [&] { + ps_roi_pool_backward_kernel_impl<<>>( grad.numel(), grad_.data_ptr(), channel_mapping.data_ptr(), diff --git a/torchvision/csrc/cuda/ps_roi_pool_kernel.h b/torchvision/csrc/cuda/ps_roi_pool_kernel.h new file mode 100644 index 00000000000..e97f0ee7065 --- /dev/null +++ b/torchvision/csrc/cuda/ps_roi_pool_kernel.h @@ -0,0 +1,23 @@ +#pragma once + +#include +#include "../macros.h" + +VISION_API std::tuple ps_roi_pool_forward_cuda( + const at::Tensor& input, + const at::Tensor& rois, + double spatial_scale, + int64_t pooled_height, + int64_t pooled_width); + +VISION_API at::Tensor ps_roi_pool_backward_cuda( + const at::Tensor& grad, + const at::Tensor& rois, + const at::Tensor& channel_mapping, + double spatial_scale, + int64_t pooled_height, + int64_t pooled_width, + int64_t batch_size, + int64_t channels, + int64_t height, + int64_t width); diff --git a/torchvision/csrc/cuda/vision_cuda.h b/torchvision/csrc/cuda/vision_cuda.h index c80386a8db1..8d411b9c67e 100644 --- a/torchvision/csrc/cuda/vision_cuda.h +++ b/torchvision/csrc/cuda/vision_cuda.h @@ -4,25 +4,6 @@ // TODO: Delete this file once all the methods are gone -VISION_API std::tuple PSROIPool_forward_cuda( - const at::Tensor& input, - const at::Tensor& rois, - double spatial_scale, - int64_t pooled_height, - int64_t pooled_width); - -VISION_API at::Tensor PSROIPool_backward_cuda( - const at::Tensor& grad, - const at::Tensor& rois, - const at::Tensor& channel_mapping, - double spatial_scale, - int64_t pooled_height, - int64_t pooled_width, - int64_t batch_size, - int64_t channels, - int64_t height, - int64_t width); - VISION_API at::Tensor ROIAlign_forward_cuda( const at::Tensor& input, const at::Tensor& rois, diff --git a/torchvision/csrc/PSROIPool.h b/torchvision/csrc/ps_roi_pool.cpp similarity index 93% rename from torchvision/csrc/PSROIPool.h rename to torchvision/csrc/ps_roi_pool.cpp index c3ced9e7842..76fb2d04be7 100644 --- a/torchvision/csrc/PSROIPool.h +++ b/torchvision/csrc/ps_roi_pool.cpp @@ -1,18 +1,10 @@ -#pragma once +#include "ps_roi_pool.h" +#include -#include "cpu/vision_cpu.h" - -#ifdef WITH_CUDA -#include "autocast.h" -#include "cuda/vision_cuda.h" -#endif -#ifdef WITH_HIP -#include "autocast.h" -#include "hip/vision_cuda.h" +#if defined(WITH_CUDA) || defined(WITH_HIP) +#include #endif -// TODO: put this stuff in torchvision namespace - std::tuple ps_roi_pool( const at::Tensor& input, const at::Tensor& rois, @@ -26,7 +18,7 @@ std::tuple ps_roi_pool( } #if defined(WITH_CUDA) || defined(WITH_HIP) -std::tuple PSROIPool_autocast( +std::tuple ps_roi_pool_autocast( const at::Tensor& input, const at::Tensor& rois, double spatial_scale, @@ -74,6 +66,8 @@ at::Tensor _ps_roi_pool_backward( width); } +namespace { + class PSROIPoolFunction : public torch::autograd::Function { public: static torch::autograd::variable_list forward( @@ -166,7 +160,9 @@ class PSROIPoolBackwardFunction } }; -std::tuple PSROIPool_autograd( +} // namespace + +std::tuple ps_roi_pool_autograd( const at::Tensor& input, const at::Tensor& rois, double spatial_scale, @@ -178,7 +174,7 @@ std::tuple PSROIPool_autograd( return std::make_tuple(result[0], result[1]); } -at::Tensor PSROIPool_backward_autograd( +at::Tensor ps_roi_pool_backward_autograd( const at::Tensor& grad, const at::Tensor& rois, const at::Tensor& channel_mapping, diff --git a/torchvision/csrc/ps_roi_pool.h b/torchvision/csrc/ps_roi_pool.h new file mode 100644 index 00000000000..0c8baef4a9a --- /dev/null +++ b/torchvision/csrc/ps_roi_pool.h @@ -0,0 +1,61 @@ +#pragma once + +#include "cpu/ps_roi_pool_kernel.h" + +#ifdef WITH_CUDA +#include "cuda/ps_roi_pool_kernel.h" +#endif +#ifdef WITH_HIP +#include "hip/ps_roi_pool_kernel.h" +#endif + +// C++ Forward +std::tuple ps_roi_pool( + const at::Tensor& input, + const at::Tensor& rois, + double spatial_scale, + int64_t pooled_height, + int64_t pooled_width); + +// Autocast Forward +#if defined(WITH_CUDA) || defined(WITH_HIP) +std::tuple ps_roi_pool_autocast( + const at::Tensor& input, + const at::Tensor& rois, + double spatial_scale, + int64_t pooled_height, + int64_t pooled_width); +#endif + +// C++ Backward +at::Tensor _ps_roi_pool_backward( + const at::Tensor& grad, + const at::Tensor& rois, + const at::Tensor& channel_mapping, + double spatial_scale, + int64_t pooled_height, + int64_t pooled_width, + int64_t batch_size, + int64_t channels, + int64_t height, + int64_t width); + +// Autograd Forward and Backward +std::tuple ps_roi_pool_autograd( + const at::Tensor& input, + const at::Tensor& rois, + double spatial_scale, + int64_t pooled_height, + int64_t pooled_width); + +at::Tensor ps_roi_pool_backward_autograd( + const at::Tensor& grad, + const at::Tensor& rois, + const at::Tensor& channel_mapping, + double spatial_scale, + int64_t pooled_height, + int64_t pooled_width, + int64_t batch_size, + int64_t channels, + int64_t height, + int64_t width); diff --git a/torchvision/csrc/vision.cpp b/torchvision/csrc/vision.cpp index c5c204aac2b..6f540c6832e 100644 --- a/torchvision/csrc/vision.cpp +++ b/torchvision/csrc/vision.cpp @@ -8,13 +8,13 @@ #include #endif -#include "PSROIPool.h" #include "ROIAlign.h" #include "ROIPool.h" #include "deform_conv2d.h" #include "empty_tensor_op.h" #include "nms.h" #include "ps_roi_align.h" +#include "ps_roi_pool.h" // If we are in a Windows environment, we need to define // initialization functions for the _custom_ops extension @@ -67,8 +67,8 @@ TORCH_LIBRARY_IMPL(torchvision, CPU, m) { m.impl("nms", nms_cpu); m.impl("ps_roi_align", ps_roi_align_forward_cpu); m.impl("_ps_roi_align_backward", ps_roi_align_backward_cpu); - m.impl("ps_roi_pool", PSROIPool_forward_cpu); - m.impl("_ps_roi_pool_backward", PSROIPool_backward_cpu); + m.impl("ps_roi_pool", ps_roi_pool_forward_cpu); + m.impl("_ps_roi_pool_backward", ps_roi_pool_backward_cpu); m.impl("roi_align", ROIAlign_forward_cpu); m.impl("_roi_align_backward", ROIAlign_backward_cpu); m.impl("roi_pool", ROIPool_forward_cpu); @@ -83,8 +83,8 @@ TORCH_LIBRARY_IMPL(torchvision, CUDA, m) { m.impl("nms", nms_cuda); m.impl("ps_roi_align", ps_roi_align_forward_cuda); m.impl("_ps_roi_align_backward", ps_roi_align_backward_cuda); - m.impl("ps_roi_pool", PSROIPool_forward_cuda); - m.impl("_ps_roi_pool_backward", PSROIPool_backward_cuda); + m.impl("ps_roi_pool", ps_roi_pool_forward_cuda); + m.impl("_ps_roi_pool_backward", ps_roi_pool_backward_cuda); m.impl("roi_align", ROIAlign_forward_cuda); m.impl("_roi_align_backward", ROIAlign_backward_cuda); m.impl("roi_pool", ROIPool_forward_cuda); @@ -98,7 +98,7 @@ TORCH_LIBRARY_IMPL(torchvision, Autocast, m) { m.impl("deform_conv2d", deform_conv2d_autocast); m.impl("nms", nms_autocast); m.impl("ps_roi_align", ps_roi_align_autocast); - m.impl("ps_roi_pool", PSROIPool_autocast); + m.impl("ps_roi_pool", ps_roi_pool_autocast); m.impl("roi_align", ROIAlign_autocast); m.impl("roi_pool", ROIPool_autocast); } @@ -109,8 +109,8 @@ TORCH_LIBRARY_IMPL(torchvision, Autograd, m) { m.impl("_deform_conv2d_backward", deform_conv2d_backward_autograd); m.impl("ps_roi_align", ps_roi_align_autograd); m.impl("_ps_roi_align_backward", ps_roi_align_backward_autograd); - m.impl("ps_roi_pool", PSROIPool_autograd); - m.impl("_ps_roi_pool_backward", PSROIPool_backward_autograd); + m.impl("ps_roi_pool", ps_roi_pool_autograd); + m.impl("_ps_roi_pool_backward", ps_roi_pool_backward_autograd); m.impl("roi_align", ROIAlign_autograd); m.impl("_roi_align_backward", ROIAlign_backward_autograd); m.impl("roi_pool", ROIPool_autograd);