-
Notifications
You must be signed in to change notification settings - Fork 21.4k
/
fused_kernel.h
59 lines (48 loc) · 1.4 KB
/
fused_kernel.h
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
#pragma once
#include <ATen/ATen.h>
#include <torch/csrc/WindowsTorchApiMacro.h>
#include <torch/csrc/jit/codegen/fuser/fused_kernel.h>
#include <cuda.h>
#include <cuda_runtime.h>
#include <nvrtc.h>
#include <cstdint>
#include <string>
#include <vector>
namespace torch {
namespace jit {
namespace fuser {
namespace cuda {
// A class holding metadata for an actual CUDA function.
// Note: CUDA functions are per device.
struct TORCH_CUDA_API FusedKernelCUDA
: public ::torch::jit::fuser::FusedKernel {
FusedKernelCUDA(
at::DeviceIndex device,
std::string name,
std::string code,
std::vector<TensorDesc> input_desc,
std::vector<TensorDesc> output_desc,
std::vector<PartitionDesc> chunk_desc,
std::vector<PartitionDesc> concat_desc,
bool has_random);
~FusedKernelCUDA() override;
void launch_raw(const uint32_t numel, std::vector<void*>& arguments)
const override;
at::Backend backend() const override {
return at::Backend::CUDA;
}
private:
static constexpr auto kBlockSize = 128;
// Note: per device to store device properties and compute launch heuristics
// Acquiring these values at launch time would be too slow
at::DeviceIndex device_;
int maxBlocks_;
cudaDeviceProp* prop_;
std::vector<char> ptx_;
CUmodule module_;
CUfunction function_;
};
} // namespace cuda
} // namespace fuser
} // namespace jit
} // namespace torch