Skip to content

How to use CUDA Dynamic Parallelism in PyTorch CPP extension? #44279

@qinjian623

Description

@qinjian623

I found discussions at discuss.pytorch.org . But there is still no solution now.

Here is the error message:

error: kernel launch from __device__ or __global__ functions requires separate compilation mode

and

error: a __device__ function call cannot be configured

Thanks.

cc @ngimel

Metadata

Metadata

Assignees

No one assigned

    Labels

    module: cudaRelated to torch.cuda, and CUDA support in generaltriagedThis issue has been looked at a team member, and triaged and prioritized into an appropriate module

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions