Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is bfloat16 going to be supportted? #11

Open
hxdtest opened this issue Aug 14, 2024 · 1 comment
Open

Is bfloat16 going to be supportted? #11

hxdtest opened this issue Aug 14, 2024 · 1 comment

Comments

@hxdtest
Copy link

hxdtest commented Aug 14, 2024

 // For now, only half_t is supported. TF32 is WIP.
  // Instantiate the function template for different HEADDIMS.
  // For now, only half_t is supported. TF32 is WIP.
  if (kHeadSize == 64) {
    testFmhaForward<cutlass::half_t, 64>(seqLength, seqLength, numHeads,
                                         batchSize, iterations, refCheck,
                                         printValues, nStreams);

Why only half_t is supported?

@hxdtest hxdtest changed the title If bfloat16 going to be supportted? Is bfloat16 going to be supportted? Aug 14, 2024
@jayhshah
Copy link
Collaborator

jayhshah commented Aug 14, 2024

Only for simplicity, you can change the template parameter yourself if you want to. However I recommend using FlashAttention-3 instead now.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants