Skip to content
This repository was archived by the owner on Sep 10, 2025. It is now read-only.

Conversation

@swolchok
Copy link
Contributor

The native runner incorrectly assumed that the output of the model
must always be float. Now it doesn't.

[ghstack-poisoned]
@swolchok
Copy link
Contributor Author

swolchok commented Sep 24, 2024

@pytorch-bot
Copy link

pytorch-bot bot commented Sep 24, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/torchchat/1198

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures

As of commit fce0c79 with merge base f343b43 (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

.to(torch::dtype(torch::kFloat32))
.to(torch::kCPU);
auto logits = result[0].data_ptr();
memcpy(s->logits, logits, p->vocab_size * sizeof(float));
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This is for AOTI right? This is needed because the rest of the runner assumes float?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

it is simply moved up from below.

@swolchok swolchok merged commit c34efd5 into main Sep 25, 2024
51 checks passed
@swolchok swolchok deleted the gh/swolchok/16/head branch September 25, 2024 00:07
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Labels

CLA Signed This label is managed by the Meta Open Source bot.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants