Skip to content

Conversation

kyujin-cho
Copy link
Contributor

This PR resolves #767. As described on #767 (comment), referencing model_type instance variable results in wrongly determining model variant. Until there's some movement to fix this behavior on huggingface side, this patch will be a simple workaround to get rid of the CUDA error.

@justinphan3110cais
Copy link

I got

    self.hf_config.to_dict().["model_type"] == "falcon"
                             ^
SyntaxError: invalid syntax

@WoosukKwon WoosukKwon self-requested a review September 10, 2023 08:38
Copy link
Collaborator

@WoosukKwon WoosukKwon left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hi @kyujin-cho, thanks for submitting the PR! I've made a small change before merging it.

@justinphan3110cais Thanks for reporting the bug!

@WoosukKwon WoosukKwon merged commit 898285c into vllm-project:main Sep 10, 2023
@kyujin-cho kyujin-cho deleted the fix/falcon-40b-cuda-error branch September 10, 2023 08:39
hongxiayang pushed a commit to hongxiayang/vllm that referenced this pull request Feb 13, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

CUDA error: an illegal memory acces with Falcon 40B
3 participants