-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
LISA Finetuning Example #10743
LISA Finetuning Example #10743
Conversation
enable_xetla = self.q_proj.enable_xetla | ||
if self.q_proj.qtype == SYM_INT4 or self.q_proj.qtype == FP8E5: | ||
enable_xetla = self.q_proj.enable_xetla | ||
else: |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Are you sure other dtypes all do not need xetla?
maybe change else
to elif BF16
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Are you sure other dtypes all do not need xetla? maybe change
else
toelif BF16
Here is the supported qtype of xetla.
supported_qtype = self.q_proj.qtype == SYM_INT4 and full_attn
supported_qtype = supported_qtype or self.q_proj.qtype == FP8E5
On XPU, the method returns True when q_proj.qtype is among supported_qtypes(SYM_INT4/FP8E5) and enable_xetla is True.
return device.type == "xpu" and enable_xetla and supported_qtype |
So if qtype is other types, just set enable_xetla to False.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
lgtm
Description
add an example of new fine-tuning algorithm on IPEX-LLM llama2 7b
#nano
'BF16Linear' object has no attribute 'enable_xetla', so 'enable_xetla' works only when q_proj.qtype == SYM_INT4/FP8E5
2. User API changes
How to run
3. Summary of the change
LISA Finetuning Example
4. How to test?