Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LinearInt8 layer for inference of int8-quantized LLMs and Arm intrinsics #5007

Closed
wants to merge 14 commits into from

Conversation

lrw04
Copy link
Contributor

@lrw04 lrw04 commented Sep 5, 2023

Please provide feedback on how the patch could be improved.

@codecov-commenter
Copy link

codecov-commenter commented Sep 5, 2023

Codecov Report

Attention: Patch coverage is 89.47368% with 4 lines in your changes missing coverage. Please review.

Project coverage is 89.52%. Comparing base (bcfec1d) to head (68b4b2d).
Report is 288 commits behind head on master.

Files with missing lines Patch % Lines
src/layer/linearint8.cpp 89.47% 4 Missing ⚠️

❗ There is a different number of reports uploaded between BASE (bcfec1d) and HEAD (68b4b2d). Click for more details.

HEAD has 28 uploads less than BASE
Flag BASE (bcfec1d) HEAD (68b4b2d)
30 2
Additional details and impacted files
@@             Coverage Diff             @@
##           master    #5007       +/-   ##
===========================================
- Coverage   94.72%   89.52%    -5.20%     
===========================================
  Files         772      303      -469     
  Lines      228777    89191   -139586     
===========================================
- Hits       216705    79848   -136857     
+ Misses      12072     9343     -2729     

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

@lrw04
Copy link
Contributor Author

lrw04 commented Sep 7, 2023

oh no, it didn't work

@lrw04
Copy link
Contributor Author

lrw04 commented Sep 7, 2023

Closing following @nihui's instructions.

@lrw04 lrw04 closed this Sep 7, 2023
@lrw04 lrw04 reopened this Sep 7, 2023
@lrw04 lrw04 force-pushed the patch-linearint8 branch 2 times, most recently from ae6ab89 to 307e635 Compare September 9, 2023 02:34
@lrw04 lrw04 changed the title Add LinearInt8 layer for inference of int8-quantized LLMs LinearInt8 layer for inference of int8-quantized LLMs and Arm intrinsics Sep 16, 2023
@nihui nihui closed this Oct 11, 2023
@nihui nihui reopened this Oct 11, 2023
@lrw04 lrw04 closed this by deleting the head repository Aug 22, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants