-
Notifications
You must be signed in to change notification settings - Fork 216
[LLM Runtime] Add Script for PPL Evaluation #685
Conversation
intel_extension_for_transformers/llm/runtime/graph/application/main_pybind.cpp
Show resolved
Hide resolved
airMeng
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
add a note on homepage readme
intel_extension_for_transformers/llm/runtime/graph/scripts/perplexity.py
Show resolved
Hide resolved
intel_extension_for_transformers/llm/runtime/graph/scripts/perplexity.py
Show resolved
Hide resolved
intel_extension_for_transformers/llm/runtime/graph/application/main_pybind.cpp
Show resolved
Hide resolved
intel_extension_for_transformers/llm/runtime/graph/application/main_pybind.cpp
Outdated
Show resolved
Hide resolved
|
Do we have PPL data of pytorch and that of cpp graph fp32/int4? |
856b6cd to
9e30626
Compare
intel_extension_for_transformers/llm/runtime/graph/application/main_pybind.cpp
Outdated
Show resolved
Hide resolved
|
c29121b to
57943ec
Compare
|
gpt-neox-20b and opt-1.3b checked in this internal extension test: https://inteltf-jenk.sh.intel.com/job/ITREX-cpp-graph-extension-WIP/33/artifact/report.html |
a32543254
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
Signed-off-by: zhenwei-intel <zhenwei.liu@intel.com>
Signed-off-by: zhenwei-intel <zhenwei.liu@intel.com>
Signed-off-by: zhenwei-intel <zhenwei.liu@intel.com>
Signed-off-by: zhenwei-intel <zhenwei.liu@intel.com>
Signed-off-by: VincyZhang <wenxin.zhang@intel.com>
hshen14
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
What's the purpose to introduce ppl script here? Can we just reuse lm-eval-harness to measure the ppl?
It is for internal usage of LLM Runtime to keep track of the inference accuracy with minimal test time addition. It also works as a usage example of the LLM Runtime (llama.cpp also has a ppl script in its |

Type of Change: Feature
API not changed
Description
cal_diff.pyscript for future CI @zhenwei-inteltest_llm_runtime.py)Expected Behavior & Potential Risk
N/A
How has this PR been tested?
Internal extension test:
Dependency Change?
No