LLMPipeline is a Python package designed to optimize the performance of tasks related to Large Language Models (LLMs). It ensures efficient parallel execution of tasks while maintaining dependency constraints, significantly enhancing the overall performance.
LLMPipeline 是一个 Python 包,旨在优化与大语言模型 (LLM) 相关任务的性能。在满足依赖关系的前提下,确保任务的高效并行执行,从而显著提高整体性能。
-
Dependency Management: Handles task dependencies efficiently, ensuring correct execution order.
依赖管理:高效处理任务依赖关系,确保正确的执行顺序。
-
Parallel Execution: Maximizes parallelism to improve performance.
并行执行:最大化并行性以提高性能。
-
Loop Handling: Supports tasks with loop structures.
循环处理:支持带有循环结构的任务。
-
Easy Integration: Simple and intuitive API for easy integration with existing projects.
易于集成:简单直观的 API,便于与现有项目集成。
You can install LLMPipeline via pip:
你可以通过 pip 安装 LLMPipeline:
pip install llmpipelineHere is a basic example to get you started:
下面是一个基本示例,帮助你快速入门:
from llmpipeline import LLMPipeline, Prompt
# set custom prompt
example_prompt = Prompt("""
...
{inp1}
xxx
""", keys=['{inp1}'])
# set api
def llm_api(inp):
...
return out
def rag_api(inp):
...
return out
# set input data
data = {
'inp': 'test input text ...',
}
# set pipeline
demo_pipe = {
'process_input': {
'prompt': example_prompt,
'format': {'out1': list, 'out2': str}, # check return json format
'inp': ['inp'],
'out': ['out1', 'out2'],
'next': ['rag1', 'loop_A'], # specify the next pipeline
},
'rag1': {
'rag_backend': rag_api2, # specific api can be set for the current pipe via 'rag_backend' or 'llm_backend'.
'inp': ['out2'],
'out': 'out8',
},
'loop_A': { # here is iterating over a list 'out1'
'inp': 'out1',
'pipe_in_loop': ['rag2', 'llm_process', 'rag3', 'rag4', 'llm_process2', 'llm_process3'],
'next': ['exit'], # 'exit' is specific pipe mean to end
},
'rag2': {
'inp': ['out1'],
'out': 'out3',
},
'llm_process2': {
'prompt': llm_process2_prompt,
'format': {'xxx': str, "xxx": str},
'inp': ['inp', 'out4', 'out8'],
'out': 'final_out1',
},
...
}
# running pipeline
pipeline = LLMPipeline(demo_pipe, llm_api, rag_api)
result, info = pipeline.run(data, core_num=4, save_pref=True)Logs are stored in the logs folder. If save_pref is true, you can see the relevant performance report.
日志存储在logs文件夹下,如果save_pref为true,你可以看到相关的性能报告。
For detailed documentation, please visit our official documentation page.
有关详细文档,请访问我们的官方文档页面。
We welcome contributions from the community. Please read our contributing guide to get started.
我们欢迎来自社区的贡献。请阅读我们的贡献指南开始。
LLMPipeline is licensed under the Apache License Version 2.0. See the LICENSE file for more details.
LLMPipeline 采用 Apache License Version 2.0 许可证。有关详细信息,请参阅许可证文件。
Special thanks to all contributors and the open-source community for their support.
特别感谢所有贡献者和开源社区的支持。
For any questions or issues, please open an issue on our GitHub repository.
如有任何问题或意见,请在我们的GitHub 仓库提交 issue。

