-
Notifications
You must be signed in to change notification settings - Fork 290
update docs #990
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
update docs #990
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Summary of Changes
Hello @hiworldwzj, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed!
This pull request updates the installation documentation for the LightLLM service, specifically focusing on Docker deployment. It provides crucial information regarding shared memory requirements to ensure proper service startup and operation, and adjusts the recommended default shared memory allocation in the provided Docker commands.
Highlights
- Docker Installation Guidance: Enhanced the Docker installation instructions by adding comprehensive notes on shared memory (
--shm-size) requirements for the LightLLM service. - Shared Memory Recommendations: Provided specific shared memory recommendations for both text-only (2GB minimum, 16GB+ recommended) and multimodal (16GB+ recommended) LightLLM deployments within Docker.
- Default Shared Memory Adjustment: Increased the default
--shm-sizeparameter in the documented Docker run commands from1gto2gto better align with the service's needs. - Performance Tuning Tip: Included a tip to adjust the
--running_max_req_sizeparameter for users with limited shared memory, suggesting it can reduce memory footprint at the cost of concurrent requests.
Using Gemini Code Assist
The full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips.
Invoking Gemini
You can request assistance from Gemini at any point in your pull request via creating an issue comment (i.e. comment on the pull request page) using either /gemini <command> or @gemini-code-assist <command>. Below is a summary of the supported commands.
| Feature | Command | Description |
|---|---|---|
| Code Review | /gemini review |
Performs a code review for the current pull request in its current state. |
| Pull Request Summary | /gemini summary |
Provides a summary of the current pull request in its current state. |
| Comment | @gemini-code-assist | Responds in comments when explicitly tagged, both in issue comments and review comments. |
| Help | /gemini help |
Displays a list of available commands. |
Customization
To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a .gemini/ folder in the base of the repository. Detailed instructions can be found here.
Limitations & Feedback
Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments or fill out our survey to provide feedback.
You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension.
Footnotes
-
Review the Privacy Notices, Generative AI Prohibited Use Policy, Terms of Service, and learn how to configure Gemini Code Assist in GitHub here. Gemini can make mistakes, so double check it and use code with caution. ↩
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code Review
此拉取请求更新了文档,以阐明在 Docker 中运行 LightLLM 的共享内存要求,这对用户来说是一个有用的补充。我的审查侧重于提高这些新说明的清晰度和一致性。我发现英文和中文文档的不同部分之间存在一些错字和不一致之处。这些建议旨在纠正这些问题,并确保无论用户如何运行 Docker 镜像,都能获得清晰一致的指导。
| $ # 手动构建镜像 | ||
| $ docker build -t <image_name> . | ||
| $ |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
为了保持一致性,建议在此处为手动构建的镜像添加关于共享内存要求的详细说明,就像为运行官方镜像所做的那样。这将避免选择从源代码构建的用户产生混淆。
$ # 运行服务, 注意现在的lightllm服务非常的依赖共享内存部分,在启动
$ # 前请确保你的docker设置中已经分配了足够的共享内存,否则可能导致
$ # 服务无法正常启动。
$ # 1.如果是纯文本服务,建议分配2GB以上的共享内存, 如果你的内存充足,建议分配16GB以上的共享内存.
$ # 2.如果是多模态服务,建议分配16GB以上的共享内存,具体可以根据实际情况进行调整.
$ # 如果你没有足够的共享内存,可以尝试在启动服务的时候调低 --running_max_req_size 参数,这会降低
$ # 服务的并发请求数量,但可以减少共享内存的占用。
| $ docker pull ghcr.io/modeltc/lightllm:main | ||
| $ | ||
| $ # Run | ||
| $ # Run,The current LightLLM service relies heavily on shared memory. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
| $ # | ||
| $ # 1. For text-only services, it is recommended to allocate more than 2GB of shared memory. | ||
| $ # If your system has sufficient RAM, allocating 16GB or more is recommended. | ||
| $ # 2.For multimodal services, it is recommended to allocate 16GB or more of shared memory. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
| $ docker build -t <image_name> . | ||
| $ | ||
| $ # Run | ||
| $ # Run, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
本节缺少为官方镜像添加的关于共享内存的详细说明。$ # Run, 这一行似乎是不完整的编辑。为了确保一致性并向用户提供完整的信息,也请在此处添加完整的说明。
$ # Run: The current LightLLM service relies heavily on shared memory.
$ # Before starting, please make sure that you have allocated enough shared memory
$ # in your Docker settings; otherwise, the service may fail to start properly.
$ #
$ # 1. For text-only services, it is recommended to allocate more than 2GB of shared memory.
$ # If your system has sufficient RAM, allocating 16GB or more is recommended.
$ # 2. For multimodal services, it is recommended to allocate 16GB or more of shared memory.
$ # You can adjust this value according to your specific requirements.
$ #
$ # If you do not have enough shared memory available, you can try lowering
$ # the --running_max_req_size parameter when starting the service.
$ # This will reduce the number of concurrent requests, but also decrease shared memory usage.
No description provided.