Skip to content

Conversation

yzh119
Copy link
Collaborator

@yzh119 yzh119 commented Oct 3, 2025

📌 Description

TVM-FFI made a new release with bugfix on potential memory leak:
apache/tvm-ffi#87
This PR bumps the apache-tvm-ffi version in docker container.

🔍 Related Issues

apache/tvm-ffi#87

🚀 Pull Request Checklist

Thank you for contributing to FlashInfer! Before we review your pull request, please make sure the following items are complete.

✅ Pre-commit Checks

  • I have installed pre-commit by running pip install pre-commit (or used your preferred method).
  • I have installed the hooks with pre-commit install.
  • I have run the hooks manually with pre-commit run --all-files and fixed any reported issues.

If you are unsure about how to set up pre-commit, see the pre-commit documentation.

🧪 Tests

  • Tests have been added or updated as needed.
  • All tests are passing (unittest, etc.).

Reviewer Notes

Copy link
Contributor

Summary of Changes

Hello @yzh119, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed!

This pull request focuses on enhancing the stability of the Docker build environment by updating a critical dependency. The primary goal is to integrate a newer version of apache-tvm-ffi that includes an important bugfix for a potential memory leak, ensuring that the Docker containers benefit from the latest improvements and operate more reliably.

Highlights

  • Dependency Upgrade: The apache-tvm-ffi package has been upgraded from version 0.1.0b11 to 0.1.0b15 within the Docker container build process.
  • Memory Leak Fix: This upgrade incorporates a bugfix in apache-tvm-ffi that addresses a potential memory leak, improving the stability and efficiency of the Docker environment.
Ignored Files
  • Ignored by pattern: .github/workflows/** (1)
    • .github/workflows/release-ci-docker.yml
Using Gemini Code Assist

The full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips.

Invoking Gemini

You can request assistance from Gemini at any point by creating a comment using either /gemini <command> or @gemini-code-assist <command>. Below is a summary of the supported commands on the current page.

Feature Command Description
Code Review /gemini review Performs a code review for the current pull request in its current state.
Pull Request Summary /gemini summary Provides a summary of the current pull request in its current state.
Comment @gemini-code-assist Responds in comments when explicitly tagged, both in pull request comments and review comments.
Help /gemini help Displays a list of available commands.

Customization

To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a .gemini/ folder in the base of the repository. Detailed instructions can be found here.

Limitations & Feedback

Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for Github and other Google products, sign up here.

You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension.

Footnotes

  1. Review the Privacy Notices, Generative AI Prohibited Use Policy, Terms of Service, and learn how to configure Gemini Code Assist in GitHub here. Gemini can make mistakes, so double check it and use code with caution.

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request upgrades the apache-tvm-ffi package to version 0.1.0b15 in the Docker build script to incorporate a memory leak fix. The change is straightforward and correct. I've added a couple of suggestions to combine consecutive pip3 install commands within the script. This will improve efficiency by reducing the number of processes launched and also enhances the script's readability.

Comment on lines +33 to +34
pip3 install --upgrade cuda-python==13.0
pip3 install "nvidia-cudnn-cu13>=9.12.0.46"
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

For better efficiency and readability, you can combine these two pip3 install commands into a single one. This reduces the number of separate processes being invoked.

Suggested change
pip3 install --upgrade cuda-python==13.0
pip3 install "nvidia-cudnn-cu13>=9.12.0.46"
pip3 install --upgrade cuda-python==13.0 "nvidia-cudnn-cu13>=9.12.0.46"

Comment on lines +36 to +37
pip3 install --upgrade cuda-python==12.*
pip3 install "nvidia-cudnn-cu12>=9.11.0.98"
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

Similar to the if branch, you can combine these two pip3 install commands into a single one for better efficiency and readability.

Suggested change
pip3 install --upgrade cuda-python==12.*
pip3 install "nvidia-cudnn-cu12>=9.11.0.98"
pip3 install --upgrade cuda-python==12.* "nvidia-cudnn-cu12>=9.11.0.98"

@yzh119 yzh119 requested a review from cyx-6 October 3, 2025 23:17
@yzh119 yzh119 merged commit ef9dd60 into flashinfer-ai:main Oct 4, 2025
12 checks passed
@yzh119 yzh119 mentioned this pull request Oct 4, 2025
5 tasks
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants