Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

FEAT: concurrent generation #417

Merged
merged 4 commits into from
Sep 13, 2023

Conversation

codingl2k1
Copy link
Contributor

@codingl2k1 codingl2k1 commented Sep 1, 2023

10 concurrent generates on one torch model get ~20% performance boost. (tested on M2 macbook)

@XprobeBot XprobeBot added this to the v0.2.1 milestone Sep 1, 2023
@aresnow1
Copy link
Contributor

aresnow1 commented Sep 1, 2023

We need to add tests to check if it works well for GGML models.

@aresnow1
Copy link
Contributor

aresnow1 commented Sep 4, 2023

Seems it always hang, it may not be caused by new tests, we can try to run all tests locally to find the reasons.

@XprobeBot XprobeBot modified the milestones: v0.2.1, v0.3.1 Sep 5, 2023
@codingl2k1
Copy link
Contributor Author

Seems it always hang, it may not be caused by new tests, we can try to run all tests locally to find the reasons.

Fixed. Thanks.

@codingl2k1 codingl2k1 marked this pull request as ready for review September 5, 2023 03:16
@XprobeBot XprobeBot modified the milestones: v0.4.0, v0.4.2 Sep 12, 2023
@UranusSeven UranusSeven changed the title FEAT: Improve torch inference performance FEAT: concurrent generation Sep 13, 2023
@UranusSeven UranusSeven merged commit f26504e into xorbitsai:main Sep 13, 2023
8 of 10 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

4 participants