Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat(0.5): update support for OpenLLM 0.5 #22442

Closed
wants to merge 7 commits into from

Conversation

aarnphm
Copy link
Contributor

@aarnphm aarnphm commented Jun 3, 2024

This PR updates the internal wrapper for OpenLLM

Since OpenLLM v0.5 is considered breaking, I have also taken this opportunity to separate the API and local inference component into two separate class:

  • OpenLLMAPI will be responsible for handing remote server going forward (async, sync, streaming supported)
  • OpenLLM will now only for local inference, and only support batching and synchronous generations.

Hopefully this would make it a bit more future-proof. In terms of testing, I believe I have tested through some manual workflow and will be include the tests on our end.

updates from #19894, cc @baskaryan

Signed-off-by: paperspace <29749331+aarnphm@users.noreply.github.com>
Copy link

vercel bot commented Jun 3, 2024

The latest updates on your projects. Learn more about Vercel for Git ↗︎

Name Status Preview Comments Updated (UTC)
langchain ✅ Ready (Inspect) Visit Preview 💬 Add feedback Jun 4, 2024 10:29pm

Signed-off-by: paperspace <29749331+aarnphm@users.noreply.github.com>
Signed-off-by: paperspace <29749331+aarnphm@users.noreply.github.com>
Signed-off-by: paperspace <29749331+aarnphm@users.noreply.github.com>
Signed-off-by: paperspace <29749331+aarnphm@users.noreply.github.com>
@aarnphm aarnphm marked this pull request as ready for review June 3, 2024 20:45
@dosubot dosubot bot added size:XL This PR changes 500-999 lines, ignoring generated files. 🤖:refactor A large refactor of a feature(s) or restructuring of many files labels Jun 3, 2024
Comment on lines -126 to -132
model_name: Optional[str] = None,
*,
model_id: Optional[str] = None,
server_url: Optional[str] = None,
server_url: str,
timeout: int = 30,
server_type: Literal["grpc", "http"] = "http",
embedded: bool = True,
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

shoudl we keep around the old args and raise a warning if they're passed in, so that this isn't a breaking change?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

sure I think we can do that.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

oh, this is for the new class OpenLLMAPI, but for OpenLLM we can keep backward compat.

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

ah missed the rename :+1

@aarnphm
Copy link
Contributor Author

aarnphm commented Jun 4, 2024

@baskaryan since the OpenLLM implementation here won't support _acall or async anymore, it would be a breaking change regardless.

@aarnphm
Copy link
Contributor Author

aarnphm commented Jun 15, 2024

bump @baskaryan when you have bandwidth. sorry for the ping :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
community Related to langchain-community 🤖:refactor A large refactor of a feature(s) or restructuring of many files size:XL This PR changes 500-999 lines, ignoring generated files.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

3 participants