Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Direct communication between Client and LLM Inference Engine #143

Closed
2 tasks
dan-homebrew opened this issue Sep 11, 2023 · 0 comments
Closed
2 tasks

Direct communication between Client and LLM Inference Engine #143

dan-homebrew opened this issue Sep 11, 2023 · 0 comments
Assignees
Labels
P0: critical Mission critical type: bug Something isn't working

Comments

@dan-homebrew
Copy link
Contributor

dan-homebrew commented Sep 11, 2023

Deliverable

To deliver by EOD Tuesday

  • Revert to initial architecture (Client and LLM communicate directly, SSEs streamed directly to Client)
  • Simple AuthZ for LLM Inference Server

Problem

Private Zenhub Image

Private Zenhub Image

@dan-homebrew dan-homebrew changed the title Streamline LLM-Client Interactions Direct communication between Client and LLM Inference Engine Sep 11, 2023
@dan-homebrew dan-homebrew added the P0: critical Mission critical label Sep 11, 2023
@0xSage 0xSage added the type: bug Something isn't working label Sep 12, 2023
dan-homebrew added a commit that referenced this issue Sep 12, 2023
…een-client-llm

fix: #143 - Direct communication between client and llm inference service
louis-jan added a commit that referenced this issue Sep 12, 2023
* chore: use OpenAI parser

* chore: access host's services

* chore: take out llm service - GGUF model for the latest llama.cpp support
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
P0: critical Mission critical type: bug Something isn't working
Projects
None yet
Development

No branches or pull requests

4 participants