Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

STORY: As a developer bringing up my own IMAGE server, I want to minimize the resource requirements necessary to run the complete IMAGE server stack, so that I can serve as many clients as possible with the lowest possible cost. #804

Closed
5 tasks
jeffbl opened this issue Feb 7, 2024 · 3 comments · Fixed by #817
Assignees

Comments

@jeffbl
Copy link
Member

jeffbl commented Feb 7, 2024

  • Profile existing server stack to find bottlenecks when serving multiple clients simultaneously
  • Create scripts/tools for load testing server on an ongoing basis
  • Review environments used by ML preprocessors, optimize where necessary
  • Design system for caching requests, e.g., if multiple clients ask for same graphic in short time window
  • Document process and best practices for running an IMAGE server (This might be significant enough to fall into another sprint?)

The goal of this story is to think broadly about where likely bottlenecks will occur, and where we can find the biggest bang for the buck in terms of making it easy and low-cost to run an IMAGE server. We are used to running on pegasus, but that is a fairly expensive machine. By optimizing, the hope is that a viable server could be run on lower-end hardware, or alternatively, scale much higher on hardware like pegasus.

@jeffbl
Copy link
Member Author

jeffbl commented Feb 7, 2024

Note some additional information in older issues, which can probably be closed/folded into this story, but might contain useful information on prior efforts in this direction:
#251
#266
#93
Shared-Reality-Lab/IMAGE-test-graphics#18

@jaydeepsingh25 jaydeepsingh25 linked a pull request May 19, 2024 that will close this issue
9 tasks
@jaydeepsingh25
Copy link
Contributor

Server cache implementation merged and deployed on unicorn

@jaydeepsingh25
Copy link
Contributor

Closing this issue as multiple issues have been created to track different items

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants