Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Nvidia Triton model packaging feature #437

Closed
karbyshevds opened this issue Nov 20, 2020 · 0 comments · Fixed by odahu/odahu-packager#36
Closed

Nvidia Triton model packaging feature #437

karbyshevds opened this issue Nov 20, 2020 · 0 comments · Fixed by odahu/odahu-packager#36
Assignees
Labels
1.4 feature [Added] for new features. WPM

Comments

@karbyshevds
Copy link
Contributor

As Data Scientist I want to be able to pack (containerize) trained mode to Nvidia Triton or other server and deploy it using ODAHU deployment.
Deployed model should be able to accept tensor, now it can only accept JSON that should be de-serialized before passing data to model.
Triton provides ability to pass raw data to models and such features as micro-batching and GPU sharing so it also may help to save costs.

@karbyshevds karbyshevds added feature [Added] for new features. 1.4 WPM labels Nov 20, 2020
@karbyshevds karbyshevds added this to Backlog in odahu-kanban via automation Nov 20, 2020
@karbyshevds karbyshevds moved this from Backlog to In development in odahu-kanban Nov 20, 2020
odahu-kanban automation moved this from In development to In QA Jan 11, 2021
@vlad-tokarev vlad-tokarev reopened this Feb 8, 2021
odahu-kanban automation moved this from In QA to To Do Feb 8, 2021
@vlad-tokarev vlad-tokarev moved this from To Do to In QA in odahu-kanban Feb 8, 2021
@BPylypenko BPylypenko moved this from In QA to Done in odahu-kanban Feb 10, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
1.4 feature [Added] for new features. WPM
Projects
odahu-kanban
  
Done
Development

Successfully merging a pull request may close this issue.

4 participants