Skip to content

Conversation

@redanrd
Copy link
Contributor

@redanrd redanrd commented Apr 17, 2024

Summary

Improve MLOps example:

  • Make MLOps pipeline automatic after first deployment using CRON schedules to: fetch data from source then train model regularly, and deploy the latest model version from model registry. The schema is useful when the data provided by source grows, which would in general improve inference server performance.

  • Code formatting.

Checklist

  • I have reviewed this myself.
  • I have attached a README to my example. You can use this template as reference.
  • I have updated the project README to link my example.

@redanrd redanrd marked this pull request as ready for review May 7, 2024 13:52
memory_limit = 2048
min_scale = 1
max_scale = 5
max_scale = 1
Copy link
Contributor Author

@redanrd redanrd May 7, 2024

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

We need to preserve the container state as we load the model in memory. Having many instances would possibly lead to ones with no model loaded in memory so they would fail when requested for inference.

@redanrd redanrd requested review from Shillaker and cyclimse May 7, 2024 14:01
@redanrd redanrd merged commit 4eecda3 into main Jun 28, 2024
@redanrd redanrd deleted the mlops-example branch June 28, 2024 08:33
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants