This is an example of how you can build your own Covid-19 End to End Big Data and ML- from ingesting stream to deploying ML model in production leveraging kafka, Apache Spark, Spark mllib and cloud services to build your system and produce machine learning model with big data.
** this doesn't include CLI/Bash/Powershell/yml files for ops.
- Azure account
- Eventhubs
- Azure Databricks with MLFlow
- Azure Machine Learning
- Azure KeyVault
- Kubernetes Environment / Azure Container Instance
7.Cognitive Services - for enriching tweet data with sentiment
- Ingest the data with Kafka on Azure
- Collect, analyze and process the data with Databricks
- Enrich the data - in out case we add sentiment analysis based on tweet text
- Train, evaluate and register machine learning models
- Deploy to production
- Observability and Monitoring
This is a simplified diagram that demonstrate a machine learning life cycle, from development to production.
The main drivers for triggering a new machine learning training process are often based on monitoring and observability layers. Three main triggers are:
- Data driven - we detect new variability of data in our systems
- Scheduled driven - we want to release an updated machine learning model every x days.
- Metrics driven - error detected - highly dependent on the model itself and our ability to detect wrong prdictions/classifications based on the use case
If you have questions/concerns or would like to chat, contact us: