argo workflow spark logs example
Hello-world workflow logs get archived in the azure storage account but spark workflow logs do not.
- kubectl create namespace freia
- kubectl create namespace argo
- kubectl apply -n argo -f https://github.com/argoproj/argo-workflows/releases/download/v3.4.2/install.yaml
- helm repo add spark-operator https://googlecloudplatform.github.io/spark-on-k8s-operator
- helm install my-release spark-operator/spark-operator --namespace spark-operator --set sparkJobNamespace=freia --create-namespace
- kubectl apply -f permissions-argo.yaml
- Create azure blob storage account. Update storage account name and container name in workflow-controller-configmap.yaml.
- kubectl create secret generic my-azure-storage-credentials
--from-literal "account-access-key=$(az storage account keys list -n storageaccountname --query '[0].value' -otsv)" -n freia
Don't forget to update the storage account name here. - kubectl get configmap/workflow-controller-configmap -n argo -o yaml > workflow-controller-configmap.yaml
and add to workflow-controller-configmap.yaml:
data:
artifactRepository: |
archiveLogs: true
azure:
endpoint: https://storageaccountname.blob.core.windows.net
container: containername
accountKeySecret:
name: my-azure-storage-credentials
key: account-access-key
kubectl delete configmap/workflow-controller-configmap -n argo
kubectl apply -f workflow-controller-configmap.yaml
10. restart argo server and workflow-controller: kubectl delete pods -n argo --all
11. kubectl create -f spark-operator-kubernetes-dag.yaml -- does not archive logs in storage account
12. kubectl create -f hello-world.yaml -- does archive logs