Repo to deploy ODF Multi-Cloud Gateway standalone using an argoCD application. This Repo has an additional helm chart to deploy the ODF Logical volume Operator.
If you don't have a running instance of argocd on your OCP cluster you can run the setup.sh script available in the repo, it will setup a basic openshift-gitops instance with the lvmo & mcg application.
- Fork & Clone the repo 'https://github.com/likid0/mcgstandalone', using the gh cli in this example:
$ gh repo fork git@github.com:likid0/mcgstandalone.git --clone
- Make modifications as needed to the argocd deployment.
$ vi argocd/values.yaml
- If you have enabled the deployment of LVMO in the argocd/values.yaml file, Make modifications to the lvmo/values.yaml file
$ vi lvmo/values.yaml
- Make modifications to the MCG configuration as needed.
$ vi mcg/values.yaml
- Push the modifications to your forked github repo.
- run the setup.sh script.
$ bash setup.sh
- Once the setup script as finished you can go into the argocd UI or use argocd cli to start the sync of the argo applications that got created, you first need to sync the LVMO app(it you are using Logical Volume Operator for your storage), and finally sync the ODF MCG app.
$ argocd app sync deploy-lvmo
$ argocd app sync deploy-mcg
- You can check that the phase state is Ready to verify MCG is up and running:
$ kubectl get noobaa noobaa -o jsonpath='{.status.phase}{"\n"}'
- MCG is ready to be used a default user has been created for testing, configure your S3 client:
$ NOOBAA_ACCESS_KEY=$(kubectl get secret noobaa-account-demo -n openshift-storage -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')
$ NOOBAA_SECRET_KEY=$(kubectl get secret noobaa-account-demo -n openshift-storage -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')
$ S3EXTENDPOINT=$(kubectl get route s3 -o jsonpath='{.spec.host}{"\n"}' -n openshift-storage)
$ alias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY aws --endpoint https://$S3EXTENDPOINT --no-verify-ssl s3'
$ s3 ls