Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Helm admin module installation timed out / failed #667

Closed
0xabdi opened this issue May 5, 2021 · 2 comments
Closed

Helm admin module installation timed out / failed #667

0xabdi opened this issue May 5, 2021 · 2 comments

Comments

@0xabdi
Copy link

0xabdi commented May 5, 2021

I have installed MOSIP 1.1.2 on-prem, only accessible over the Internal network/VPN, and using self-signed SSL certs.

When I try to access the pre-reg-ui, I am greeted with the below error:

image

When I run helm1 list and helm2 list to check if all modules were correctly installed, it shows that the admin module's installation failed.

image

I tried deleting the module, re-running it, and starting the admin playbook using these commands:
helm1 delete admin
an playbooks/admin.yml

But it times out by throwing the error: Error: timed out waiting for the condition

I tried increasing the timeout duration from 20 to 120 as suggested here #503 but this only makes it wait longer before it eventually throws the same error.

@0xabdi
Copy link
Author

0xabdi commented May 19, 2021

The docker image version for the admin playbook is incorrect.

To fix this, find the relevant docker images in versions.yml, replace 1.1.3 with 1.1.2.

The above only fixes the admin module failing to install. The pre-registration UI page is still not loading properly and some of the pods are not running properly as shown below:

[mosipuser@console ~]$ kc2 get pods -A
NAMESPACE              NAME                                                      READY   STATUS    RESTARTS   AGE
default                clamav-8bdf76568-r2gx6                                    1/1     Running   0          12d
default                healthcheck-app-6fc6d9b86d-hq49t                          1/1     Running   0          12d
default                nfs-provisioner-nfs-client-provisioner-6bc4c68b58-klcpx   1/1     Running   0          12d
default                reg-client-downloader-6cdd489444-q82bp                    1/1     Running   0          12d
default                regproc-camel-stage-556685f467-nml2q                      2/2     Running   0          12d
default                regproc-dmz-packet-server-55cdc9f9dd-w5kdv                1/1     Running   0          12d
default                regproc-packet-receiver-stage-f995799b7-kh4xs             0/2     Pending   0          12d
default                regproc-registration-status-service-747cbcb85-8g9fj       1/2     Running   377        12d
ingress-nginx          nginx-ingress-controller-54b86f8f7b-d89fr                 1/1     Running   0          12d
kube-system            coredns-f9fd979d6-chkmj                                   1/1     Running   0          12d
kube-system            coredns-f9fd979d6-dn7v5                                   1/1     Running   0          12d
kube-system            etcd-dmzmaster.sb                                         1/1     Running   0          12d
kube-system            kube-apiserver-dmzmaster.sb                               1/1     Running   0          12d
kube-system            kube-controller-manager-dmzmaster.sb                      1/1     Running   0          12d
kube-system            kube-flannel-ds-amd64-8m2nd                               1/1     Running   1          12d
kube-system            kube-flannel-ds-amd64-wdds5                               1/1     Running   0          12d
kube-system            kube-proxy-bx5zs                                          1/1     Running   0          12d
kube-system            kube-proxy-psdr8                                          1/1     Running   0          12d
kube-system            kube-scheduler-dmzmaster.sb                               1/1     Running   0          12d
kubernetes-dashboard   dashboard-metrics-scraper-7b59f7d4df-9fqcf                1/1     Running   0          12d
kubernetes-dashboard   kubernetes-dashboard-77c9766b-q9s6f                       1/1     Running   1          12d
monitoring             filebeat-filebeat-gz5qj                                   0/1     Running   557        12d
monitoring             graf-grafana-75786f6bc8-q9knf                             1/1     Running   0          12d
monitoring             metrics-server-74884b4657-jpgcj                           1/1     Running   1          12d
monitoring             prom-prometheus-alertmanager-9dc6c75c6-qcknk              2/2     Running   0          12d
monitoring             prom-prometheus-node-exporter-njxrf                       1/1     Running   0          12d
monitoring             prom-prometheus-pushgateway-58c98cb767-tcsbk              1/1     Running   0          12d
monitoring             prom-prometheus-server-7668b5c898-2xsvb                   2/2     Running   0          12d
[mosipuser@console ~]$ kc2 describe pod regproc-packet-receiver-stage-f995799b7-kh4xs
Name:           regproc-packet-receiver-stage-f995799b7-kh4xs
Namespace:      default
Priority:       0
Node:           <none>
Labels:         app=regproc-packet-receiver-stage
                component=service-hazelcast-server
                exportLogs=enabled
                pod-template-hash=f995799b7
Annotations:    <none>
Status:         Pending
IP:             
IPs:            <none>
Controlled By:  ReplicaSet/regproc-packet-receiver-stage-f995799b7
Containers:
  regproc-packet-receiver-stage:
    Image:       mosipid/registration-processor-packet-receiver-stage:1.1.2
    Ports:       5711/TCP, 5701/TCP, 8081/TCP
    Host Ports:  0/TCP, 0/TCP, 0/TCP
    Environment:
      active_profile_env:       dmz
      spring_config_label_env:  1.1.2
      spring_config_url_env:    http://mzworker0.sb:30080/config
      artifactory_url_env:      http://artifactory-service/
      is_glowroot_env:          absent
    Mounts:
      /home/logs from applogs (rw)
      /mnt/regproc/landing from landing-zone (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from default-token-hbtxc (ro)
  logger-sidecar:
    Image:      busybox
    Port:       <none>
    Host Port:  <none>
    Args:
      /bin/sh
      -c
      tail -F /home/logs/registrationProcessor.log
    Environment:  <none>
    Mounts:
      /home/logs from applogs (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from default-token-hbtxc (ro)
Conditions:
  Type           Status
  PodScheduled   False 
Volumes:
  landing-zone:
    Type:       PersistentVolumeClaim (a reference to a PersistentVolumeClaim in the same namespace)
    ClaimName:  landing-pvc
    ReadOnly:   false
  applogs:
    Type:       EmptyDir (a temporary directory that shares a pod's lifetime)
    Medium:     
    SizeLimit:  <unset>
  default-token-hbtxc:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  default-token-hbtxc
    Optional:    false
QoS Class:       BestEffort
Node-Selectors:  <none>
Tolerations:     node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
                 node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
  Type     Reason            Age                          From               Message
  ----     ------            ----                         ----               -------
  Warning  FailedScheduling  <invalid> (x17728 over 12d)  default-scheduler  persistentvolumeclaim "landing-pvc" not found
[mosipuser@console ~]$ kc2 describe pod regproc-registration-status-service-747cbcb85-8g9fj
Name:         regproc-registration-status-service-747cbcb85-8g9fj
Namespace:    default
Priority:     0
Node:         dmzworker0/172.29.108.23
Start Time:   Wed, 05 May 2021 14:38:20 +0000
Labels:       app=regproc-registration-status-service
              exportLogs=enabled
              pod-template-hash=747cbcb85
Annotations:  <none>
Status:       Running
IP:           10.244.1.19
IPs:
  IP:           10.244.1.19
Controlled By:  ReplicaSet/regproc-registration-status-service-747cbcb85
Containers:
  regproc-registration-status-service:
    Container ID:   docker://633569863edd3ea9417814eb900b5814466f961d77f9274b922657b9b00844de
    Image:          mosipid/registration-processor-registration-status-service:1.1.2
    Image ID:       docker-pullable://mosipid/registration-processor-registration-status-service@sha256:49d834a399e50af14394f6d9580813ea809d4959ca82cf34dccfd2b2326f4499
    Port:           8083/TCP
    Host Port:      0/TCP
    State:          Running
      Started:      Thu, 06 May 2021 23:47:11 +0000
    Last State:     Terminated
      Reason:       Error
      Exit Code:    1
      Started:      Thu, 06 May 2021 23:42:07 +0000
      Finished:     Thu, 06 May 2021 23:42:07 +0000
    Ready:          False
    Restart Count:  377
    Readiness:      http-get http://:8083/registrationprocessor/v1/registrationstatus/actuator/health delay=30s timeout=1s period=30s #success=1 #failure=3
    Environment:
      active_profile_env:       dmz
      spring_config_label_env:  1.1.2
      spring_config_url_env:    http://mzworker0.sb:30080/config
      artifactory_url_env:      http://artifactory-service/
      is_glowroot_env:          absent
      iam_adapter_url_env:      http://mzworker0.sb:30080/artifactory/libs-snapshot-local/io/mosip/kernel/kernel-auth-adapter/1.2.0/kernel-auth-adapter-1.2.0-20200918.164425-65.jar
    Mounts:
      /home/logs from applogs (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from default-token-hbtxc (ro)
  logger-sidecar:
    Container ID:  docker://e5e2af13b3143266382b4808dc71c6b337582087deaaf619f15251dc2e1d6591
    Image:         busybox
    Image ID:      docker-pullable://busybox@sha256:be4684e4004560b2cd1f12148b7120b0ea69c385bcc9b12a637537a2c60f97fb
    Port:          <none>
    Host Port:     <none>
    Args:
      /bin/sh
      -c
      tail -F /home/logs/registrationProcessor.log
    State:          Running
      Started:      Wed, 05 May 2021 14:38:37 +0000
    Ready:          True
    Restart Count:  0
    Environment:    <none>
    Mounts:
      /home/logs from applogs (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from default-token-hbtxc (ro)
Conditions:
  Type              Status
  Initialized       True 
  Ready             False 
  ContainersReady   False 
  PodScheduled      True 
Volumes:
  applogs:
    Type:       EmptyDir (a temporary directory that shares a pod's lifetime)
    Medium:     
    SizeLimit:  <unset>
  default-token-hbtxc:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  default-token-hbtxc
    Optional:    false
QoS Class:       BestEffort
Node-Selectors:  <none>
Tolerations:     node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
                 node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
  Type     Reason     Age                          From     Message
  ----     ------     ----                         ----     -------
  Warning  Unhealthy  <invalid> (x31802 over 11d)  kubelet  Readiness probe failed: Get "http://10.244.1.19:8083/registrationprocessor/v1/registrationstatus/actuator/health": dial tcp 10.244.1.19:8083: connect: connection refused

@0xabdi 0xabdi closed this as completed May 19, 2021
@0xabdi 0xabdi reopened this May 19, 2021
@0xabdi
Copy link
Author

0xabdi commented May 25, 2021

So, the issue of some of the pods not deploying properly was fixed by resetting the deployment and doing a fresh installation, and then for the pre-registration page error, I realized it requires you to access the page using the domain name of the console machine and not its IP address. In my case, I didn't have control over the network and could not configure the main DNS server on the network to do the mapping but ended up creating a local mapping on my machine (/etc/hosts).

@0xabdi 0xabdi closed this as completed May 25, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant