-
Notifications
You must be signed in to change notification settings - Fork 33
Error while deploying simple nifi cluster. #49
Comments
Same issue here. |
hello! |
Upon setting networkPolicy.enabled=false for Zookeeper, issue got resolved for me.
From: Jose Ramon Mañes ***@***.***>
Reply to: Orange-OpenSource/nifikop ***@***.***>
Date: Thursday, 5 August 2021 at 2:00 PM
To: Orange-OpenSource/nifikop ***@***.***>
Cc: Vipul Vipul Mahajan ***@***.***>, Author ***@***.***>
Subject: [EXTERNAL] Re: [Orange-OpenSource/nifikop] Error while deploying simple nifi cluster. (#49)
hello! same issue here, could you solve it? I have zk and operator pods running but the nifi pods are in init state.. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub, or ZjQcmQRYFpfptBannerStart
This Message Is From an External Sender
This message came from outside your organization.
ZjQcmQRYFpfptBannerEnd
hello!
same issue here, could you solve it?
I have zk and operator pods running but the nifi pods are in init state..
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub<#49 (comment)>, or unsubscribe<https://github.com/notifications/unsubscribe-auth/AR5BNQTSJDOH4BQCR57KME3T3JD2JANCNFSM4UERWY7Q>.
Triage notifications on the go with GitHub Mobile for iOS<https://apps.apple.com/app/apple-store/id1477376905?ct=notification-email&mt=8&pt=524675> or Android<https://play.google.com/store/apps/details?id=com.github.android&utm_campaign=notification-email>.
|
I just fixed it, I change the zkAddress in the NifiCluster manifest |
Hi Error: Logs:
Cluster state:
My configuration:
|
Hi All I am facing the same issue : in secured nifi clustered if I remove the OIDC config with OIDC the cluster is running Fine Really appreciate it if some could help out here. LOGS:
Nificluster state:
MY CONFIGURATION:
|
Hello @Sreenivas-Ratakonda In a first check, I can see that your
In the logs, you can see that you're trying to reach: Instead of: Try to modify this one and deploy it again Let me know how is after this change ;) |
Check this comment: #49 (comment) I think is the same issue, the var reference is not the right one |
@jrmanes below is the aws loadbalancer config, I am mapping aws load balancer with headless service (sslnifi-headless) created by nifikop
the problem here is nifikop is not able to verify the cluster state: This the logs I am getting :
Below is the cluster status:
As you can see in cluster all nodes are connected but if we describe the nificluster status: status portion of the
As per nifikop one node is still in connecting status because of that nifikop says nificluster is not ready yet to take on any actions like creating users groups even we cant perform any upscaling or downscaling. the #49 issue is almost similar one so i have mentioned issue here, @jrmanes you mentioned that by changing Zkaddress it was working for what changes you have done I have mentioned My config below, this time i am not creating any aws loadbalancer.
the managed users and groups are not getting created as well. Nifikop is saying nificluster is not reachable becase nifi node which is elected to be the cluster cordinator is node1 and nifikop is not able to reach that node thats why nifikop doesnt know about the actual cluster status, as per nifikop node 1 is not ready so its assuming that the cluster is also not ready. I think there some issue in the nifikop when we run it without OIDC, its not able to reach the cluster cordinator. |
Hello @Sreenivas-Ratakonda Regarding the issue, I'll recommend you to do some checks before in order to validate the issue:
Once you have verify that the zk cluster is running properly, I'd recommend you to verify that you can resolve the dns reference from the pods. I'll check the code from my side to check if I got the same issue. @erdrix is this issue familiar to you? |
@jrmanes I have checked the zookeeper status, zookeeper cluster is working fine weather the network policy true or false the cluster is working fine for me i am doing this poc in aws.
I have tried with nifikop version 0.7.0 and 0.7.5 in both version the result is same if I exclude the OIDC the nifikop is getting disengaged from the cluster and it says the cluster is not ready. When I checked the nifikop logs below mentioned error is keep on coming up:
Mean while when i describe the nificluster its still says : connecting.
@erdrix just wanted to check with you if the nifikop only works with OIDC, your response would be really appreciated. @jrmanes can you please try to replicate same scenario may be you might get different results. |
Type of question
Getting error while deploying simple nifi cluster.
Unable to resolve simplenifi-headless serviec within DNS. Below message displayed in Operator logs.
"error":"Get "http://simplenifi-headless.nifi.svc.cluster.local:8080/nifi-api/controller/cluster\": dial tcp: lookup simplenifi-headless.nifi.svc.cluster.local on 10.96.0.10:53: no such host",
Question
What did you do?
Executed below steps:
kubectl apply -f https://raw.githubusercontent.com/Orange-OpenSource/nifikop/master/deploy/crds/v1/nifi.orange.com_nificlusters_crd.yaml
kubectl apply -f https://raw.githubusercontent.com/Orange-OpenSource/nifikop/master/deploy/crds/v1/nifi.orange.com_nifiusers_crd.yaml
kubectl apply -f https://raw.githubusercontent.com/Orange-OpenSource/nifikop/master/deploy/crds/v1/nifi.orange.com_nifiusergroups_crd.yaml
kubectl apply -f https://raw.githubusercontent.com/Orange-OpenSource/nifikop/master/deploy/crds/v1/nifi.orange.com_nifidataflows_crd.yaml
kubectl apply -f https://raw.githubusercontent.com/Orange-OpenSource/nifikop/master/deploy/crds/v1/nifi.orange.com_nifiparametercontexts_crd.yaml
kubectl apply -f https://raw.githubusercontent.com/Orange-OpenSource/nifikop/master/deploy/crds/v1/nifi.orange.com_nifiregistryclients_crd.yaml
helm install nifikop
orange-incubator/nifikop
--namespace=nifi
--set namespaces={"nifi"}
--set image.tag=v0.4.1-alpha-release
Edit below properties:
spec.zkAddress: "nifikop-zk-zookeeper:2181"
spec.nodeConfigGroups.default_group.erviceAccountName: "nifikop"
spec.nodeConfigGroups.default_group.storageConfigs[].pvcSpec.storageClassName: "nfs-client"
Executed the deployment:
kubectl create -n nifi -f config/samples/simplenificluster.yaml
What did you expect to see?
Expecting two clusters nodes to be in running state as part of simplenifi deployment.
What did you see instead? Under which circumstances?
The simplenifi pod is in init state, there is no error reported in pod description.
Inspecting the operator logs, observing below error logs:
"error":"Get "http://simplenifi-headless.nifi.svc.cluster.local:8080/nifi-api/controller/cluster\": dial tcp: lookup simplenifi-headless.nifi.svc.cluster.local on 10.96.0.10:53: no such host",
Here is cluster status:
(base) ~/config/ams kubectl -n nifi get all
NAME READY STATUS RESTARTS AGE
pod/nifikop-68646cd785-nxhkm 1/1 Running 0 17m
pod/nifikop-zk-zookeeper-0 1/1 Running 30 16d
pod/nifikop-zk-zookeeper-1 1/1 Running 30 16d
pod/nifikop-zk-zookeeper-2 1/1 Running 30 16d
pod/simplenifi-1-nodemn8fg 0/1 Init:0/1 0 6m11s
pod/simplenifi-2-node78nxj 0/1 Init:0/1 0 6m11s
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
service/nifikop-zk-zookeeper ClusterIP 10.98.53.129 2181/TCP,2888/TCP,3888/TCP 16d
service/nifikop-zk-zookeeper-headless ClusterIP None 2181/TCP,2888/TCP,3888/TCP 16d
service/simplenifi LoadBalancer 10.96.189.36 8080:32413/TCP,6007:32125/TCP,10000:32286/TCP 6m12s
service/simplenifi-headless ClusterIP None 8080/TCP,6007/TCP,10000/TCP 6m12s
NAME READY UP-TO-DATE AVAILABLE AGE
deployment.apps/nifikop 1/1 1 1 17m
NAME DESIRED CURRENT READY AGE
replicaset.apps/nifikop-68646cd785 1 1 1 17m
NAME READY AGE
statefulset.apps/nifikop-zk-zookeeper 3/3 16d
(base) ~/config/ams
Environment
nifikop version:
0.4.1-alpha-release
Kubernetes version information:
(base) ~/config/ams kubectl --kubeconfig version
Client Version: version.Info{Major:"1", Minor:"19", GitVersion:"v1.19.4", GitCommit:"d360454c9bcd1634cf4cc52d1867af5491dc9c5f", GitTreeState:"clean", BuildDate:"2020-11-12T01:09:16Z", GoVersion:"go1.15.4", Compiler:"gc", Platform:"darwin/amd64"}
Server Version: version.Info{Major:"1", Minor:"16", GitVersion:"v1.16.3", GitCommit:"b3cbbae08ec52a7fc73d334838e18d17e8512749", GitTreeState:"clean", BuildDate:"2019-11-13T11:13:49Z", GoVersion:"go1.12.12", Compiler:"gc", Platform:"linux/amd64"}
Kubernetes cluster kind:
On-prem k8s cluster, 3-nodes matter and 3-node worker
NiFi version:
apache/nifi:1.12.1
Additional context
Observing similar behaviour in two separate environments.
The text was updated successfully, but these errors were encountered: