Skip to content
This repository has been archived by the owner on Jun 20, 2024. It is now read-only.

prefer node ip for commnunicating with cilium , rather than DNS #86

Merged
merged 1 commit into from
Feb 2, 2022

Conversation

weizhoublue
Copy link
Contributor

@weizhoublue weizhoublue commented Jan 27, 2022

Signed-off-by: weizhou.lan@daocloud.io weizhou.lan@daocloud.io

prefer node ip for commnunicating with cilium , rather than DNS

when I deploy hubble-otel , I encounter the following issue reported by the hubble otel collector
DNS failed to resolve Node hostname .
It much make sense to use nodeIp for collector communicating directly with cilium-agent for most cases

# kubectl logs -n kube-system -l app.kubernetes.io/name=otelcol-hubble-collector
2022-01-27T13:22:38.295Z	info	v3@v3.2103.1/logger.go:46	Discard stats nextEmptySlot: 0
	{"kind": "receiver", "name": "hubble"}
2022-01-27T13:22:38.295Z	info	v3@v3.2103.1/logger.go:46	Set nextTxnTs to 0	{"kind": "receiver", "name": "hubble"}
2022-01-27T13:22:38.298Z	info	service/telemetry.go:116	Serving Prometheus metrics	{"address": ":8888", "level": "basic", "service.instance.id": "cae4e6c1-bbca-41f6-93b6-7d9e0c310551", "service.version": "latest"}
2022-01-27T13:22:38.298Z	info	service/collector.go:230	Starting otelcol-hubble...	{"Version": "0.1.0", "NumCPU": 40}
2022-01-27T13:22:38.298Z	info	service/collector.go:132	Everything is ready. Begin running and processing data.
2022-01-27T13:22:39.282Z	info	jaegerexporter@v0.38.0/exporter.go:186	State of the connection with the Jaeger Collector backend	{"kind": "exporter", "name": "jaeger", "state": "READY"}
2022-01-27T13:22:42.290Z	error	receiver@v0.0.0-00010101000000-000000000000/receiver.go:94	hubble reciever error	{"kind": "receiver", "name": "hubble", "error": "GetFlows failed: rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing dial tcp: lookup 172-81-0-20 on 172.22.0.10:53: server misbehaving\""}
github.com/cilium/hubble-otel/receiver.(*hubbleReceiver).Start.func1
	github.com/cilium/hubble-otel/receiver@v0.0.0-00010101000000-000000000000/receiver.go:94
2022-01-27T13:22:38.600Z	info	v3@v3.2103.1/logger.go:46	Discard stats nextEmptySlot: 0
	{"kind": "receiver", "name": "hubble"}
2022-01-27T13:22:38.600Z	info	v3@v3.2103.1/logger.go:46	Set nextTxnTs to 0	{"kind": "receiver", "name": "hubble"}
2022-01-27T13:22:38.606Z	info	service/telemetry.go:116	Serving Prometheus metrics	{"address": ":8888", "level": "basic", "service.instance.id": "b597c5cb-2d86-4173-a8b1-0a5fecfd3635", "service.version": "latest"}
2022-01-27T13:22:38.606Z	info	service/collector.go:230	Starting otelcol-hubble...	{"Version": "0.1.0", "NumCPU": 40}
2022-01-27T13:22:38.606Z	info	service/collector.go:132	Everything is ready. Begin running and processing data.
2022-01-27T13:22:39.590Z	info	jaegerexporter@v0.38.0/exporter.go:186	State of the connection with the Jaeger Collector backend	{"kind": "exporter", "name": "jaeger", "state": "READY"}
2022-01-27T13:22:42.593Z	error	receiver@v0.0.0-00010101000000-000000000000/receiver.go:94	hubble reciever error	{"kind": "receiver", "name": "hubble", "error": "GetFlows failed: rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing dial tcp: lookup 172-81-0-10 on 172.22.0.10:53: server misbehaving\""}
github.com/cilium/hubble-otel/receiver.(*hubbleReceiver).Start.func1
	github.com/cilium/hubble-otel/receiver@v0.0.0-00010101000000-000000000000/receiver.go:94

Signed-off-by: weizhou Lan weizhou.lan@daocloud.io

@weizhoublue
Copy link
Contributor Author

weizhoublue commented Jan 28, 2022

@lizrice

@errordeveloper
Copy link
Contributor

Thanks for this PR @weizhoublue! I am not working on this project any more, but surely @lizrice can help you with this. The change looks good to me.

Signed-off-by: weizhou.lan@daocloud.io <weizhou.lan@daocloud.io>

2

Signed-off-by: weizhou.lan@daocloud.io <weizhou.lan@daocloud.io>
@weizhoublue
Copy link
Contributor Author

weizhoublue commented Jan 29, 2022

@errordeveloper thanks for reply

@glibsm glibsm requested a review from lizrice January 31, 2022 17:47
Copy link
Member

@lizrice lizrice left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM, thanks!

The tests failed but I don't think that's anything to do with this change - looks like a permissions issue. Please bear with us while we get that sorted before merging this PR

@lizrice
Copy link
Member

lizrice commented Feb 2, 2022

The permissions issue is nothing to do with this PR so I'm going ahead and merging this

@lizrice lizrice merged commit 6e522e4 into cilium:main Feb 2, 2022
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

3 participants