-
Notifications
You must be signed in to change notification settings - Fork 16
/
Makefile
468 lines (393 loc) · 21.8 KB
/
Makefile
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
# Image URL to use all building/pushing image targets
IMG_CONTROLLER ?= controller:dev
IMG_VXLAN_CONTROLLER ?= vxlan-controller:dev
IMG_RUNNER ?= runner:dev
IMG_SCHEDULER ?= autoscale-scheduler:dev
IMG_AUTOSCALER_AGENT ?= autoscaler-agent:dev
E2E_TESTS_VM_IMG ?= vm-postgres:15-bullseye
PG16_DISK_TEST_IMG ?= pg16-disk-test:dev
## Golang details
GOARCH ?= $(shell go env GOARCH)
GOOS ?= $(shell go env GOOS)
# Get the currently used golang base path
GOPATH=$(shell go env GOPATH)
# Get the currently used golang install path (in GOPATH/bin, unless GOBIN is set)
ifeq (,$(shell go env GOBIN))
GOBIN=$(GOPATH)/bin
else
GOBIN=$(shell go env GOBIN)
endif
# Go 1.20 changed the handling of git worktrees:
# https://github.com/neondatabase/autoscaling/pull/130#issuecomment-1496276620
export GOFLAGS=-buildvcs=false
# Setting SHELL to bash allows bash commands to be executed by recipes.
# Options are set to exit when a recipe line exits non-zero or a piped command fails.
SHELL = /usr/bin/env bash -o pipefail
.SHELLFLAGS = -ec
GIT_INFO := $(shell git describe --long --dirty)
# in CI environment use 'neonvm' as cluster name
# in other cases add $USER as cluster name suffix
# or fallback to 'neonvm' if $USER variable absent
ifdef CI
CLUSTER_NAME = neonvm
else ifdef USER
CLUSTER_NAME = neonvm-$(USER)
else
CLUSTER_NAME = neonvm
endif
.PHONY: all
all: build lint
##@ General
# The help target prints out all targets with their descriptions organized
# beneath their categories. The categories are represented by '##@' and the
# target descriptions by '##'. The awk commands is responsible for reading the
# entire set of makefiles included in this invocation, looking for lines of the
# file as xyz: ## something, and then pretty-format the target and help. Then,
# if there's a line with ##@ something, that gets pretty-printed as a category.
# More info on the usage of ANSI control characters for terminal formatting:
# https://en.wikipedia.org/wiki/ANSI_escape_code#SGR_parameters
# More info on the awk command:
# http://linuxcommand.org/lc3_adv_awk.php
.PHONY: help
help: ## Display this help.
@awk 'BEGIN {FS = ":.*##"; printf "\nUsage:\n make \033[36m<target>\033[0m\n"} /^[a-zA-Z_0-9-]+:.*?##/ { printf " \033[36m%-15s\033[0m %s\n", $$1, $$2 } /^##@/ { printf "\n\033[1m%s\033[0m\n", substr($$0, 5) } ' $(MAKEFILE_LIST)
##@ Development
# Generate a number of things:
# * Code containing DeepCopy, DeepCopyInto, and DeepCopyObject method implementations
# * WebhookConfiguration, ClusterRole, and CustomResourceDefinition objects
# * Go client
.PHONY: generate
generate: ## Generate boilerplate DeepCopy methods, manifests, and Go client
# Use uid and gid of current user to avoid mismatched permissions
iidfile=$$(mktemp /tmp/iid-XXXXXX) && \
docker build \
--build-arg USER_ID=$(shell id -u $(USER)) \
--build-arg GROUP_ID=$(shell id -g $(USER)) \
--build-arg CONTROLLER_TOOLS_VERSION=$(CONTROLLER_TOOLS_VERSION) \
--build-arg CODE_GENERATOR_VERSION=$(CODE_GENERATOR_VERSION) \
--file neonvm/hack/Dockerfile.generate \
--iidfile $$iidfile . && \
docker run --rm \
--volume $$PWD:/go/src/github.com/neondatabase/autoscaling \
--workdir /go/src/github.com/neondatabase/autoscaling \
--user $(shell id -u $(USER)):$(shell id -g $(USER)) \
$$(cat $$iidfile) \
./neonvm/hack/generate.sh && \
docker rmi $$(cat $$iidfile)
rm -rf $$iidfile
go fmt ./...
.PHONY: fmt
fmt: ## Run go fmt against code.
go fmt ./...
.PHONY: vet
vet: ## Run go vet against code.
# `go vet` requires gcc
# ref https://github.com/golang/go/issues/56755
CGO_ENABLED=0 go vet ./...
.PHONY: test
test: fmt vet envtest ## Run tests.
# chmodding KUBEBUILDER_ASSETS dir to make it deletable by owner,
# otherwise it fails with actions/checkout on self-hosted GitHub runners
# ref: https://github.com/kubernetes-sigs/controller-runtime/pull/2245
export KUBEBUILDER_ASSETS="$(shell $(ENVTEST) use $(ENVTEST_K8S_VERSION) --bin-dir $(LOCALBIN) -p path)"; \
find $(KUBEBUILDER_ASSETS) -type d -exec chmod 0755 {} \; ; \
CGO_ENABLED=0 \
go test ./... -coverprofile cover.out
##@ Build
.PHONY: build
build: fmt vet bin/vm-builder ## Build all neonvm binaries.
go build -o bin/controller neonvm/main.go
go build -o bin/vxlan-controller neonvm/tools/vxlan/controller/main.go
go build -o bin/runner neonvm/runner/*.go
.PHONY: bin/vm-builder
bin/vm-builder: ## Build vm-builder binary.
CGO_ENABLED=0 go build -o bin/vm-builder -ldflags "-X main.Version=${GIT_INFO}" neonvm/tools/vm-builder/main.go
.PHONY: run
run: fmt vet ## Run a controller from your host.
go run ./neonvm/main.go
.PHONY: lint
lint: ## Run golangci-lint against code.
golangci-lint run
# If you wish built the controller image targeting other platforms you can use the --platform flag.
# (i.e. docker build --platform linux/arm64 ). However, you must enable docker buildKit for it.
# More info: https://docs.docker.com/develop/develop-images/build_enhancements/
.PHONY: docker-build
docker-build: docker-build-controller docker-build-runner docker-build-vxlan-controller docker-build-autoscaler-agent docker-build-scheduler ## Build docker images for NeonVM controllers, NeonVM runner, autoscaler-agent, scheduler
.PHONY: docker-push
docker-push: docker-build ## Push docker images to docker registry
docker push -q $(IMG_CONTROLLER)
docker push -q $(IMG_RUNNER)
docker push -q $(IMG_VXLAN_CONTROLLER)
docker push -q $(IMG_SCHEDULER)
docker push -q $(IMG_AUTOSCALER_AGENT)
.PHONY: docker-build-controller
docker-build-controller: ## Build docker image for NeonVM controller
docker build --build-arg VM_RUNNER_IMAGE=$(IMG_RUNNER) --build-arg BUILDTAGS=$(if $(PRESERVE_RUNNER_PODS),nodelete) -t $(IMG_CONTROLLER) -f neonvm/Dockerfile .
.PHONY: docker-build-runner
docker-build-runner: ## Build docker image for NeonVM runner
docker build -t $(IMG_RUNNER) -f neonvm/runner/Dockerfile .
.PHONY: docker-build-vxlan-controller
docker-build-vxlan-controller: ## Build docker image for NeonVM vxlan controller
docker build -t $(IMG_VXLAN_CONTROLLER) -f neonvm/tools/vxlan/Dockerfile .
.PHONY: docker-build-autoscaler-agent
docker-build-autoscaler-agent: ## Build docker image for autoscaler-agent
docker buildx build \
--tag $(IMG_AUTOSCALER_AGENT) \
--load \
--build-arg "GIT_INFO=$(GIT_INFO)" \
--file build/autoscaler-agent/Dockerfile \
.
.PHONY: docker-build-scheduler
docker-build-scheduler: ## Build docker image for (autoscaling) scheduler
docker buildx build \
--tag $(IMG_SCHEDULER) \
--load \
--build-arg "GIT_INFO=$(GIT_INFO)" \
--file build/autoscale-scheduler/Dockerfile \
.
.PHONY: docker-build-examples
docker-build-examples: bin/vm-builder ## Build docker images for testing VMs
./bin/vm-builder -src postgres:15-bullseye -dst $(E2E_TESTS_VM_IMG) -spec tests/e2e/image-spec.yaml
.PHONY: docker-build-pg16-disk-test
docker-build-pg16-disk-test: bin/vm-builder ## Build a VM image for testing
./bin/vm-builder -src alpine:3.19 -dst $(PG16_DISK_TEST_IMG) -spec vm-examples/pg16-disk-test/image-spec.yaml
#.PHONY: docker-push
#docker-push: ## Push docker image with the controller.
# docker push ${IMG_CONTROLLER}
# PLATFORMS defines the target platforms for the controller image be build to provide support to multiple
# architectures. (i.e. make docker-buildx IMG=myregistry/mypoperator:0.0.1). To use this option you need to:
# - able to use docker buildx . More info: https://docs.docker.com/build/buildx/
# - have enable BuildKit, More info: https://docs.docker.com/develop/develop-images/build_enhancements/
# - be able to push the image for your registry (i.e. if you do not inform a valid value via IMG=<myregistry/image:<tag>> than the export will fail)
# To properly provided solutions that supports more than one platform you should use this option.
#PLATFORMS ?= linux/arm64,linux/amd64,linux/s390x,linux/ppc64le
#.PHONY: docker-buildx
#docker-buildx: test ## Build and push docker image for the controller for cross-platform support
# # copy existing Dockerfile and insert --platform=${BUILDPLATFORM} into Dockerfile.cross, and preserve the original Dockerfile
# sed -e '1 s/\(^FROM\)/FROM --platform=\$$\{BUILDPLATFORM\}/; t' -e ' 1,// s//FROM --platform=\$$\{BUILDPLATFORM\}/' Dockerfile > Dockerfile.cross
# - docker buildx create --name project-v3-builder
# docker buildx use project-v3-builder
# - docker buildx build --push --platform=$(PLATFORMS) --tag ${IMG_CONTROLLER} -f Dockerfile.cross
# - docker buildx rm project-v3-builder
# rm Dockerfile.cross
##@ Deployment
ifndef ignore-not-found
ignore-not-found = false
endif
.PHONY: kernel
kernel: ## Build linux kernel.
rm -f neonvm/hack/kernel/vmlinuz; \
linux_config=$$(ls neonvm/hack/kernel/linux-config-*) \
kernel_version=$${linux_config##*-} \
iidfile=$$(mktemp /tmp/iid-XXXXXX); \
trap "rm $$iidfile" EXIT; \
docker buildx build \
--build-arg KERNEL_VERSION=$$kernel_version \
--platform linux/amd64 \
--pull \
--load \
--iidfile $$iidfile \
--file neonvm/hack/kernel/Dockerfile.kernel-builder \
neonvm/hack/kernel; \
id=$$(docker create $$(cat $$iidfile)); \
docker cp $$id:/vmlinuz neonvm/hack/kernel/vmlinuz; \
docker rm -f $$id
.PHONY: check-local-context
check-local-context: ## Asserts that the current kubectl context is pointing at k3d or kind, to avoid accidentally applying to prod
@if [ "$$($(KUBECTL) config current-context)" != 'k3d-$(CLUSTER_NAME)' ] && [ "$$($(KUBECTL) config current-context)" != 'kind-$(CLUSTER_NAME)' ]; then echo "kubectl context is not pointing to local k3d or kind cluster (must be k3d-$(CLUSTER_NAME) or kind-$(CLUSTER_NAME))"; exit 1; fi
.PHONY: install
install: check-local-context kustomize ## Install CRDs into the K8s cluster specified in ~/.kube/config.
$(KUSTOMIZE) build neonvm/config/crd | kubectl apply -f -
.PHONY: uninstall
uninstall: check-local-context kustomize ## Uninstall CRDs from the K8s cluster specified in ~/.kube/config.
$(KUSTOMIZE) build neonvm/config/crd | kubectl delete --ignore-not-found=$(ignore-not-found) -f -
BUILDTS := $(shell date +%s)
RENDERED ?= $(shell pwd)/rendered_manifests
$(RENDERED):
mkdir -p $(RENDERED)
.PHONY: render-manifests
render-manifests: $(RENDERED) kustomize
# Prepare:
cd neonvm/config/controller && $(KUSTOMIZE) edit set image controller=$(IMG_CONTROLLER) && $(KUSTOMIZE) edit add annotation buildtime:$(BUILDTS) --force
cd neonvm/config/vxlan-controller && $(KUSTOMIZE) edit set image vxlan-controller=$(IMG_VXLAN_CONTROLLER) && $(KUSTOMIZE) edit add annotation buildtime:$(BUILDTS) --force
cd neonvm/runner-image-loader && $(KUSTOMIZE) edit set image runner=$(IMG_RUNNER) && $(KUSTOMIZE) edit add annotation buildtime:$(BUILDTS) --force
cd deploy/scheduler && $(KUSTOMIZE) edit set image autoscale-scheduler=$(IMG_SCHEDULER) && $(KUSTOMIZE) edit add annotation buildtime:$(BUILDTS) --force
cd deploy/agent && $(KUSTOMIZE) edit set image autoscaler-agent=$(IMG_AUTOSCALER_AGENT) && $(KUSTOMIZE) edit add annotation buildtime:$(BUILDTS) --force
# Build:
$(KUSTOMIZE) build neonvm/config/whereabouts > $(RENDERED)/whereabouts.yaml
$(KUSTOMIZE) build neonvm/config/multus-eks > $(RENDERED)/multus-eks.yaml
$(KUSTOMIZE) build neonvm/config/multus > $(RENDERED)/multus.yaml
$(KUSTOMIZE) build neonvm/config > $(RENDERED)/neonvm.yaml
$(KUSTOMIZE) build neonvm/runner-image-loader > $(RENDERED)/neonvm-runner-image-loader.yaml
$(KUSTOMIZE) build deploy/scheduler > $(RENDERED)/autoscale-scheduler.yaml
$(KUSTOMIZE) build deploy/agent > $(RENDERED)/autoscaler-agent.yaml
# Cleanup:
cd neonvm/config/controller && $(KUSTOMIZE) edit set image controller=controller:dev && $(KUSTOMIZE) edit remove annotation buildtime --ignore-non-existence
cd neonvm/config/vxlan-controller && $(KUSTOMIZE) edit set image vxlan-controller=vxlan-controller:dev && $(KUSTOMIZE) edit remove annotation buildtime --ignore-non-existence
cd neonvm/runner-image-loader && $(KUSTOMIZE) edit set image runner=runner:dev && $(KUSTOMIZE) edit remove annotation buildtime --ignore-non-existence
cd deploy/scheduler && $(KUSTOMIZE) edit set image autoscale-scheduler=autoscale-scheduler:dev && $(KUSTOMIZE) edit remove annotation buildtime --ignore-non-existence
cd deploy/agent && $(KUSTOMIZE) edit set image autoscaler-agent=autoscaler-agent:dev && $(KUSTOMIZE) edit remove annotation buildtime --ignore-non-existence
render-release: $(RENDERED) kustomize
# Prepare:
cd neonvm/config/controller && $(KUSTOMIZE) edit set image controller=$(IMG_CONTROLLER)
cd neonvm/config/vxlan-controller && $(KUSTOMIZE) edit set image vxlan-controller=$(IMG_VXLAN_CONTROLLER)
cd neonvm/runner-image-loader && $(KUSTOMIZE) edit set image runner=$(IMG_RUNNER)
cd deploy/scheduler && $(KUSTOMIZE) edit set image autoscale-scheduler=$(IMG_SCHEDULER)
cd deploy/agent && $(KUSTOMIZE) edit set image autoscaler-agent=$(IMG_AUTOSCALER_AGENT)
# Build:
$(KUSTOMIZE) build neonvm/config/whereabouts > $(RENDERED)/whereabouts.yaml
$(KUSTOMIZE) build neonvm/config/multus-eks > $(RENDERED)/multus-eks.yaml
$(KUSTOMIZE) build neonvm/config/multus > $(RENDERED)/multus.yaml
$(KUSTOMIZE) build neonvm/config > $(RENDERED)/neonvm.yaml
$(KUSTOMIZE) build neonvm/runner-image-loader > $(RENDERED)/neonvm-runner-image-loader.yaml
$(KUSTOMIZE) build deploy/scheduler > $(RENDERED)/autoscale-scheduler.yaml
$(KUSTOMIZE) build deploy/agent > $(RENDERED)/autoscaler-agent.yaml
# Cleanup:
cd neonvm/config/controller && $(KUSTOMIZE) edit set image controller=controller:dev
cd neonvm/config/vxlan-controller && $(KUSTOMIZE) edit set image vxlan-controller=vxlan-controller:dev
cd neonvm/runner-image-loader && $(KUSTOMIZE) edit set image runner=runner:dev
cd deploy/scheduler && $(KUSTOMIZE) edit set image autoscale-scheduler=autoscale-scheduler:dev
cd deploy/agent && $(KUSTOMIZE) edit set image autoscaler-agent=autoscaler-agent:dev
.PHONY: deploy
deploy: check-local-context docker-build load-images render-manifests kubectl ## Deploy controller to the K8s cluster specified in ~/.kube/config.
$(KUBECTL) apply -f $(RENDERED)/multus.yaml
$(KUBECTL) -n kube-system rollout status daemonset kube-multus-ds
$(KUBECTL) apply -f $(RENDERED)/whereabouts.yaml
$(KUBECTL) -n kube-system rollout status daemonset whereabouts
$(KUBECTL) apply -f $(RENDERED)/neonvm-runner-image-loader.yaml
$(KUBECTL) -n neonvm-system rollout status daemonset neonvm-runner-image-loader
$(KUBECTL) apply -f $(RENDERED)/neonvm.yaml
$(KUBECTL) -n neonvm-system rollout status daemonset neonvm-device-plugin
$(KUBECTL) -n neonvm-system rollout status daemonset neonvm-vxlan-controller
$(KUBECTL) -n neonvm-system rollout status deployment neonvm-controller
# NB: typical upgrade path requires updated scheduler before autoscaler-agents.
$(KUBECTL) apply -f $(RENDERED)/autoscale-scheduler.yaml
$(KUBECTL) -n kube-system rollout status deployment autoscale-scheduler
$(KUBECTL) apply -f $(RENDERED)/autoscaler-agent.yaml
$(KUBECTL) -n kube-system rollout status daemonset autoscaler-agent
.PHONY: load-images
load-images: check-local-context kubectl kind k3d ## Push docker images to the local kind/k3d cluster
@if [ $$($(KUBECTL) config current-context) = k3d-$(CLUSTER_NAME) ]; then make k3d-load; fi
@if [ $$($(KUBECTL) config current-context) = kind-$(CLUSTER_NAME) ]; then make kind-load; fi
.PHONY: load-example-vms
load-example-vms: check-local-context kubectl kind k3d ## Load the testing VM image to the kind/k3d cluster.
@if [ $$($(KUBECTL) config current-context) = k3d-$(CLUSTER_NAME) ]; then $(K3D) image import $(E2E_TESTS_VM_IMG) --cluster $(CLUSTER_NAME) --mode direct; fi
@if [ $$($(KUBECTL) config current-context) = kind-$(CLUSTER_NAME) ]; then $(KIND) load docker-image $(E2E_TESTS_VM_IMG) --name $(CLUSTER_NAME); fi
.PHONY: example-vms
example-vms: docker-build-examples load-example-vms ## Build and push the testing VM images to the kind/k3d cluster.
.PHONY: load-pg16-disk-test
load-pg16-disk-test: check-local-context kubectl kind k3d ## Load the pg16-disk-test VM image to the kind/k3d cluster.
@if [ $$($(KUBECTL) config current-context) = k3d-$(CLUSTER_NAME) ]; then $(K3D) image import $(PG16_DISK_TEST_IMG) --cluster $(CLUSTER_NAME) --mode direct; fi
@if [ $$($(KUBECTL) config current-context) = kind-$(CLUSTER_NAME) ]; then $(KIND) load docker-image $(PG16_DISK_TEST_IMG) --name $(CLUSTER_NAME); fi
.PHONY: pg16-disk-test
pg16-disk-test: docker-build-pg16-disk-test load-pg16-disk-test ## Build and push the pg16-disk-test VM test image to the kind/k3d cluster.
.PHONY: kind-load
kind-load: kind # Push docker images to the kind cluster.
$(KIND) load docker-image \
$(IMG_CONTROLLER) \
$(IMG_RUNNER) \
$(IMG_VXLAN_CONTROLLER) \
$(IMG_SCHEDULER) \
$(IMG_AUTOSCALER_AGENT) \
--name $(CLUSTER_NAME)
.PHONY: k3d-load
k3d-load: k3d # Push docker images to the k3d cluster.
$(K3D) image import \
$(IMG_CONTROLLER) \
$(IMG_RUNNER) \
$(IMG_VXLAN_CONTROLLER) \
$(IMG_SCHEDULER) \
$(IMG_AUTOSCALER_AGENT) \
--cluster $(CLUSTER_NAME) --mode direct
##@ End-to-End tests
.PHONE: e2e-tools
e2e-tools: k3d kind kubectl kuttl python-init ## Donwnload tools for e2e tests locally if necessary.
.PHONE: e2e
e2e: check-local-context e2e-tools ## Run e2e kuttl tests
$(KUTTL) test --config tests/e2e/kuttl-test.yaml $(if $(CI),--skip-delete)
rm -f kubeconfig
##@ Local kind cluster
.PHONY: kind-setup
kind-setup: kind kubectl ## Create local cluster by kind tool and prepared config
$(KIND) create cluster --name $(CLUSTER_NAME) --config kind/config.yaml
$(KUBECTL) --context kind-$(CLUSTER_NAME) apply -f https://github.com/cert-manager/cert-manager/releases/latest/download/cert-manager.yaml
$(KUBECTL) --context kind-$(CLUSTER_NAME) -n cert-manager rollout status deployment cert-manager
$(KUBECTL) --context kind-$(CLUSTER_NAME) apply -f https://github.com/kubernetes-sigs/metrics-server/releases/latest/download/components.yaml
$(KUBECTL) --context kind-$(CLUSTER_NAME) patch -n kube-system deployment metrics-server --type=json -p '[{"op":"add","path":"/spec/template/spec/containers/0/args/-","value":"--kubelet-insecure-tls"}]'
$(KUBECTL) --context kind-$(CLUSTER_NAME) -n kube-system rollout status deployment metrics-server
.PHONY: kind-destroy
kind-destroy: kind ## Destroy local kind cluster
$(KIND) delete cluster --name $(CLUSTER_NAME)
##@ Local k3d cluster
# K3D_FIX_MOUNTS=1 used to allow foreign CNI (cilium, multus and so on), https://github.com/k3d-io/k3d/pull/1268
.PHONY: k3d-setup
k3d-setup: k3d kubectl ## Create local cluster by k3d tool and prepared config
K3D_FIX_MOUNTS=1 $(K3D) cluster create $(CLUSTER_NAME) --config k3d/config.yaml
$(KUBECTL) --context k3d-$(CLUSTER_NAME) apply -f k3d/cilium.yaml
$(KUBECTL) --context k3d-$(CLUSTER_NAME) -n kube-system rollout status daemonset cilium
$(KUBECTL) --context k3d-$(CLUSTER_NAME) -n kube-system rollout status deployment cilium-operator
$(KUBECTL) --context k3d-$(CLUSTER_NAME) apply -f https://github.com/cert-manager/cert-manager/releases/latest/download/cert-manager.yaml
$(KUBECTL) --context k3d-$(CLUSTER_NAME) -n cert-manager rollout status deployment cert-manager
.PHONY: k3d-destroy
k3d-destroy: k3d ## Destroy local k3d cluster
$(K3D) cluster delete $(CLUSTER_NAME)
##@ Build Dependencies
## Location to install dependencies to
LOCALBIN ?= $(shell pwd)/bin
$(LOCALBIN):
mkdir -p $(LOCALBIN)
## Tools
KUSTOMIZE ?= $(LOCALBIN)/kustomize
KUSTOMIZE_VERSION ?= v4.5.7
ENVTEST ?= $(LOCALBIN)/setup-envtest
# ENVTEST_K8S_VERSION refers to the version of kubebuilder assets to be downloaded by envtest binary.
# List of available versions: https://storage.googleapis.com/kubebuilder-tools
ENVTEST_K8S_VERSION = 1.25.0
CONTROLLER_GEN ?= $(LOCALBIN)/controller-gen
CONTROLLER_TOOLS_VERSION ?= v0.10.0
CODE_GENERATOR_VERSION ?= v0.25.16
KUTTL ?= $(LOCALBIN)/kuttl
KUTTL_VERSION ?= v0.15.0
KUBECTL ?= $(LOCALBIN)/kubectl
KUBECTL_VERSION ?= v1.25.16
KIND ?= $(LOCALBIN)/kind
KIND_VERSION ?= v0.20.0
K3D ?= $(LOCALBIN)/k3d
K3D_VERSION ?= v5.5.1
## Install tools
KUSTOMIZE_INSTALL_SCRIPT ?= "https://raw.githubusercontent.com/kubernetes-sigs/kustomize/master/hack/install_kustomize.sh"
.PHONY: kustomize
kustomize: $(KUSTOMIZE) ## Download kustomize locally if necessary.
$(KUSTOMIZE): $(LOCALBIN)
@test -s $(LOCALBIN)/kustomize || { curl -Ss $(KUSTOMIZE_INSTALL_SCRIPT) | bash -s -- $(subst v,,$(KUSTOMIZE_VERSION)) $(LOCALBIN); }
.PHONY: envtest
envtest: $(ENVTEST) ## Download envtest-setup locally if necessary.
$(ENVTEST): $(LOCALBIN)
@test -s $(LOCALBIN)/setup-envtest || GOBIN=$(LOCALBIN) go install sigs.k8s.io/controller-runtime/tools/setup-envtest@latest
.PHONY: controller-gen
controller-gen: $(CONTROLLER_GEN) ## Download controller-gen locally if necessary.
$(CONTROLLER_GEN): $(LOCALBIN)
@test -s $(LOCALBIN)/controller-gen || GOBIN=$(LOCALBIN) go install sigs.k8s.io/controller-tools/cmd/controller-gen@$(CONTROLLER_TOOLS_VERSION)
.PHONY: kind
kind: $(KIND) ## Download kind locally if necessary.
$(KIND): $(LOCALBIN)
@test -s $(LOCALBIN)/kind || { curl -sfSLo $(KIND) https://kind.sigs.k8s.io/dl/$(KIND_VERSION)/kind-$(GOOS)-$(GOARCH) && chmod +x $(KIND); }
.PHONY: kubectl
kubectl: $(KUBECTL) ## Download kubectl locally if necessary.
$(KUBECTL): $(LOCALBIN)
@test -s $(LOCALBIN)/kubectl || { curl -sfSLo $(KUBECTL) https://dl.k8s.io/release/$(KUBECTL_VERSION)/bin/$(GOOS)/$(GOARCH)/kubectl && chmod +x $(KUBECTL); }
.PHONY: kuttl
kuttl: $(KUTTL) ## Download kuttl locally if necessary.
$(KUTTL): $(LOCALBIN)
@test -s $(LOCALBIN)/kuttl || { curl -sfSLo $(KUTTL) https://github.com/kudobuilder/kuttl/releases/download/$(KUTTL_VERSION)/kubectl-kuttl_$(subst v,,$(KUTTL_VERSION))_$(GOOS)_$(shell uname -m) && chmod +x $(KUTTL); }
.PHONY: k3d
k3d: $(K3D) ## Download k3d locally if necessary.
$(K3D): $(LOCALBIN)
@test -s $(LOCALBIN)/k3d || { curl -sfSLo $(K3D) https://github.com/k3d-io/k3d/releases/download/$(K3D_VERSION)/k3d-$(GOOS)-$(GOARCH) && chmod +x $(K3D); }
.PHONY: cert-manager
cert-manager: check-local-context kubectl ## install cert-manager to cluster
$(KUBECTL) apply -f https://github.com/cert-manager/cert-manager/releases/latest/download/cert-manager.yaml
.PHONY: python-init
python-init:
python3 -m venv tests/e2e/.venv
tests/e2e/.venv/bin/pip install -r requirements.txt