Skip to content
This repository has been archived by the owner on Oct 29, 2019. It is now read-only.

[bsc#1125100] update image refs #758

Merged
merged 4 commits into from
Mar 6, 2019
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion pillar/cni.sls
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
# the flannel backend ('udp', 'vxlan', 'host-gw', etc)
flannel:
image: '{{ salt.caasp_registry.base_image_url() }}/flannel:0.9.1'
image: 'registry.suse.com/caasp/v4/flannel:0.9.1'
backend: 'vxlan'
port: '8472' # UDP port to use for sending encapsulated packets. Defaults to kernel default, currently 8472.
healthz_port: '8471' # TCP port used for flannel healthchecks
Expand Down
2 changes: 1 addition & 1 deletion pillar/params.sls
Original file line number Diff line number Diff line change
Expand Up @@ -198,4 +198,4 @@ ldap:
mail_attribute: ''

# infra container to use instead of downloading gcr.io/google_containers/pause
pod_infra_container_image: {{ salt.caasp_registry.base_image_url() }}/pause:1.0.0
pod_infra_container_image: registry.suse.com/caasp/v4/pause:1.0.0
75 changes: 0 additions & 75 deletions salt/_modules/caasp_registry.py

This file was deleted.

2 changes: 1 addition & 1 deletion salt/addons/dex/manifests/20-deployment.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -58,7 +58,7 @@ spec:
operator: Exists

containers:
- image: {{ salt.caasp_registry.base_image_url() }}/caasp-dex:2.7.1
- image: registry.suse.com/caasp/v4/caasp-dex:2.7.1
name: dex
command: ["/usr/bin/caasp-dex", "serve", "/etc/dex/cfg/config.yaml"]

Expand Down
6 changes: 3 additions & 3 deletions salt/addons/dns/manifests/20-deployment.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -47,7 +47,7 @@ spec:
name: kube-dns
containers:
- name: kubedns
image: {{ salt.caasp_registry.base_image_url() }}/kubedns:1.0.0
image: registry.suse.com/caasp/v4/kubedns:1.14.1
resources:
# TODO: Set memory limits when we've profiled the container for large
# clusters, then set request = limit to keep this container in
Expand Down Expand Up @@ -98,7 +98,7 @@ spec:
- name: kube-dns-config
mountPath: /kube-dns-config
- name: dnsmasq
image: {{ salt.caasp_registry.base_image_url() }}/dnsmasq-nanny:1.0.0
image: registry.suse.com/caasp/v4/dnsmasq-nanny:2.78
livenessProbe:
httpGet:
path: /healthcheck/dnsmasq
Expand Down Expand Up @@ -136,7 +136,7 @@ spec:
- name: kube-dns-config
mountPath: /etc/k8s/dns/dnsmasq-nanny
- name: sidecar
image: {{ salt.caasp_registry.base_image_url() }}/sidecar:1.0.0
image: registry.suse.com/caasp/v4/sidecar:1.14.1
livenessProbe:
httpGet:
path: /metrics
Expand Down
2 changes: 1 addition & 1 deletion salt/addons/tiller/manifests/20-deployment.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -44,7 +44,7 @@ spec:
- env:
- name: TILLER_NAMESPACE
value: kube-system
image: {{ salt.caasp_registry.base_image_url() }}/tiller:2.8.2
image: registry.suse.com/caasp/v4/tiller:2.8.2
imagePullPolicy: IfNotPresent
livenessProbe:
httpGet:
Expand Down
6 changes: 3 additions & 3 deletions salt/cni/cilium-ds.yaml.jinja
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@ spec:
serviceAccountName: cilium
initContainers:
- name: install-cni-conf
image: {{ salt.caasp_registry.base_image_url() }}/{{ pillar['cilium']['image'] }}
image: registry.suse.com/caasp/v4/{{ pillar['cilium']['image'] }}
command:
- /bin/sh
- "-c"
Expand All @@ -36,7 +36,7 @@ spec:
- name: host-cni-conf
mountPath: /host/etc/cni/net.d
- name: install-cni-bin
image: {{ salt.caasp_registry.base_image_url() }}/{{ pillar['cilium']['image'] }}
image: registry.suse.com/caasp/v4/{{ pillar['cilium']['image'] }}
command:
- /bin/sh
- "-c"
Expand All @@ -49,7 +49,7 @@ spec:
priorityClassName: system-node-critical

containers:
- image: {{ salt.caasp_registry.base_image_url() }}/{{ pillar['cilium']['image'] }}
- image: registry.suse.com/caasp/v4/{{ pillar['cilium']['image'] }}
imagePullPolicy: IfNotPresent
name: cilium-agent
command: [ "cilium-agent" ]
Expand Down
23 changes: 0 additions & 23 deletions salt/container-feeder/container-feeder.json.jinja

This file was deleted.

31 changes: 0 additions & 31 deletions salt/container-feeder/init.sls

This file was deleted.

4 changes: 0 additions & 4 deletions salt/container-feeder/stop.sls

This file was deleted.

1 change: 0 additions & 1 deletion salt/container-feeder/sysconfig

This file was deleted.

7 changes: 0 additions & 7 deletions salt/crio/init.sls
Original file line number Diff line number Diff line change
@@ -1,8 +1,5 @@
include:
- kubelet
{%- if not salt.caasp_registry.use_registry_images() %}
- container-feeder
{%- endif %}

crio:
file.managed:
Expand All @@ -23,10 +20,6 @@ crio:
- makedirs: True
- require_in:
- kubelet
{%- if not salt.caasp_registry.use_registry_images() %}
- require:
- service: container-feeder
{%- endif %}
module.run:
- name: service.systemctl_reload
- onchanges:
Expand Down
2 changes: 1 addition & 1 deletion salt/haproxy/haproxy.yaml.jinja
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ spec:
priorityClassName: system-node-critical
containers:
- name: haproxy
image: {{ salt.caasp_registry.base_image_url() }}/haproxy:1.6.0
image: registry.suse.com/caasp/v4/haproxy:1.6.11
resources:
requests:
memory: 128Mi
Expand Down
6 changes: 0 additions & 6 deletions salt/haproxy/init.sls
Original file line number Diff line number Diff line change
Expand Up @@ -11,9 +11,6 @@ include:
# failing orchestrations.
- kubelet
- {{ salt['pillar.get']('cri:chosen', 'docker') }}
{%- if not salt.caasp_registry.use_registry_images() %}
- container-feeder
{%- endif %}
{% endif %}

/etc/caasp/haproxy:
Expand Down Expand Up @@ -86,9 +83,6 @@ haproxy-restart:
{% if not salt.caasp_nodes.is_admin_node() %}
- require:
- service: kubelet
{%- if not salt.caasp_registry.use_registry_images() %}
- service: container-feeder
{%- endif %}
{% endif %}


Expand Down
3 changes: 0 additions & 3 deletions salt/orch/removal.sls
Original file line number Diff line number Diff line change
Expand Up @@ -306,9 +306,6 @@ stop-services-in-target:
salt.state:
- tgt: '{{ target }}'
- sls:
{%- if not salt.caasp_registry.use_registry_images() %}
- container-feeder.stop
{%- endif %}
{%- if target in masters %}
- kube-apiserver.stop
- kube-controller-manager.stop
Expand Down
6 changes: 0 additions & 6 deletions salt/orch/update.sls
Original file line number Diff line number Diff line change
Expand Up @@ -215,9 +215,6 @@ early-services-setup:
salt.state:
- tgt: '{{ master_id }}'
- sls:
{%- if not salt.caasp_registry.use_registry_images() %}
- container-feeder.stop
{%- endif %}
- kube-apiserver.stop
- kube-controller-manager.stop
- kube-scheduler.stop
Expand Down Expand Up @@ -377,9 +374,6 @@ all-workers-2.0-pre-clean-shutdown:
salt.state:
- tgt: '{{ worker_id }}'
- sls:
{%- if not salt.caasp_registry.use_registry_images() %}
- container-feeder.stop
{%- endif %}
- kube-proxy.stop
- cri.stop
- etcd.stop
Expand Down
3 changes: 0 additions & 3 deletions salt/top.sls
Original file line number Diff line number Diff line change
Expand Up @@ -6,9 +6,6 @@ base:
- match: grain_pcre
- ca-cert
- cri
{%- if not salt.caasp_registry.use_registry_images() %}
- container-feeder
{%- endif %}
{% if not salt.caasp_nodes.is_admin_node() %}
# the admin node uses docker as CRI, requiring its state
# will cause the docker daemon to be restarted, which will
Expand Down