Kustomize provides two ways of adding ConfigMap in one kustomization
, either by declaring ConfigMap as a resource or declaring ConfigMap from a ConfigMapGenerator. The formats inside kustomization.yaml
are
# declare ConfigMap as a resource resources: - configmap.yaml # declare ConfigMap from a ConfigMapGenerator configMapGenerator: - name: a-configmap files: # configfile is used as key - configs/configfile # configkey is used as key - configkey=configs/another_configfile
The ConfigMaps declared as resource are treated the same way as other resources. Kustomize doesn't append any hash to the ConfigMap name. The ConfigMap declared from a ConfigMapGenerator is treated differently. A hash is appended to the name and any change in the ConfigMap will trigger a rolling update.
In this demo, the same hello_world is used while the ConfigMap declared as [resources] is replaced by a ConfigMap declared from a ConfigMapGenerator. The change in this ConfigMap will result in a hash change and a rolling update.
Establish the base with a configMapGenerator
:
DEMO_HOME=$(mktemp -d)
BASE=$DEMO_HOME/base
mkdir -p $BASE
curl -s -o "$BASE/#1.yaml" "https://raw.githubusercontent.com\
/kubernetes-sigs/kustomize\
/master/examples/helloWorld\
/{deployment,service}.yaml"
cat <<'EOF' >$BASE/kustomization.yaml
commonLabels:
app: hello
resources:
- deployment.yaml
- service.yaml
configMapGenerator:
- name: the-map
literals:
- altGreeting=Good Morning!
- enableRisky="false"
EOF
Establish the staging with a patch applied to the ConfigMap
OVERLAYS=$DEMO_HOME/overlays
mkdir -p $OVERLAYS/staging
cat <<'EOF' >$OVERLAYS/staging/kustomization.yaml
namePrefix: staging-
nameSuffix: -v1
commonLabels:
variant: staging
org: acmeCorporation
commonAnnotations:
note: Hello, I am staging!
resources:
- ../../base
patches:
- path: map.yaml
EOF
cat <<EOF >$OVERLAYS/staging/map.yaml
apiVersion: v1
kind: ConfigMap
metadata:
name: the-map
data:
altGreeting: "Have a pineapple!"
enableRisky: "true"
EOF
The hello-world deployment running in this cluster is configured with data from a ConfigMap.
The deployment refers to this map by name:
grep -C 2 configMapKeyRef $BASE/deployment.yaml
Changing the data held by a live ConfigMap in a cluster is considered bad practice. Deployments have no means to know that the ConfigMaps they refer to have changed, so such updates have no effect.
The recommended way to change a deployment's configuration is to
- create a new ConfigMap with a new name,
- patch the deployment, modifying the name value of
the appropriate
configMapKeyRef
field.
This latter change initiates rolling update to the pods in the deployment.
The staging variant here has a ConfigMap patch:
cat $OVERLAYS/staging/map.yaml
This patch is by definition a named but not necessarily complete resource spec intended to modify a complete resource spec.
The ConfigMap it modifies is declared from a configMapGenerator
.
grep -C 4 configMapGenerator $BASE/kustomization.yaml
For a patch to work, the names in the metadata/name
fields must match.
However, the name values specified in the file are not what gets used in the cluster. By design, kustomize modifies names of ConfigMaps declared from ConfigMapGenerator. To see the names ultimately used in the cluster, just run kustomize:
kustomize build $OVERLAYS/staging |\
grep -B 8 -A 1 staging-the-map
The ConfigMap name is prefixed by staging-, per the
namePrefix
field in
$OVERLAYS/staging/kustomization.yaml
.
The ConfigMap name is suffixed by -v1, per the
nameSuffix
field in
$OVERLAYS/staging/kustomization.yaml
.
The suffix to the ConfigMap name is generated from a hash of the maps content - in this case the name suffix is 5276h4th55:
kustomize build $OVERLAYS/staging | grep 5276h4th55
Now modify the map patch, to change the greeting the server will use:
sed -i.bak 's/pineapple/kiwi/' $OVERLAYS/staging/map.yaml
See the new greeting:
kustomize build $OVERLAYS/staging |\
grep -B 2 -A 3 kiwi
Run kustomize again to see the new ConfigMap names:
kustomize build $OVERLAYS/staging |\
grep -B 8 -A 1 staging-the-map
Confirm that the change in ConfigMap content resulted in three new names ending in c2g8fcbf88 - one in the ConfigMap name itself, and two in the deployment that uses the map:
test 3 == \
$(kustomize build $OVERLAYS/staging | grep c2g8fcbf88 | wc -l); \
echo $?
Applying these resources to the cluster will result in a rolling update of the deployments pods, retargetting them from the 5276h4th55 maps to the c2g8fcbf88 maps. The system will later garbage collect the unused maps.
To rollback, one would undo whatever edits were made to the configuation in source control, then rerun kustomize on the reverted configuration and apply it to the cluster.