Professional Documents
Culture Documents
Core Concepts
Scheduling
Logging Monitoring
Security
Storage
Networking
Troubleshooting
3
Operating
System Upgrade
M W W W
Scheduling
Logging Monitoring
Security
Storage
Networking
Troubleshooting
7
Kubernetes
Releases
kubectl get nodes
NAME STATUS ROLES AGE VERSION
master Ready master 1d v1.11.3
node-1 Ready <none> 1d v1.11.3
node-2 Ready <none> 1d v1.11.3
v1.11.3
MAJOR MINOR PATCH
V1.10.0-beta
V1.10.0-alpha
https://github.com/kubernetes/kubernetes/releases
v1.13.4
v3.2.18 v1.1.3
Course Objectives
Core Concepts
Scheduling
Logging Monitoring
Security
Storage
Networking
Troubleshooting
17
Cluster Upgrade
Process
kube-apiserver Controller-manager kube-scheduler kubelet kube-proxy kubectl
v3.2.18 v1.1.3
kube-apiserver
X v1.10
kubectl
Controller-manager kube-scheduler
kubelet kube-proxy
X-2 X-2
v1.8 or v1.9 or v.110 V1.8 or v1.9 or v1.10
kube-apiserver
v1.13 Supported
v1.10
v1.11 Supported
kubelet kube-proxy
v1.10 v1.10
v1.10 Un-Supported
kubeadm “The hard way”
M W W W
M W W
v1.10
Strategy - 2
2
1 3
M W W
v1.10
Strategy - 2
2
1 3
M W W
v1.10
Strategy - 2
2
1 3
M W W W
M W W W W
M W W W
W
v1.11 v1.10 v1.10 v1.11
v1.10
Strategy - 3
M W W W W
You can now apply the upgrade by executing the following command:
kubeadm - upgrade
kubeadm upgrade plan
[preflight] Running pre-flight checks.
[upgrade] Making sure the cluster is healthy:
[upgrade/config] Making sure the configuration is correct:
[upgrade] Fetching available versions to upgrade to
[upgrade/versions] Cluster version: v1.11.8
[upgrade/versions] kubeadm version: v1.11.3
[upgrade/versions] Latest stable version: v1.13.4
[upgrade/versions] Latest version in the v1.11 series: v1.11.8
You can now apply the upgrade by executing the following command:
Note: Before you can perform this upgrade, you have to update kubeadm to v1.13.4.
kubeadm - upgrade
apt-get upgrade -y kubeadm=1.12.0-00
[upgrade/kubelet] Now that your control plane is upgraded, please proceed with
upgrading your kubelets if you haven't already done so.
M
kubectl get nodes
NAME STATUS ROLES AGE VERSION
v1.12 master Ready master 1d v1.12.0
node-1 Ready <none> 1d v1.11.3
node-2 Ready <none> 1d v1.11.3
kubeadm - upgrade
M W W W
Scheduling
Logging Monitoring
Security
Storage
Networking
Troubleshooting
38
Backup and
Restore
Backup Candidates
Resource Configuration
kubectl create configmap
Declarative pod-definition.yml
apiVersion: v1
kind: Pod
metadata:
name: myapp-pod
labels:
app: myapp
Resource Configuration type: front-end
spec:
containers:
- name: nginx-container
image: nginx
kube-apiserver
Resource Configuration
ETCD Cluster
Backup - ETCD
ETCD Cluster
etcd.service
ExecStart=/usr/local/bin/etcd \\
--name ${ETCD_NAME} \\
--cert-file=/etc/etcd/kubernetes.pem \\
ETCD ETCD ETCD --key-file=/etc/etcd/kubernetes-key.pem \\
--peer-cert-file=/etc/etcd/kubernetes.pem \\
M M M --peer-key-file=/etc/etcd/kubernetes-key.pem \\
--trusted-ca-file=/etc/etcd/ca.pem \\
--peer-trusted-ca-file=/etc/etcd/ca.pem \\
--peer-client-cert-auth \\
--client-cert-auth \\
--initial-advertise-peer-urls https://${INTERNAL_IP}:2
--listen-peer-urls https://${INTERNAL_IP}:2380 \\
--listen-client-urls https://${INTERNAL_IP}:2379,https
--advertise-client-urls https://${INTERNAL_IP}:2379 \\
--initial-cluster-token etcd-cluster-0 \\
--initial-cluster controller-0=https://${CONTROLLER0_I
--initial-cluster-state new \\
--data-dir=/var/lib/etcd
Backup - ETCD
ETCD Cluster
ETCDCTL_API=3 etcdctl \
snapshot save snapshot.db
ETCDCTL_API=3 etcdctl \
snapshot status snapshot.db
+----------+----------+------------+------------+
| HASH | REVISION | TOTAL KEYS | TOTAL SIZE |
+----------+----------+------------+------------+
| e63b3fc5 | 473353 | 875 | 4.1 MB |
+----------+----------+------------+------------+
ETCDCTL_API=3 etcdctl \
ls
snapshot.db
ETCD Cluster
service kube-apiserver stop
Service kube-apiserver stopped
ETCDCTL_API=3 etcdctl \
snapshot restore snapshot.db \
--data-dir /var/lib/etcd-from-backup \
--initial-cluster master-1=https://192.168.5.11:2380,master-2=https://192.168.5.12:2380 \
--initial-cluster-token etcd-cluster-1 \
--initial-advertise-peer-urls https://${INTERNAL_IP}:2380
I | mvcc: restore compact to 475629
I | etcdserver/membership: added member 5e89ccdfe3 [https://192.168.5.12:2380] to cluster 894c7131f5165a78
I | etcdserver/membership: added member c8246cee7c [https://192.168.5.11:2380] to cluster 894c7131f5165a78
ETCDCTL_API=3 etcdctl \
Restore - ETCD ls
snapshot save snapshot.db
snapshot.db
ETCDCTL_API=3 etcdctl \
snapshot restore snapshot.db \
etcd.service
--data-dir /var/lib/etcd-from-backup \ ExecStart=/usr/local/bin/etcd \\
--initial-cluster master- --name ${ETCD_NAME} \\
1=https://192.168.5.11:2380,master- --cert-file=/etc/etcd/kubernetes.pem \\
2=https://192.168.5.12:2380 \ --key-file=/etc/etcd/kubernetes-key.pem \\
--initial-cluster-token etcd-cluster-1 \ --peer-cert-file=/etc/etcd/kubernetes.pem \\
--initial-advertise-peer-urls --peer-key-file=/etc/etcd/kubernetes-key.pem \\
https://${INTERNAL_IP}:2380 --trusted-ca-file=/etc/etcd/ca.pem \\
--peer-trusted-ca-file=/etc/etcd/ca.pem \\
I | mvcc: restore compact to 475629 --peer-client-cert-auth \\
I | etcdserver/membership: added member 5e89ccdfe3 --client-cert-auth \\
[https://192.168.5.12:2380] to cluster 894c7131f5165a78 --initial-advertise-peer-urls https://${INTERNAL_I
I | etcdserver/membership: added member c8246cee7c --listen-peer-urls https://${INTERNAL_IP}:2380 \\
[https://192.168.5.11:2380] to cluster 894c7131f5165a78 --listen-client-urls https://${INTERNAL_IP}:2379,h
--advertise-client-urls https://${INTERNAL_IP}:237
systemctl daemon-reload --initial-cluster-token etcd-cluster-1
etcd-cluster-0 \ \
--initial-cluster controller-0=https://${CONTROLLE
--initial-cluster-state new \\
service etcd restart
--data-dir=/var/lib/etcd
--data-dir=/var/lib/etcd-from-backup
Service etcd restarted
ETCDCTL_API=3 etcdctl \
Restore - ETCD ls
snapshot save snapshot.db
snapshot.db
ETCDCTL_API=3 etcdctl \
snapshot restore snapshot.db \
--data-dir /var/lib/etcd-from-backup \
--initial-cluster master-1=https://192.168.5.11:2380,master-2=https://192.168.5.12:2380 \
--initial-cluster-token etcd-cluster-1 \
--initial-advertise-peer-urls https://${INTERNAL_IP}:2380
I | mvcc: restore compact to 475629
I | etcdserver/membership: added member 5e89ccdfe3 [https://192.168.5.12:2380] to cluster 894c7131f5165a78
I | etcdserver/membership: added member c8246cee7c [https://192.168.5.11:2380] to cluster 894c7131f5165a78
--endpoints=https://127.0.0.1:2379 \
--cacert=/etc/etcd/ca.crt \
--cert=/etc/etcd/etcd-server.crt \
--key=/etc/etcd/etcd-server.key
51
KUBERNETES
ARCHITECTURE
52
Master Worker Nodes
Manage, Plan, Schedule, Monitor Host Application as Containers
Nodes
ETCD
CLUSTER
Kubernetes Architecture
Master Worker Nodes
Manage, Plan, Schedule, Monitor Host Application as Containers
Nodes
ETCD kube-
Kube-proxy
CLUSTER apiserver
kube-scheduler
Container Runtime Engine
kubelet
Run containers
Kube
Controller
Kube-proxy
Manager