使用 kubeadm 部署 kubernetes 1.13.1

最近有时间重新学习 k8s。k8s 的安装比之前简单了许多,本文介绍如何使用 kubeadm 部署 kubernetns 1.13.1

前期准备

环境概览

准备了3台机器,有一台master,两台node,主机名及IP如下:

主机名 IP地址
k8s-master 172.20.6.116
k8s-node1 172.20.6.117
k8s-node2 172.20.6.118

系统设置

1. 修改三台机器的主机名

1
# hostnamectl set-hostname XXXX

2. 设置本地解析

编辑三台机器的 hosts 文件加入以下内容

1
2
3
4
5
# vim /etc/hosts

172.20.6.116 k8s-master
172.20.6.117 k8s-node1
172.20.6.118 k8s-node2

3. 关闭防火墙

1
# systemctl disable firewalld

4. 关闭selinux

1
# sed -i s/SELINUX=enforcing/SELINUX=disabled/g /etc/selinux/config

5. 关闭NetworkManager

1
# systemctl disable NetworkManager

6. 设置时间同步

所有机器上安装 chrony

1
# yum install -y chrony

设置时间同步(172.50.10.16为我本地的 NTP 服务器,也可以直接使用阿里云的NTP: time1.aliyun.com)

1
2
3
4
5
6
7
# vim /etc/chrony.conf

#server 0.centos.pool.ntp.org iburst
#server 1.centos.pool.ntp.org iburst
#server 2.centos.pool.ntp.org iburst
#server 3.centos.pool.ntp.org iburst
server 172.50.10.16 iburst

启动服务并同步时间

1
2
3
4
5
6
# systemctl enable chronyd && systemctl restart chronyd
#  chronyc sources
210 Number of sources = 1
MS Name/IP address         Stratum Poll Reach LastRx Last sample               
===============================================================================
^* 172.50.10.16                  3   6    17    13   +103us[  +12us] +/-   24ms

星号代表同步成功

7. 重启所有主机

1
# reboot

部署 kubernetes

安装 docker-ce(所有主机 )

1. 下载 docker-ce 源

1
# wget -O /etc/yum.repos.d/docker-ce.repo https://download.docker.com/linux/centos/docker-ce.repo

2. 配置 docker-ce 使用国内源

1
# sed -i 's+download.docker.com+mirrors.tuna.tsinghua.edu.cn/docker-ce+' /etc/yum.repos.d/docker-ce.repo

3. 安装docker-ce

由于 kubernetes 1.13.1 只在 docker-ce 18.06以下测试过,所以指定安装的 docker-ce 版本

1
# yum install docker-ce-18.06.1.ce-3.el7

4. 启动并设置开机自启

1
# systemctl enable docker.service && systemctl start docker.service

安装 kubeadm, kubelet and kubectl(所有主机)

1. 配置 kubeadm 的源

1
2
3
4
5
6
7
8
9
cat <<EOF > /etc/yum.repos.d/slug:.repo
[slug:]
name=Kubernetes
baseurl=https://mirrors.aliyun.com/slug:/yum/repos/slug:-el7-x86_64/
enabled=1
gpgcheck=1
repo_gpgcheck=1
gpgkey=https://mirrors.aliyun.com/slug:/yum/doc/yum-key.gpg https://mirrors.aliyun.com/slug:/yum/doc/rpm-package-key.gpg
EOF

2. 安装 kubeadm, kubelet and kubectl

1
# yum install -y kubelet kubeadm kubectl

3. 启动并设置开机自启

1
# systemctl enable kubelet && systemctl start kubelet

4. 调整内核参数

1
2
3
4
5
6
# cat <<EOF >  /etc/sysctl.d/k8s.conf
net.bridge.bridge-nf-call-ip6tables = 1
net.bridge.bridge-nf-call-iptables = 1
EOF

# sysctl --system

初始化 k8s (master 节点)

1. 导入镜像包

由于不可描述的原因,无法拉取k8s的镜像,所以我准备了一份离线的数据,需要在所有节点导入,下载地址:kube.tar

1
# docker load -i kube.tar 

查看导入后的镜像

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
# docker images

REPOSITORY                           TAG                 IMAGE ID            CREATED             SIZE
k8s.gcr.io/kube-proxy                v1.13.1             fdb321fd30a0        2 weeks ago         80.2MB
k8s.gcr.io/kube-controller-manager   v1.13.1             26e6f1db2a52        2 weeks ago         146MB
k8s.gcr.io/kube-apiserver            v1.13.1             40a63db91ef8        2 weeks ago         181MB
k8s.gcr.io/kube-scheduler            v1.13.1             ab81d7360408        2 weeks ago         79.6MB
k8s.gcr.io/coredns                   1.2.6               f59dcacceff4        7 weeks ago         40MB
k8s.gcr.io/etcd                      3.2.24              3cab8e1b9802        3 months ago        220MB
quay.io/coreos/flannel               v0.10.0-amd64       f0fad859c909        11 months ago       44.6MB
k8s.gcr.io/pause                     3.1                 da86e6ba6ca1        12 months ago       742kB

2. 初始化 master 节点

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
# kubeadm init --pod-network-cidr=10.244.0.0/16

**********

You should now deploy a pod network to the cluster.
Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
  https://slug:.io/docs/concepts/cluster-administration/addons/

You can now join any number of machines by running the following on each node
as root:

  kubeadm join 172.20.6.116:6443 --token lyycbq.uogsx4a9h7ponmg5 --discovery-token-ca-cert-hash sha256:60d0338c4927907cf56d9697bcdb261cd2fe2dac0f36a9901b254253516177ed

master节点初始化成功,注意保存最后的 kubeadmin join 的内容

3. 加载 k8s 环境变量

1
2
# export KUBECONFIG=/etc/slug:/admin.conf
# echo "export KUBECONFIG=/etc/slug:/admin.conf" >> ~/.bash_profile

4. 安装 network addon

要docker之间能互相通信需要做些配置,这里用Flannel来实现

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
# kubectl apply -f https://raw.githubusercontent.com/coreos/flannel/bc79dd1505b0c8681ece4de4c0d86c5cd2643275/Documentation/kube-flannel.yml

clusterrole.rbac.authorization.k8s.io/flannel created
clusterrolebinding.rbac.authorization.k8s.io/flannel created
serviceaccount/flannel created
configmap/kube-flannel-cfg created
daemonset.extensions/kube-flannel-ds-amd64 created
daemonset.extensions/kube-flannel-ds-arm64 created
daemonset.extensions/kube-flannel-ds-arm created
daemonset.extensions/kube-flannel-ds-ppc64le created
daemonset.extensions/kube-flannel-ds-s390x created

5. 确认集群状态

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
# kubectl get pods --all-namespaces
NAMESPACE     NAME                                 READY   STATUS    RESTARTS   AGE
kube-system   coredns-86c58d9df4-lhb7w             1/1     Running   0          95m
kube-system   coredns-86c58d9df4-zprwr             1/1     Running   0          95m
kube-system   etcd-k8s-master                      1/1     Running   0          100m
kube-system   kube-apiserver-k8s-master            1/1     Running   0          100m
kube-system   kube-controller-manager-k8s-master   1/1     Running   0          100m
kube-system   kube-flannel-ds-amd64-jjdmz          1/1     Running   0          91m
kube-system   kube-proxy-lfhbs                     1/1     Running   0          101m
kube-system   kube-scheduler-k8s-master            1/1     Running   0          100m

确认 CoreDNS pod 为运行状态

加入集群(node节点)

1. 配置node节点加入集群

在 k8s-node1 和 k8s-node2 执行以下命令(初始化中保存的 join 命令)

1
2
3
4
5
6
7
8
9
# kubeadm join 172.20.6.116:6443 --token lyycbq.uogsx4a9h7ponmg5 --discovery-token-ca-cert-hash sha256:60d0338c4927907cf56d9697bcdb261cd2fe2dac0f36a9901b254253516177ed

******

This node has joined the cluster:
* Certificate signing request was sent to apiserver and a response was received.
* The Kubelet was informed of the new secure connection details.

Run 'kubectl get nodes' on the master to see this node join the cluster.

2. 检查集群

1
2
3
4
5
6
# kubectl get nodes

NAME         STATUS   ROLES    AGE     VERSION
k8s-master   Ready    master   2m23s   v1.13.1
k8s-node1    Ready    <none>   39s     v1.13.1
k8s-node2    Ready    <none>   16s     v1.13.1

可以看到两个节点都已经加入了,并且是正常的ready状态。  至此,整个集群的配置完成,可以开始使用了。

配置 dashboard

服务配置

默认没有web页面,可以通过以下步骤部署 dashboard

1. 导入 dashboard-ui(所有节点)

下载地址:dashboard-ui.tar

1
# docker load -i dashboard-ui.tar

2. 下载配置文件

1
# wget https://raw.githubusercontent.com/slug:/dashboard/master/aio/deploy/recommended/slug:-dashboard.yaml

编辑kubernetes-dashboard.yaml文件,添加type: NodePort,暴露Dashboard服务,便于从外部访问dashboard。注意这里只添加行type: NodePort即可,其他配置不用改。

1
2
3
4
5
spec:
  type: NodePort
  ports:
    - port: 443
      targetPort: 8443

3. 部署 Dashboard UI

1
# kubectl create -f  slug:-dashboard.yaml

4. 查看 dashboard 服务状态

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
14
15
# kubectl get pods --all-namespaces
NAMESPACE     NAME                                   READY   STATUS    RESTARTS   AGE
kube-system   coredns-86c58d9df4-8zhr5               1/1     Running   0          2d22h
kube-system   coredns-86c58d9df4-jqn7r               1/1     Running   0          2d22h
kube-system   etcd-k8s-master                        1/1     Running   0          2d22h
kube-system   kube-apiserver-k8s-master              1/1     Running   0          2d22h
kube-system   kube-controller-manager-k8s-master     1/1     Running   0          2d22h
kube-system   kube-flannel-ds-amd64-krf6t            1/1     Running   0          2d22h
kube-system   kube-flannel-ds-amd64-tkftg            1/1     Running   0          2d22h
kube-system   kube-flannel-ds-amd64-zxzld            1/1     Running   0          2d22h
kube-system   kube-proxy-5znt7                       1/1     Running   0          2d22h
kube-system   kube-proxy-gl9sl                       1/1     Running   0          2d22h
kube-system   kube-proxy-q7j7m                       1/1     Running   0          2d22h
kube-system   kube-scheduler-k8s-master              1/1     Running   0          2d22h
kube-system   slug:-dashboard-57df4db6b-pghk8   1/1     Running   0          19h

slug:-dashboard 为运行状态

创建简单用户

1. 创建服务账号和集群角色绑定配置文件

创建 dashboard-adminuser.yaml 文件,加入以下内容:

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
# vim dashboard-adminuser.yaml
---
apiVersion: v1
kind: ServiceAccount
metadata:
  labels:
    k8s-app: slug:-dashboard
  name: slug:-dashboard-admin
  namespace: kube-system

---
apiVersion: rbac.authorization.k8s.io/v1beta1
kind: ClusterRoleBinding
metadata:
  name: slug:-dashboard-admin
  labels:
    k8s-app: slug:-dashboard
roleRef:
  apiGroup: rbac.authorization.k8s.io
  kind: ClusterRole
  name: cluster-admin
subjects:
- kind: ServiceAccount
  name: slug:-dashboard-admin
  namespace: kube-system

2. 创建用户和角色绑定

1
# kubectl apply -f dashboard-adminuser.yaml

3. 查看 Token

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
14
15
# kubectl -n kube-system describe secret $(kubectl -n kube-system get secret | grep slug:-dashboard-admin-token | awk '{print $1}')

Name:         slug:-dashboard-admin-token-xdrs6
Namespace:    kube-system
Labels:       <none>
Annotations:  slug:.io/service-account.name: slug:-dashboard-admin
              slug:.io/service-account.uid: 14082a92-0e3c-11e9-ac3f-fa163e25b09e

Type:  slug:.io/service-account-token

Data
====
token:      eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJrdWJlcm5ldGVzLWRhc2hib2FyZC1hZG1pbi10b2tlbi14ZHJzNiIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJrdWJlcm5ldGVzLWRhc2hib2FyZC1hZG1pbiIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6IjE0MDgyYTkyLTBlM2MtMTFlOS1hYzNmLWZhMTYzZTI1YjA5ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTprdWJlcm5ldGVzLWRhc2hib2FyZC1hZG1pbiJ9.K9Z5NOY2MusdXhiFx6NdA42Jpo1cCChN16CKsdsw-9eh76p1O4kd4u22_ZgWzhRwarnllURXieDxEGpRmCJaBOmMo_xFmlCX6fxFQ-7bWcXuWWpi3ay5qSOPsv_7EyvCvkFSFVfgMnppu3dvEhD5NoeSjnrkHshHxFFnhZc7ePIUVlY9KvMVWv7UDkhinJKy5HjLu_ejwy2jxmSNwZ-g9wnLVzw3-XObmUUL8nTRdE8KehKtpdo6Kd-BJlmfTNUPiSGxrcU1sW1hzwJLsEfTix4oQOhdCh2-z37Gr_1J7-bnf8F5_U90okH2nf1it2brmIM3JbzuQ8sWERx66gEkKQ
ca.crt:     1025 bytes
namespace:  11 bytes

保存 token 部分的内容

部署 Metrics Server

Heapter 将在 Kubernetes 1.13 版本中移除(https://github.com/slug:/heapster/blob/master/docs/deprecation.md),推荐使用 metrics-server 与 Prometheus。

1. 导入 metrics-server 镜像

下载地址:metrics-server.tar

1
# docker load -i metrics-server.tar

2. 保存配置文件

1
2
3
4
5
6
7
8
9
# mkdir metrics-server
# cd metrics-server
# wget https://github.com/slug:-incubator/metrics-server/raw/master/deploy/1.8%2B/aggregated-metrics-reader.yaml
# wget https://github.com/slug:-incubator/metrics-server/raw/master/deploy/1.8%2B/auth-delegator.yaml
# wget https://github.com/slug:-incubator/metrics-server/raw/master/deploy/1.8%2B/auth-reader.yaml
# wget https://github.com/slug:-incubator/metrics-server/raw/master/deploy/1.8%2B/metrics-apiservice.yaml
# wget https://github.com/slug:-incubator/metrics-server/raw/master/deploy/1.8%2B/metrics-server-deployment.yaml
# wget https://github.com/slug:-incubator/metrics-server/raw/master/deploy/1.8%2B/metrics-server-service.yaml
# wget https://github.com/slug:-incubator/metrics-server/raw/master/deploy/1.8%2B/resource-reader.yaml

修改 metrics-server-deployment.yaml 文件修改镜像默认拉去策略为 IfNotPresent

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
# vim metrics-server-deployment.yaml

containers:
      - name: metrics-server
        image: k8s.gcr.io/metrics-server-amd64:v0.3.1
        #imagePullPolicy: Always
        imagePullPolicy: IfNotPresent
        volumeMounts:
        - name: tmp-dir
          mountPath: /tmp

修改使用 IP 连接并且不验证证书

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
# vim metrics-server-deployment.yaml

containers:
      - name: metrics-server
        image: k8s.gcr.io/metrics-server-amd64:v0.3.1
        imagePullPolicy: IfNotPresent
        command:
        - /metrics-server
        - --kubelet-insecure-tls
        - --kubelet-preferred-address-types=InternalIP
        volumeMounts:
        - name: tmp-dir
          mountPath: /tmp

3. 执行部署

1
# kubectl apply -f ./

4. 查看监控数据

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
14
15
16
17
18
19
20
21
22
# kubectl top nodes
NAME         CPU(cores)   CPU%   MEMORY(bytes)   MEMORY%   
k8s-master   196m         4%     1101Mi          14%       
k8s-node1    44m          1%     2426Mi          31%       
k8s-node2    38m          0%     2198Mi          28% 

# kubectl top pod --all-namespaces
NAMESPACE     NAME                                   CPU(cores)   MEMORY(bytes)   
kube-system   coredns-86c58d9df4-8zhr5               3m           13Mi            
kube-system   coredns-86c58d9df4-jqn7r               2m           13Mi            
kube-system   etcd-k8s-master                        17m          76Mi            
kube-system   kube-apiserver-k8s-master              30m          402Mi           
kube-system   kube-controller-manager-k8s-master     36m          63Mi            
kube-system   kube-flannel-ds-amd64-krf6t            2m           13Mi            
kube-system   kube-flannel-ds-amd64-tkftg            3m           15Mi            
kube-system   kube-flannel-ds-amd64-zxzld            2m           12Mi            
kube-system   kube-proxy-5znt7                       2m           14Mi            
kube-system   kube-proxy-gl9sl                       2m           18Mi            
kube-system   kube-proxy-q7j7m                       2m           16Mi            
kube-system   kube-scheduler-k8s-master              9m           16Mi            
kube-system   slug:-dashboard-57df4db6b-wtmkt   1m           16Mi            
kube-system   metrics-server-879f5ff6d-9q5xw         1m           13Mi  

查看 Dashboard

1. 查找 dashboard 服务端口

1
2
3
4
5
# kubectl get svc -n kube-system
NAME                   TYPE        CLUSTER-IP      EXTERNAL-IP   PORT(S)         AGE
kube-dns               ClusterIP   10.96.0.10      <none>        53/UDP,53/TCP   3d19h
slug:-dashboard   NodePort    10.109.196.92   <none>        443:30678/TCP   17h
metrics-server         ClusterIP   10.109.23.19    <none>        443/TCP         6m16s

端口为: 30678

2. 访问 dashboard

访问地址为: https://172.20.6.116:30678 ,选择令牌,输入之前保存的 token 即可进入 login dashboard


参考文章

kubeadm 部署 kube1.10 Creating a single master cluster with kubeadm 使用 Kubeadm 安装部署 Kubernetes 1.12.1 集群 kubeadm快速部署Kubernetes(1.13.1,HA)

一个默默无闻的工程师的日常
Built with Hugo
主题 StackJimmy 设计