• Kubernates集群搭建


    1、Kubernates集群组件

      - etcd:在Master节点上,一个高可用的K/V键值对存储和服务发现系统

      - kube-apiserver:在Master节点上,实现跨主机的容器网络通信

      - kube-controller-manager:在Master节点上,确保集群服务

      - kube-scheduler:在Master节点上,调度容器,分配到Node节点上

      

      - kube-proxy:在Node节点上,提供网络代理服务

      - kubelet:在Node节点上,按照配置文件中定义的容器规格启动容器

      - flannel:在Node节点上,实现跨主机的容器网络的通信

    2、集群示意图

    3、环境安装

    • 安装epel-release源
    yum install -y epel-release
    • 关闭防火墙和selinx(与docker容器的放火请规则冲突)
    systemctl stop firewalld
    systemctl disable firewalld
    setenforce 0

    4、安装Master节点

    • 安装etcd和kubernates-master
    yum -y install etcd kubernetes-master
    • 编辑/etc/etcd/etcd.conf文件,设置etcd监听客户端的URLS
    ETCD_NAME=default
    ETCD_DATA_DIR="/var/lib/etcd/default.etcd"
    ETCD_LISTEN_CLIENT_URLS="http://0.0.0.0:2379"
    ETCD_ADVERTISE_CLIENT_URLS="http://localhost:2379"
    • 配置ServiceAccount,用于Master使用yaml文件在Node上生成Pod
    openssl genrsa -out /etc/kubernetes/serviceaccount.key 2048
    • 编辑/etc/kubernetes/apiserver文件,设置kube-api的地址和端口和ServiceAccount
    KUBE_API_ADDRESS="--insecure-bind-address=0.0.0.0"
    KUBE_API_PORT="--port=8080"
    KUBELET_PORT="--kubelet-port=10250"
    KUBE_ETCD_SERVERS="--etcd-servers=http://127.0.0.1:2379"
    KUBE_SERVICE_ADDRESSES="--service-cluster-ip-range=10.254.0.0/16"
    KUBE_ADMISSION_CONTROL="--admission-control=NamespaceLifecycle,NamespaceExists,LimitRanger,SecurityContextDeny,ResourceQuota"
    KUBE_API_ARGS="--service_account_key_file=/etc/kubernetes/serviceaccount.key"
    • 编辑/etc/kubernetes/controller-manager文件,配置ServiceAccount
    KUBE_CONTROLLER_MANAGER_ARGS="--service_account_private_key_file=/etc/kubernetes/serviceaccount.key"
    • 启动etcd、kube-apiserver、kube-controller-manager、kube-scheduler等服务,并设置开机启动
    for SERVICES in etcd kube-apiserver kube-controller-manager kube-scheduler; 
        do 
            systemctl restart $SERVICES;
            systemctl enable $SERVICES;
            systemctl status $SERVICES ; 
        done
    • 在etcd中定义flannel网络
    etcdctl mk /atomic.io/network/config '{"Network":"172.17.0.0/16"}'

    5、安装Node节点

    • 安装flannel和kubernetes-node
    yum -y install flannel kubernetes-node
    • 修改/etc/sysconfig/flanneld文件,为flannel网络指定etcd服务
    FLANNEL_ETCD="http://192.168.30.20:2379"
    FLANNEL_ETCD_KEY="/atomic.io/network"
    • 修改/etc/kubernetes/config文件,指定kube master地址
    KUBE_LOGTOSTDERR="--logtostderr=true"
    KUBE_LOG_LEVEL="--v=0"
    KUBE_ALLOW_PRIV="--allow-privileged=false"
    KUBE_MASTER="--master=http://192.168.30.20:8080"
    • 修改/etc/kubernetes/kubelet文件,配置相应Node节点的信息
      • node1
    KUBELET_ADDRESS="--address=0.0.0.0"
    KUBELET_PORT="--port=10250"
    KUBELET_HOSTNAME="--hostname-override=192.168.30.21" #修改成对应Node的IP
    KUBELET_API_SERVER="--api-servers=http://192.168.30.20:8080" #指定Master节点的API Server
    KUBELET_POD_INFRA_CONTAINER="--pod-infra-container-image=registry.access.redhat.com/rhel7/pod-infrastructure:latest"
    KUBELET_ARGS=""
      • node2
    KUBELET_ADDRESS="--address=0.0.0.0"
    KUBELET_PORT="--port=10250"
    KUBELET_HOSTNAME="--hostname-override=192.168.30.22" 
    KUBELET_API_SERVER="--api-servers=http://192.168.30.20:8080"
    KUBELET_POD_INFRA_CONTAINER="--pod-infra-container-image=registry.access.redhat.com/rhel7/pod-infrastructure:latest"
    KUBELET_ARGS=""
      • node3
    KUBELET_ADDRESS="--address=0.0.0.0"
    KUBELET_PORT="--port=10250"
    KUBELET_HOSTNAME="--hostname-override=192.168.30.23"
    KUBELET_API_SERVER="--api-servers=http://192.168.30.20:8080"
    KUBELET_POD_INFRA_CONTAINER="--pod-infra-container-image=registry.access.redhat.com/rhel7/pod-infrastructure:latest"
    KUBELET_ARGS=""
      • node4
    KUBELET_ADDRESS="--address=0.0.0.0"
    KUBELET_PORT="--port=10250"
    KUBELET_HOSTNAME="--hostname-override=192.168.30.24"
    KUBELET_API_SERVER="--api-servers=http://192.168.30.20:8080"
    KUBELET_POD_INFRA_CONTAINER="--pod-infra-container-image=registry.access.redhat.com/rhel7/pod-infrastructure:latest"
    KUBELET_ARGS=""
    • 启动kube-proxy,kubelet,docker,flanneld等服务,并设置开机启动
    for SERVICES in kube-proxy kubelet docker flanneld;
        do
            systemctl restart $SERVICES;
            systemctl enable $SERVICES;
            systemctl status $SERVICES;
        done
    • 安装rhsm,解决Pod一直处于ContainerCreating状态
    wget http://mirror.centos.org/centos/7/os/x86_64/Packages/python-rhsm-certificates-1.19.10-1.el7_4.x86_64.rpm
    rpm2cpio python-rhsm-certificates-1.19.10-1.el7_4.x86_64.rpm | cpio -iv --to-stdout ./etc/rhsm/ca/redhat-uep.pem | tee /etc/rhsm/ca/redhat-uep.pem

    6、验证集群是否安装成功(在Master上执行)

    kubectl get node

     文章来自于:韩德田Tivens-k8s入门系列之集群安装篇

  • 相关阅读:
    Hyperledger Fabric 1.0 从零开始(七)——启动Fabric多节点集群
    Hyperledger Fabric 1.0 从零开始(六)——创建Fabric多节点集群
    Hyperledger Fabric 1.0 从零开始(五)——运行测试e2e
    Hyperledger Fabric 1.0 从零开始(四)——Fabric源码及镜像文件处理
    Hyperledger Fabric 1.0 从零开始(三)——环境构建(内网/准离线)
    Hyperledger Fabric 1.0 从零开始(二)——环境构建(公网)
    Hyperledger Fabric 1.0 从零开始(一)
    Springboot 多数据源配置,结合tk-mybatis
    使用mybatis-generator插件结合tk.mybatis自动生成mapper二三事
    使用IDEA创建一个springboot项目
  • 原文地址:https://www.cnblogs.com/pascall/p/10499862.html
Copyright © 2020-2023  润新知