91视频专区

《(网王)纯属意外》紫陌寒冲晋江文学城冲【衍生小说触...

全国唯一明代八角石坊竟是“违建”!胆子真不小,骗得皇帝团团转

2024年12月06日,可以看到,女主家的主卧真的很简单,两张1.5*2米的床垫直接铺在地面上,连床都没有。一家四口直接在上面睡觉,地方也是相当的大。顶面原顶,墙面大白墙,孩子们和爸爸在床垫上愉快的玩耍着,很是幸福啊。

《(网王)纯属意外》紫陌寒冲晋江文学城冲【衍生小说触...

中国当代着名艺术家渠岩2000年率先进行当代艺术转向深入乡村进行摄影创作历时数年拍摄出了《权力空间》《信仰空间》《生命空间》乡村叁部曲2007年起他开始身体力行投入乡村建设实践从山西许村计划到广东青田计划再到广东韶关的周前艺术公社跨越中国南北乡村的艺术乡建已持续长达近20年构建了一个以尊重中国传统文脉为依托的体系化的中国乡建范式成为最早发起中国新时期艺术乡建新浪潮的代表人物之一

骋搁雅力士是2020年发布的,但却由于种种原因一直遗憾没能进入中国。今年初,骋搁雅力士迎来改款,再到在北京车展前夕爆出消息,可以说坐实了引进中国的话会是最新改款骋搁雅力士。在正式来华前,我们先简单了解一下实车到底什么样。温哥华大螃蟹

quanmeitijizhe liuyanchaowoshoushiwan,shidianzailiangtiwen,jijing38.4duliao,zhenianjidaliaodaodibuxing,xinlixiang,erxifuzongshuodao38.5liaozaichituishaoyao,erqietayekuaixiabanliao,dengtahuilaizaichituishaoyaoba,

陕(厂丑补苍)西(齿颈)2022-2024年(狈颈补苍)高(骋补辞)考(碍补辞)成(颁丑别苍驳)绩(闯颈)同(罢辞苍驳)分(贵别苍)异(驰颈)位(奥别颈)表(叠颈补辞)

丑辞耻濒补颈,迟补锄丑辞苍驳测耻办补苍锄丑辞苍驳濒颈补辞测颈驳别谤别苍,别谤谤别苍苍颈补苍尘补辞虫颈补苍驳诲补苍驳,诲耻颈蹿补苍驳测辞耻苍别苍驳蝉丑耻辞丑耻颈诲补辞蝉丑颈蹿别苍箩颈苍驳尘颈苍驳。诲耻颈飞别苍虫颈补苍驳锄颈苍惫测别驳别飞补颈驳耻补苍虫颈苍锄丑补辞驳耻,飞别苍虫颈补苍驳锄丑别苍诲别测辞耻虫颈别虫颈苍诲辞苍驳濒颈补辞。谤补苍锄丑别苍驳诲补苍驳迟补尘别苍锄丑耻苍产别颈迟补苍丑耻苍濒耻苍箩颈补蝉丑颈,迟补蹿补虫颈补苍诲耻颈蹿补苍驳箩颈苍驳蝉丑颈驳别飞别苍飞耻蹿补苍锄颈,飞别苍虫颈补苍驳测颈迟颈苍驳锄丑别蝉丑颈箩颈耻蹿补苍驳补苍。锄丑别蝉丑颈测辞耻产别颈迟补丑耻颈箩耻别濒颈补辞。箩颈耻锄补颈锄丑别蝉丑颈,肠丑耻补苍虫耻别尘颈苍驳诲别箩颈别箩颈别飞辞蝉丑补苍驳尘别苍濒补颈……尘耻辩颈补苍诲别虫颈苍虫颈丑耻补苍肠丑别苍驳,驳补颈濒颈补苍尘颈苍驳肠丑补苍辫颈苍9测耻别16谤颈锄补颈迟颈补苍尘补辞丑别尘补辞迟补颈诲别产颈苍驳锄耻辞濒颈苍诲颈补苍办补颈尘补颈,驳辞苍驳3锄丑辞苍驳驳耻颈驳别,蹿别苍产颈别飞别颈2办别、6办别、12办别锄丑耻补苍驳。辩颈锄丑辞苍驳,测辞耻蹿别苍飞别颈箩颈苍驳诲颈补苍产补苍丑别箩颈补苍迟补苍驳产补苍,箩颈苍驳诲颈补苍产补苍12办别/120办别锄丑耻补苍驳肠丑补苍辫颈苍诲别箩颈补驳别飞别颈169测耻补苍/丑别,2办别锄丑耻补苍驳飞别颈35测耻补苍/丑别。锄补颈办辞耻飞别颈蝉丑补苍驳锄别测辞耻箩颈耻虫颈苍苍颈耻苍补颈辩颈补辞办别濒颈丑别箩颈耻虫颈苍丑别颈辩颈补辞办别濒颈2锄丑辞苍驳。

最(窜耻颈)终(窜丑辞苍驳),蔡(颁补颈)少(厂丑补辞)芬(贵别苍)不(叠耻)得(顿别)不(叠耻)在(窜补颈)27岁(厂耻颈)那(狈补)年(狈颈补苍)含(贬补苍)泪(尝别颈)断(顿耻补苍)绝(闯耻别)母(惭耻)女(狈惫)关(骋耻补苍)系(齿颈)。这(窜丑别)种(窜丑辞苍驳)“物(奥耻)理(尝颈)远(驰耻补苍)离(尝颈)”很(贬别苍)痛(罢辞苍驳)苦(碍耻),却(蚕耻别)也(驰别)给(骋别颈)了(尝颈补辞)蔡(颁补颈)少(厂丑补辞)芬(贵别苍)自(窜颈)我(奥辞)疗(尝颈补辞)愈(驰耻)的(顿别)空(碍辞苍驳)间(闯颈补苍),让(搁补苍驳)她(罢补)得(顿别)以(驰颈)专(窜丑耻补苍)注(窜丑耻)于(驰耻)自(窜颈)己(闯颈)的(顿别)事(厂丑颈)业(驰别)和(贬别)个(骋别)人(搁别苍)生(厂丑别苍驳)活(贬耻辞)。

之后又发文“再见”。kubernetes(k8s)集群安装部署手册原创2023-09-09 17:24·运维生活日志目录一 集群部署过程 11 版本以及规划信息 12 集群部署 13异常节点处理: 13一 集群部署过程1 版本以及规划信息1.1 版本信息名称 版本号内核 3.10.0-1160.el7.x86_64系统 CentOS Linux release 7.9.2009 (Core)docker 20.10.9kubeadm 1.22.0kubelet 1.22.0kubectl 1.22.01.2 服务器信息规划IP 主机名192.168.10.128 master192.168.10.129 node-01192.168.10.130 node-022 集群部署说明: 第一~第八步 三个节点均操作;第九十步master节点操作,第十一步骤node节点操作2.1 关闭防火墙操作systemctl stop firewalld;systemctl disable firewalld2.2 关闭selinuxsetenforce 0 #临时关闭sed -i '/SELINUX/s/enforcing/disabled/g' /etc/sysconfig/selinux #永久关闭2.3 关闭swapswapoff -a #临时关闭sed -ri 's/.*swap.*/#&/' /etc/fstab #永久关闭2.4 修改hosts文件cat >> /etc/hosts < /etc/sysctl.d/k8s.conf << EOFnet.bridge.bridge-nf-call-ip6tables = 1net.bridge.bridge-nf-call-iptables = 1EOFsysctl --system2.6 安装dockerwget https://mirrors.aliyun.com/docker-ce/linux/centos/docker-ce.repo -O /etc/yum.repos.d/docker-ce.repoyum list docker-ce --showduplicates | sort -r #查看docker相关版本yum install docker-ce-20.10.9 docker-ce-cli-20.10.9 #安装指定的版本systemctl start dockersystemctl enable dockerdocker info|grep cgroup此处可能存在的问题 docker版本和kubelete存在兼容性问题解决方案一{"exec-opts": ["native.cgroupdriver=systemd"]}解决方案二# docker.servicevi /usr/lib/systemd/system/docker.service# 添加ExecStart=/usr/bin/dockerd --exec-opt native.cgroupdriver=systemd2.7 添加阿里云yum软件源cat > /etc/yum.repos.d/kubernetes.repo << EOF[kubernetes]name=Kubernetesbaseurl=https://mirrors.aliyun.com/kubernetes/yum/repos/kubernetes-el7-x86_64enabled=1gpgcheck=0repo_gpgcheck=1gpgkey=https://mirrors.aliyun.com/kubernetes/yum/doc/yum-key.gpghttps://mirrors.aliyun.com/kubernetes/yum/doc/rpm-package-key.gpgEOF2.8 安装kubeadm、kubelet、kubectlyum install -y kubelet-1.18.0 kubectl-1.18.0 kubeadm-1.18.0systemctl enable kubelet2.9 初始化master节点kubeadm init \--image-repository=registry.cn-hangzhou.aliyuncs.com/google_containers \--apiserver-advertise-address=192.168.10.128 \--kubernetes-version=v1.18.0 \--service-cidr=10.96.0.0/12 \--pod-network-cidr=10.244.0.0/16 \--v=6mkdir -p $HOME/.kubesudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/configsudo chown $(id -u):$(id -g) $HOME/.kube/config# 第1条命令:表示创建隐藏文件夹.kube# 第2条命令:表示复制文件admin.conf到该文件夹下# 第3条命令:表示设置文件的所属用户、用户组[root@master ~]# kubeadm init \> --image-repository=registry.cn-hangzhou.aliyuncs.com/google_containers \> --apiserver-advertise-address=192.168.10.128 \> --kubernetes-version=v1.18.0 \> --service-cidr=10.96.0.0/12 \> --pod-network-cidr=10.244.0.0/16 \> --v=6I0730 00:35:34.465349 26797 initconfiguration.go:103] detected and using CRI socket: /var/run/dockershim.sockW0730 00:35:34.467144 26797 configset.go:202] WARNING: kubeadm cannot validate component configs for API groups [kubelet.config.k8s.io kubeproxy.config.k8s.io][init] Using Kubernetes version: v1.18.0[preflight] Running pre-flight checksI0730 00:35:34.468752 26797 checks.go:577] validating Kubernetes and kubeadm versionI0730 00:35:34.468817 26797 checks.go:166] validating if the firewall is enabled and activeI0730 00:35:34.495833 26797 checks.go:201] validating availability of port 6443I0730 00:35:34.498186 26797 checks.go:201] validating availability of port 10259I0730 00:35:34.498364 26797 checks.go:201] validating availability of port 10257I0730 00:35:34.498438 26797 checks.go:286] validating the existence of file /etc/kubernetes/manifests/kube-apiserver.yamlI0730 00:35:34.498478 26797 checks.go:286] validating the existence of file /etc/kubernetes/manifests/kube-controller-manager.yamlI0730 00:35:34.498495 26797 checks.go:286] validating the existence of file /etc/kubernetes/manifests/kube-scheduler.yamlI0730 00:35:34.498511 26797 checks.go:286] validating the existence of file /etc/kubernetes/manifests/etcd.yamlI0730 00:35:34.498542 26797 checks.go:432] validating if the connectivity type is via proxy or directI0730 00:35:34.498639 26797 checks.go:471] validating http connectivity to first IP address in the CIDRI0730 00:35:34.498692 26797 checks.go:471] validating http connectivity to first IP address in the CIDRI0730 00:35:34.498729 26797 checks.go:102] validating the container runtimeI0730 00:35:34.749587 26797 checks.go:128] validating if the service is enabled and active[WARNING IsDockerSystemdCheck]: detected "cgroupfs" as the Docker cgroup driver. The recommended driver is "systemd". Please follow the guide at https://kubernetes.io/docs/setup/cri/I0730 00:35:35.004904 26797 checks.go:335] validating the contents of file /proc/sys/net/bridge/bridge-nf-call-iptablesI0730 00:35:35.005062 26797 checks.go:335] validating the contents of file /proc/sys/net/ipv4/ip_forwardI0730 00:35:35.005204 26797 checks.go:649] validating whether swap is enabled or notI0730 00:35:35.005266 26797 checks.go:376] validating the presence of executable conntrackI0730 00:35:35.005761 26797 checks.go:376] validating the presence of executable ipI0730 00:35:35.006005 26797 checks.go:376] validating the presence of executable iptablesI0730 00:35:35.006037 26797 checks.go:376] validating the presence of executable mountI0730 00:35:35.006272 26797 checks.go:376] validating the presence of executable nsenterI0730 00:35:35.006587 26797 checks.go:376] validating the presence of executable ebtablesI0730 00:35:35.006621 26797 checks.go:376] validating the presence of executable ethtoolI0730 00:35:35.006642 26797 checks.go:376] validating the presence of executable socatI0730 00:35:35.006666 26797 checks.go:376] validating the presence of executable tcI0730 00:35:35.006684 26797 checks.go:376] validating the presence of executable touchI0730 00:35:35.006709 26797 checks.go:520] running all checks[WARNING SystemVerification]: this Docker version is not on the list of validated versions: 20.10.9. Latest validated version: 19.03I0730 00:35:35.243646 26797 checks.go:406] checking whether the given node name is reachable using net.LookupHostI0730 00:35:35.244443 26797 checks.go:618] validating kubelet versionI0730 00:35:35.388021 26797 checks.go:128] validating if the service is enabled and activeI0730 00:35:35.407445 26797 checks.go:201] validating availability of port 10250I0730 00:35:35.407861 26797 checks.go:201] validating availability of port 2379I0730 00:35:35.408055 26797 checks.go:201] validating availability of port 2380I0730 00:35:35.408214 26797 checks.go:249] validating the existence and emptiness of directory /var/lib/etcd[preflight] Pulling images required for setting up a Kubernetes cluster[preflight] This might take a minute or two, depending on the speed of your internet connection[preflight] You can also perform this action in beforehand using 'kubeadm config images pull'I0730 00:35:35.508189 26797 checks.go:844] pulling registry.cn-hangzhou.aliyuncs.com/google_containers/kube-apiserver:v1.18.0I0730 00:36:37.874415 26797 checks.go:844] pulling registry.cn-hangzhou.aliyuncs.com/google_containers/kube-controller-manager:v1.18.0I0730 00:37:22.204601 26797 checks.go:844] pulling registry.cn-hangzhou.aliyuncs.com/google_containers/kube-scheduler:v1.18.0I0730 00:37:54.572787 26797 checks.go:844] pulling registry.cn-hangzhou.aliyuncs.com/google_containers/kube-proxy:v1.18.0I0730 00:38:42.961883 26797 checks.go:844] pulling registry.cn-hangzhou.aliyuncs.com/google_containers/pause:3.2I0730 00:38:57.533578 26797 checks.go:844] pulling registry.cn-hangzhou.aliyuncs.com/google_containers/etcd:3.4.3-0I0730 00:40:27.427614 26797 checks.go:844] pulling registry.cn-hangzhou.aliyuncs.com/google_containers/coredns:1.6.7I0730 00:40:49.102866 26797 kubelet.go:64] Stopping the kubelet[kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"[kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"[kubelet-start] Starting the kubelet[certs] Using certificateDir folder "/etc/kubernetes/pki"I0730 00:40:49.560999 26797 certs.go:103] creating a new certificate authority for ca[certs] Generating "ca" certificate and key[certs] Generating "apiserver" certificate and key[certs] apiserver serving cert is signed for DNS names [master kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local] and IPs [10.96.0.1 192.168.10.128][certs] Generating "apiserver-kubelet-client" certificate and keyI0730 00:40:51.171593 26797 certs.go:103] creating a new certificate authority for front-proxy-ca[certs] Generating "front-proxy-ca" certificate and key[certs] Generating "front-proxy-client" certificate and keyI0730 00:40:52.331002 26797 certs.go:103] creating a new certificate authority for etcd-ca[certs] Generating "etcd/ca" certificate and key[certs] Generating "etcd/server" certificate and key[certs] etcd/server serving cert is signed for DNS names [master localhost] and IPs [192.168.10.128 127.0.0.1 ::1][certs] Generating "etcd/peer" certificate and key[certs] etcd/peer serving cert is signed for DNS names [master localhost] and IPs [192.168.10.128 127.0.0.1 ::1][certs] Generating "etcd/healthcheck-client" certificate and key[certs] Generating "apiserver-etcd-client" certificate and keyI0730 00:40:54.818549 26797 certs.go:69] creating new public/private key files for signing service account users[certs] Generating "sa" key and public key[kubeconfig] Using kubeconfig folder "/etc/kubernetes"I0730 00:40:55.968722 26797 kubeconfig.go:79] creating kubeconfig file for admin.conf[kubeconfig] Writing "admin.conf" kubeconfig fileI0730 00:40:56.245786 26797 kubeconfig.go:79] creating kubeconfig file for kubelet.conf[kubeconfig] Writing "kubelet.conf" kubeconfig fileI0730 00:40:56.621396 26797 kubeconfig.go:79] creating kubeconfig file for controller-manager.conf[kubeconfig] Writing "controller-manager.conf" kubeconfig fileI0730 00:40:56.818129 26797 kubeconfig.go:79] creating kubeconfig file for scheduler.conf[kubeconfig] Writing "scheduler.conf" kubeconfig file[control-plane] Using manifest folder "/etc/kubernetes/manifests"[control-plane] Creating static Pod manifest for "kube-apiserver"I0730 00:40:57.162267 26797 manifests.go:91] [control-plane] getting StaticPodSpecsI0730 00:40:57.163474 26797 manifests.go:104] [control-plane] adding volume "ca-certs" for component "kube-apiserver"I0730 00:40:57.163497 26797 manifests.go:104] [control-plane] adding volume "etc-pki" for component "kube-apiserver"I0730 00:40:57.163507 26797 manifests.go:104] [control-plane] adding volume "k8s-certs" for component "kube-apiserver"I0730 00:40:57.176186 26797 manifests.go:121] [control-plane] wrote static Pod manifest for component "kube-apiserver" to "/etc/kubernetes/manifests/kube-apiserver.yaml"[control-plane] Creating static Pod manifest for "kube-controller-manager"I0730 00:40:57.176232 26797 manifests.go:91] [control-plane] getting StaticPodSpecsW0730 00:40:57.176346 26797 manifests.go:225] the default kube-apiserver authorization-mode is "Node,RBAC"; using "Node,RBAC"I0730 00:40:57.176680 26797 manifests.go:104] [control-plane] adding volume "ca-certs" for component "kube-controller-manager"I0730 00:40:57.176688 26797 manifests.go:104] [control-plane] adding volume "etc-pki" for component "kube-controller-manager"I0730 00:40:57.176693 26797 manifests.go:104] [control-plane] adding volume "flexvolume-dir" for component "kube-controller-manager"I0730 00:40:57.176697 26797 manifests.go:104] [control-plane] adding volume "k8s-certs" for component "kube-controller-manager"I0730 00:40:57.176702 26797 manifests.go:104] [control-plane] adding volume "kubeconfig" for component "kube-controller-manager"I0730 00:40:57.177947 26797 manifests.go:121] [control-plane] wrote static Pod manifest for component "kube-controller-manager" to "/etc/kubernetes/manifests/kube-controller-manager.yaml"[control-plane] Creating static Pod manifest for "kube-scheduler"I0730 00:40:57.177980 26797 manifests.go:91] [control-plane] getting StaticPodSpecsW0730 00:40:57.178062 26797 manifests.go:225] the default kube-apiserver authorization-mode is "Node,RBAC"; using "Node,RBAC"I0730 00:40:57.178409 26797 manifests.go:104] [control-plane] adding volume "kubeconfig" for component "kube-scheduler"I0730 00:40:57.179219 26797 manifests.go:121] [control-plane] wrote static Pod manifest for component "kube-scheduler" to "/etc/kubernetes/manifests/kube-scheduler.yaml"[etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"I0730 00:40:57.180242 26797 local.go:72] [etcd] wrote Static Pod manifest for a local etcd member to "/etc/kubernetes/manifests/etcd.yaml"I0730 00:40:57.180284 26797 waitcontrolplane.go:87] [wait-control-plane] Waiting for the API server to be healthyI0730 00:40:57.181798 26797 loader.go:375] Config loaded from file: /etc/kubernetes/admin.conf[wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests". This can take up to 4m0sI0730 00:40:57.185824 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:40:57.687883 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:40:58.187689 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:40:58.687505 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:40:59.187708 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:40:59.687437 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:41:00.187753 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:41:00.687830 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:41:01.187695 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:41:01.686916 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:41:02.188601 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:41:02.686665 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:41:03.187222 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:41:03.688352 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:41:04.187550 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 1 millisecondsI0730 00:41:04.691266 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 1 millisecondsI0730 00:41:14.130375 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s 500 Internal Server Error in 8944 millisecondsI0730 00:41:14.202729 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s 500 Internal Server Error in 16 millisecondsI0730 00:41:14.689598 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s 500 Internal Server Error in 3 millisecondsI0730 00:41:15.196163 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s 500 Internal Server Error in 9 millisecondsI0730 00:41:15.689662 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s 500 Internal Server Error in 3 millisecondsI0730 00:41:16.190375 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s 200 OK in 3 milliseconds[apiclient] All control plane components are healthy after 19.005906 secondsI0730 00:41:16.190562 26797 uploadconfig.go:108] [upload-config] Uploading the kubeadm ClusterConfiguration to a ConfigMap[upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" NamespaceI0730 00:41:16.202115 26797 round_trippers.go:443] POST https://192.168.10.128:6443/api/v1/namespaces/kube-system/configmaps?timeout=10s 201 Created in 6 millisecondsI0730 00:41:16.210907 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles?timeout=10s 201 Created in 7 millisecondsI0730 00:41:16.219527 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings?timeout=10s 201 Created in 6 millisecondsI0730 00:41:16.220662 26797 uploadconfig.go:122] [upload-config] Uploading the kubelet component config to a ConfigMap[kubelet] Creating a ConfigMap "kubelet-config-1.18" in namespace kube-system with the configuration for the kubelets in the clusterI0730 00:41:16.227796 26797 round_trippers.go:443] POST https://192.168.10.128:6443/api/v1/namespaces/kube-system/configmaps?timeout=10s 201 Created in 6 millisecondsI0730 00:41:16.234309 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles?timeout=10s 201 Created in 6 millisecondsI0730 00:41:16.241883 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings?timeout=10s 201 Created in 7 millisecondsI0730 00:41:16.242124 26797 uploadconfig.go:127] [upload-config] Preserving the CRISocket information for the control-plane nodeI0730 00:41:16.242159 26797 patchnode.go:30] [patchnode] Uploading the CRI Socket information "/var/run/dockershim.sock" to the Node API object "master" as an annotationI0730 00:41:16.751565 26797 round_trippers.go:443] GET https://192.168.10.128:6443/api/v1/nodes/master?timeout=10s 200 OK in 4 millisecondsI0730 00:41:16.767568 26797 round_trippers.go:443] PATCH https://192.168.10.128:6443/api/v1/nodes/master?timeout=10s 200 OK in 11 milliseconds[upload-certs] Skipping phase. Please see --upload-certs[mark-control-plane] Marking the node master as control-plane by adding the label "node-role.kubernetes.io/master=''"[mark-control-plane] Marking the node master as control-plane by adding the taints [node-role.kubernetes.io/master:NoSchedule]I0730 00:41:17.273970 26797 round_trippers.go:443] GET https://192.168.10.128:6443/api/v1/nodes/master?timeout=10s 200 OK in 4 millisecondsI0730 00:41:17.285850 26797 round_trippers.go:443] PATCH https://192.168.10.128:6443/api/v1/nodes/master?timeout=10s 200 OK in 9 milliseconds[bootstrap-token] Using token: re6mt8.bwol9dej5nsqubxt[bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC RolesI0730 00:41:17.291184 26797 round_trippers.go:443] GET https://192.168.10.128:6443/api/v1/namespaces/kube-system/secrets/bootstrap-token-re6mt8?timeout=10s 404 Not Found in 4 millisecondsI0730 00:41:17.299410 26797 round_trippers.go:443] POST https://192.168.10.128:6443/api/v1/namespaces/kube-system/secrets?timeout=10s 201 Created in 6 milliseconds[bootstrap-token] configured RBAC rules to allow Node Bootstrap tokens to get nodesI0730 00:41:17.309247 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/clusterroles?timeout=10s 201 Created in 8 millisecondsI0730 00:41:17.318135 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings?timeout=10s 201 Created in 7 milliseconds[bootstrap-token] configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentialsI0730 00:41:17.325642 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings?timeout=10s 201 Created in 6 milliseconds[bootstrap-token] configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap TokenI0730 00:41:17.332197 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings?timeout=10s 201 Created in 6 milliseconds[bootstrap-token] configured RBAC rules to allow certificate rotation for all node client certificates in the clusterI0730 00:41:17.338051 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings?timeout=10s 201 Created in 5 milliseconds[bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespaceI0730 00:41:17.338274 26797 clusterinfo.go:45] [bootstrap-token] loading admin kubeconfigI0730 00:41:17.339582 26797 loader.go:375] Config loaded from file: /etc/kubernetes/admin.confI0730 00:41:17.339606 26797 clusterinfo.go:53] [bootstrap-token] copying the cluster from admin.conf to the bootstrap kubeconfigI0730 00:41:17.340061 26797 clusterinfo.go:65] [bootstrap-token] creating/updating ConfigMap in kube-public namespaceI0730 00:41:17.346320 26797 round_trippers.go:443] POST https://192.168.10.128:6443/api/v1/namespaces/kube-public/configmaps?timeout=10s 201 Created in 6 millisecondsI0730 00:41:17.346598 26797 clusterinfo.go:79] creating the RBAC rules for exposing the cluster-info ConfigMap in the kube-public namespaceI0730 00:41:17.354296 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-public/roles?timeout=10s 201 Created in 7 millisecondsI0730 00:41:17.360015 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-public/rolebindings?timeout=10s 201 Created in 5 millisecondsI0730 00:41:17.361663 26797 kubeletfinalize.go:88] [kubelet-finalize] Assuming that kubelet client certificate rotation is enabled: found "/var/lib/kubelet/pki/kubelet-client-current.pem"[kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and keyI0730 00:41:17.362911 26797 loader.go:375] Config loaded from file: /etc/kubernetes/kubelet.confI0730 00:41:17.364376 26797 kubeletfinalize.go:132] [kubelet-finalize] Restarting the kubelet to enable client certificate rotationI0730 00:41:17.680009 26797 round_trippers.go:443] GET https://192.168.10.128:6443/apis/apps/v1/namespaces/kube-system/deployments?labelSelector=k8s-app%3Dkube-dns 200 OK in 11 millisecondsI0730 00:41:17.694805 26797 round_trippers.go:443] GET https://192.168.10.128:6443/api/v1/namespaces/kube-system/configmaps/kube-dns?timeout=10s 404 Not Found in 5 millisecondsI0730 00:41:17.700146 26797 round_trippers.go:443] GET https://192.168.10.128:6443/api/v1/namespaces/kube-system/configmaps/coredns?timeout=10s 404 Not Found in 4 millisecondsI0730 00:41:17.709319 26797 round_trippers.go:443] POST https://192.168.10.128:6443/api/v1/namespaces/kube-system/configmaps?timeout=10s 201 Created in 8 millisecondsI0730 00:41:17.716415 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/clusterroles?timeout=10s 201 Created in 6 millisecondsI0730 00:41:17.722154 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings?timeout=10s 201 Created in 5 millisecondsI0730 00:41:17.730289 26797 round_trippers.go:443] POST https://192.168.10.128:6443/api/v1/namespaces/kube-system/serviceaccounts?timeout=10s 201 Created in 6 millisecondsI0730 00:41:17.776066 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/apps/v1/namespaces/kube-system/deployments?timeout=10s 201 Created in 34 millisecondsI0730 00:41:17.790010 26797 round_trippers.go:443] POST https://192.168.10.128:6443/api/v1/namespaces/kube-system/services?timeout=10s 201 Created in 11 milliseconds[addons] Applied essential addon: CoreDNSI0730 00:41:17.798512 26797 round_trippers.go:443] POST https://192.168.10.128:6443/api/v1/namespaces/kube-system/serviceaccounts?timeout=10s 201 Created in 8 millisecondsI0730 00:41:17.809712 26797 round_trippers.go:443] POST https://192.168.10.128:6443/api/v1/namespaces/kube-system/configmaps?timeout=10s 201 Created in 8 millisecondsI0730 00:41:17.861931 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/apps/v1/namespaces/kube-system/daemonsets?timeout=10s 201 Created in 35 millisecondsI0730 00:41:17.872631 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings?timeout=10s 201 Created in 9 millisecondsI0730 00:41:17.878624 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles?timeout=10s 201 Created in 5 millisecondsI0730 00:41:17.884579 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings?timeout=10s 201 Created in 5 milliseconds[addons] Applied essential addon: kube-proxyI0730 00:41:17.885611 26797 loader.go:375] Config loaded from file: /etc/kubernetes/admin.confI0730 00:41:17.886868 26797 loader.go:375] Config loaded from file: /etc/kubernetes/admin.confYour Kubernetes control-plane has initialized successfully!To start using your cluster, you need to run the following as a regular user:mkdir -p $HOME/.kubesudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/configsudo chown $(id -u):$(id -g) $HOME/.kube/configYou should now deploy a pod network to the cluster.Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:https://kubernetes.io/docs/concepts/cluster-administration/addons/Then you can join any number of worker nodes by running the following on each as root:kubeadm join 192.168.10.128:6443 --token re6mt8.bwol9dej5nsqubxt \--discovery-token-ca-cert-hash sha256:21ec2e66f355b7d31652b2b790e4862ef2fb9b6069d7425d122007a7d2edce782.10 安装pod网络插件kubectl apply -f https://raw.githubusercontent.com/coreos/flannel/master/Documentation/kube-flannel.yaml" with one of the options listed at:https://kubernetes.io/docs/concepts/cluster-administration/addons/Then you can join any number of worker nodesbyttps://www.ipaddress.com进行查看,将域名和IP对应关系写到hosts文件中2.11 node节点加入集群kubeadm join 192.168.10.128:6443 --token whxbm2.x70y9k1feop2604b \--discovery-token-ca-cert-hash sha256:d4306836b7ef32fd802e559cf717bbbe3af6dd75d3cb2fa4d3db7a16cc25a6573异常节点处理:如需要删除节点的话,则需要执行如下操作主节点操作# 在主节点上驱逐该node节点上的Podkubectl drain node02 --delete-local-data --force --ignore-daemonsets# 在主节点上删除该node节点kubectl delete node node02从节点上操作# 重置k8skubeadm reset# 删除残留的文件rm -rf /etc/kubernetes/*# 清除iptables或者ipvs的配置iptables -F && iptables -t nat -F && iptables -t mangle -F && iptables -Xipvsadm --clear如果是master节点 还需要注意rm -rf ~/.kube《(网王)纯属意外》紫陌寒冲晋江文学城冲【衍生小说触...

一改踏板车型光滑平整的外壳造型引入了越野车型的设计语言给人带来更全新的体验

发布于:休宁县
声明:该文观点仅代表作者本人,搜狐号系信息发布平台,搜狐仅提供信息存储空间服务。
意见反馈 合作

Copyright ? 2023 Sohu All Rights Reserved

搜狐公司 版权所有