标签:比较 int systemctl inux shang 工作 clust x86 mamicode
[kube-proxy]http://www.cnblogs.com/xuxinkun/p/5799986.html
[flannel]
[root@master ~]# cd ~/k8s
[root@master ~]# wget https://raw.githubusercontent.com/coreos/flannel/master/Documentation/kube-flannel.yml
[root@master ~]# kubectl apply -f kube-flannel.yml
clusterrole "flannel" created
clusterrolebinding "flannel" created
serviceaccount "flannel" created
configmap "kube-flannel-cfg" created
daemonset "kube-flannel-ds" created
- 指定网卡
如果有多个网卡,需要在kube-flannel.yml中使用–iface参数指定集群主机内网网卡的名称,否则可能会出现dns无法解析。需要将kube-flannel.yml下载到本地,flanneld启动参数加上–iface=。
......
apiVersion: extensions/v1beta1
kind: DaemonSet
metadata:
name: kube-flannel-ds
......
containers:
- name: kube-flannel
image: quay.io/coreos/flannel:v0.9.0-amd64
command: [ "/opt/bin/flanneld", "--ip-masq", "--kube-subnet-mgr", "--iface=eth1" ]
k8s 官方文档:
- kubectl 大家应该都知道,是跟 k8s 服务交互的命令行工具。
- kubeadm 就是安装 k8s 测试环境的命令行工具
- kubelet 就比较重要了,没有 kubelet,kubeadm 啥也干不了。kubelet 其实就类似于 Nova 中的
nova-compute 进程(管理 VM),负责管理 container。安装完 kubelet,系统中就多了一个
kubelet 服务。关于 kubelet,
kubeadm init 干了什么:
- 系统状态检查
- 生成 token
- 生成自签名证书
- 生成 kubeconfig 用于跟 api talk
- 为管理面服务容器生成 manifest,放在
/etc/kubernetes/manifests
目录下
- 配置 RBAC,并设置master node只运行管理容器
- 创建附加服务,比如 kube-proxy 和 kube-dns 等
安装成功后就可以查看系统中创建的 container 和 pod:
docker ps | grep -v ‘/pause‘
kubectl get pods --all-namespaces
熟悉 k8s 的朋友其实已经知道那些处于 pause 的 container 的作用。当你在 k8s
中创建包含一个 container 的 pod 时,其实 k8s 会在这个 pod 里偷偷创建一个叫 infra-container
的容器,初始化 pod 的网络、命名空间,pod 中的其他 container 就会共享这个网络和命名空间。所以完成网络初始化后,这些
infra-container 就会永久睡眠,直到收到 SIGINT 或 SIGTERM 信号。
这里看到 kube-dns 卡在 Pending 状态,是因为它必须在安装 pod network 组件后才能启动成功。
Revert k8s master and nodes
当我再往下准备安装网络组件时,发现 calico 要求执行 kubeadm 时有额外参数,所以我就回退了 kubeadm 的安装:
kubectl get nodes
kubectl drain lingxian-XXXX-kubeadm --delete-local-data --force --ignore-daemonsets
kubectl delete node lingxian-XXXX-kubeadm
kubeadm reset
kubeadm init --pod-network-cidr=192.168.0.0/16
kubeadm reset
ifconfig cni0 down
ip link delete cni0
ifconfig flannel.1 down
ip link delete flannel.1
rm -rf /var/lib/cni/
[Kubeadm reset]
find /var/lib/kubelet | xargs -n 1 findmnt -n -t tmpfs -o TARGET -T | uniq | xargs -r umount -v;
rm -r -f /etc/kubernetes /var/lib/kubelet /var/lib/etcd;
kubeadm reset
Allow Pod schedule to Master
我只有一个 node,所以就把 master 直接当 worker 用:
root@lingxian-test-kubeadm:~# kubectl taint nodes --all node-role.kubernetes.io/master-
node "lingxian-test-kubeadm" untainted
Master节点参与工作负载
使用kubeadm初始化的集群,出于安全考虑Pod不会被调度到Master Node上,可使用如下命令使Master节点参与工作负载。
kubectl taint nodes node1 node-role.kubernetes.io/master-
Run
(XX)kubectl run mynginx --image=nginx --expose --port 8088
kubectl delete deployment mynginx
kubectl delete svc mynginx
kubectl run unginx --image=nginx --expose --port 80
kubectl run centos --image=cu.eshore.cn/library/java:jdk8 --command -- vi
kubectl scale --replicas=4 deployment/centos
向Kubernetes集群添加Node
kubeadm token list | grep authentication,signing | awk ‘{print $1}‘
- 查看discovery-token-ca-cert-hash
openssl x509 -pubkey -in /etc/kubernetes/pki/ca.crt | openssl rsa -pubin -outform der 2>/dev/null | openssl dgst -sha256 -hex | sed ‘s/^.* //‘
kubeadm join --token=a20844.654ef6410d60d465 --discovery-token-ca-cert-hash sha256:0c2dbe69a2721870a59171c6b5158bd1c04bc27665535ebf295c918a96de0bb1 master.k8s.samwong.im:6443
token失效被删除。在Master上查看token,结果为空。
kubeadm token list
- 解决方法
重新生成token,默认token有效期为24小时,生成token时通过指定--ttl 0可设置token永久有效。
[root@master ~]# kubeadm token create --ttl 0
https://lingxiankong.github.io/2018-01-20-install-k8s.html[step by step]
时间同步,主机名,/etc/hosts,防火墙,selinux, 无密钥登录,安装docker-1.12.6 这些都已经配置好了的。
https://anthonychu.ca/post/api-versioning-kubernetes-nginx-ingress/
[dashboard]
https://www.zybuluo.com/ncepuwanghui/note/953929
授予Dashboard账户集群管理权限
创建一个kubernetes-dashboard-admin的ServiceAccount并授予集群admin的权限,创建kubernetes-dashboard-admin.rbac.yaml。
kubectl create -f kubernetes-dashboard-admin.rbac.yaml
- 查看kubernete-dashboard-admin的token
kubectl -n kube-system get secret | grep kubernetes-dashboard-admin
kubectl describe -n kube-system secret/xxxxxxx
- 查看Dashboard服务端口
kubectl get svc -n kube-system
[heapster]
[heapster]http://www.mamicode.com/info-detail-1715935.html
https://www.slahser.com/2016/11/18/k8s%E5%90%8E%E6%97%A5%E8%B0%88-Heaspter%E7%9B%91%E6%8E%A7/
安装Heapster为集群添加使用统计和监控功能,为Dashboard添加仪表盘。
mkdir -p ~/k8s/heapster
cd ~/k8s/heapster
wget https://raw.githubusercontent.com/kubernetes/heapster/master/deploy/kube-config/influxdb/grafana.yaml
wget https://raw.githubusercontent.com/kubernetes/heapster/master/deploy/kube-config/rbac/heapster-rbac.yaml
wget https://raw.githubusercontent.com/kubernetes/heapster/master/deploy/kube-config/influxdb/heapster.yaml
wget https://raw.githubusercontent.com/kubernetes/heapster/master/deploy/kube-config/influxdb/influxdb.yaml
kubectl create -f ./
docker pull mirrorgooglecontainers/heapster-grafana-amd64:v4.4.3
docker tag mirrorgooglecontainers/heapster-grafana-amd64:v4.4.3
k8s.gcr.io/heapster-grafana-amd64:v4.4.3
docker pull mirrorgooglecontainers/heapster-amd64:v1.4.2
docker tag mirrorgooglecontainers/heapster-amd64:v1.4.2
k8s.gcr.io/heapster-amd64:v1.4.2
docker pull mirrorgooglecontainers/heapster-influxdb-amd64:v1.3.3
docker tag mirrorgooglecontainers/heapster-influxdb-amd64:v1.3.3
k8s.gcr.io/heapster-influxdb-amd64:v1.3.3
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++=
docker pull ist0ne/kubernetes-dashboard-amd64:v1.8.0
docker tag ist0ne/kubernetes-dashboard-amd64:v1.8.0
k8s.gcr.io/kubernetes-dashboard-amd64:v1.8.3
https://kubernetes.io/docs/setup/independent/create-cluster-kubeadm/
v1.8.4
kubectl replace --force -f
sysctl -P 报错解决办法
问题症状
修改 linux 内核文件
#vi /etc/sysctl.conf后执行sysctl -P 报错
error: "net.bridge.bridge-nf-call-ip6tables" is an unknown key
error: "net.bridge.bridge-nf-call-iptables" is an unknown key
error: "net.bridge.bridge-nf-call-arptables" is an unknown key
解决方法如下:
modprobe bridge
lsmod|grep bridge
Environment="KUBELET_NETWORK_ARGS=--network-plugin=cni
--cni-conf-dir=/etc/cni/net.d --cni-bin-dir=/opt/cni/bin"
https://www.zybuluo.com/ncepuwanghui/note/953929
[kexueshangwang--]
docker tag
registry.cn-hangzhou.aliyuncs.com/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.14.5
gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.14.5
刚刚查看了一下,二者确实不一样。docker的cgroup-driver是cgroupfs。将此conf文件对应修改成cgroupfs,正在重新运行,检查kubelet运行正常。但又卡在
跳过(依赖问题):
docker-engine.x86_64
0:1.12.6-1.el7.centos
docker-engine-selinux.noarch
0:17.05.0.ce-1.el7.centos
libtool-ltdl.x86_64 0:2.4.2-22.el7_3
images=(
kube-apiserver-amd64:v1.8.4 kube-controller-manager-amd64:v1.8.4 kube-scheduler-amd64:v1.8.4 kube-proxy-amd64:v1.8.4 etcd-amd64:3.0.17 pause-amd64:3.0 k8s-dns-sidecar-amd64:1.14.5 k8s-dns-kube-dns-amd64:1.14.5 k8s-dns-dnsmasq-nanny-amd64:1.14.5 flannel:v0.9.1-amd64
)
for imageName in ${images[@]} ; do
docker pull mritd/$imageName
docker tag mritd/$imageName gcr.io/google_containers/$imageName
docker rmi mritd/$imageName
done
问题2
新打开一个窗口,查看 /var/log/messages 有如下错误:
1
|
Aug 12 23:40:10 cu3 kubelet: error: failed to run Kubelet: failed to create kubelet: misconfiguration: kubelet cgroup driver: "systemd" is different from docker cgroup driver: "cgroupfs"
|
docker和kubelet的cgroup driver不一样,修改kubelet的配置。把docker启动参数 masq 一起改了。
1
2
3
4
|
[root@cu3 ~]# sed -i ‘s/KUBELET_CGROUP_ARGS=--cgroup-driver=systemd/KUBELET_CGROUP_ARGS=--cgroup-driver=cgroupfs/‘ /etc/systemd/system/kubelet.service.d/10-kubeadm.conf
[root@cu3 ~]# sed -i ‘s#/usr/bin/dockerd.*#/usr/bin/dockerd --ip-masq=false#‘ /usr/lib/systemd/system/docker.service
[root@cu3 ~]# systemctl daemon-reload; systemctl restart docker kubelet
|
sed -i ‘s#/usr/bin/dockerd.*#/usr/bin/dockerd --ip-masq=false#‘ /usr/lib/systemd/system/docker.service
注意:加了 ip-masq=false 后,docker0就不能上外网了。也就是单独起的docker容器不能上外网!
1
|
ExecStart=/usr/bin/dockerd --ip-masq=false
|
节点防火墙(由于是云主机,增加防火墙):
1
2
3
|
firewall-cmd --zone=trusted --add-source=192.168.0.0/16 --permanent
firewall-cmd --zone=trusted --add-source=10.0.0.0/8 --permanent
firewall-cmd --complete-reload
|
[download k8s]
https://github.com/kubernetes/kubernetes/blob/master/CHANGELOG-1.8.md#v181
[k8s metrics]
http://www.cnblogs.com/iiiiher/p/7999761.html
[k8s cookbook]
https://github.com/kubernetes/kubernetes/tree/master/examples/guestbook
[infra esta]
http://blog.csdn.net/qq_32971807/article/details/54693254
[ranchor]
https://www.cnrancher.com/rancher-k8s-accelerate-installation-document/
docker run -d
--restart always --name rancher_server -p 6080:8080
rancher/server && docker logs -f rancher-server
[k8s 1.6.1 install]
https://www.jianshu.com/p/8ce11f947410
[CNI]
https://segmentfault.com/a/1190000008803805
http://www.cnblogs.com/whtydn/p/4353695.html
(+++)https://www.jianshu.com/p/a2039a8855ec
(!!!!!)https://www.cnblogs.com/liangDream/p/7358847.html
http://ju.outofmemory.cn/entry/231591
https://segmentfault.com/a/1190000007074726
http://www.infoq.com/cn/articles/centos7-practical-kubernetes-deployment
http://www.cnblogs.com/zhenyuyaodidiao/p/6500720.html
(kernel concept)
https://www.jianshu.com/p/8d3204b96cf9
http://blog.csdn.net/hackstoic/article/details/50574886
(mesos)
kubenets installation--ranchor-mesos
标签:比较 int systemctl inux shang 工作 clust x86 mamicode
原文地址:https://www.cnblogs.com/SZLLQ2000/p/9771080.html