Dear Fabien and Fabio, Puppet configuration is now fixed and Kubernetes 1.9.1 is running on all nodes. However I now use internal kubernetes DNS and it seems something it killing it in my back after a few seconds on CC-IN2P3 cluster (see logs below). FYI, I have successfully tested this setup on 10 Centos7 boxes in Openstack. Would you please have an idea about the thing which kill kube-dns process on CC-IN2P3 cluster? Cheers, Fabrice # Logs for the process inside kubedns pod: root@ccosvms0070:~/admin# kubectl logs kube-dns-6f4fd4bdf-rb4x8 -c kubedns --namespace kube-system I0201 13:29:23.613716 1 dns.go:48] version: 1.14.6-3-gc36cb11 I0201 13:29:23.615050 1 server.go:69] Using configuration read from directory: /kube-dns-config with period 10s I0201 13:29:23.615128 1 server.go:112] FLAG: --alsologtostderr="false" I0201 13:29:23.615144 1 server.go:112] FLAG: --config-dir="/kube-dns-config" I0201 13:29:23.615153 1 server.go:112] FLAG: --config-map="" I0201 13:29:23.615158 1 server.go:112] FLAG: --config-map-namespace="kube-system" I0201 13:29:23.615165 1 server.go:112] FLAG: --config-period="10s" I0201 13:29:23.615173 1 server.go:112] FLAG: --dns-bind-address="0.0.0.0" I0201 13:29:23.615180 1 server.go:112] FLAG: --dns-port="10053" I0201 13:29:23.615189 1 server.go:112] FLAG: --domain="cluster.local." I0201 13:29:23.615198 1 server.go:112] FLAG: --federations="" I0201 13:29:23.615204 1 server.go:112] FLAG: --healthz-port="8081" I0201 13:29:23.615209 1 server.go:112] FLAG: --initial-sync-timeout="1m0s" I0201 13:29:23.615213 1 server.go:112] FLAG: --kube-master-url="" I0201 13:29:23.615221 1 server.go:112] FLAG: --kubecfg-file="" I0201 13:29:23.615227 1 server.go:112] FLAG: --log-backtrace-at=":0" I0201 13:29:23.615237 1 server.go:112] FLAG: --log-dir="" I0201 13:29:23.615244 1 server.go:112] FLAG: --log-flush-frequency="5s" I0201 13:29:23.615251 1 server.go:112] FLAG: --logtostderr="true" I0201 13:29:23.615256 1 server.go:112] FLAG: --nameservers="" I0201 13:29:23.615263 1 server.go:112] FLAG: --stderrthreshold="2" I0201 13:29:23.615268 1 server.go:112] FLAG: --v="2" I0201 13:29:23.615274 1 server.go:112] FLAG: --version="false" I0201 13:29:23.615283 1 server.go:112] FLAG: --vmodule="" I0201 13:29:23.615368 1 server.go:194] Starting SkyDNS server (0.0.0.0:10053) I0201 13:29:23.615753 1 server.go:213] Skydns metrics enabled (/metrics:10055) I0201 13:29:23.615781 1 dns.go:146] Starting endpointsController I0201 13:29:23.615791 1 dns.go:149] Starting serviceController I0201 13:29:23.615880 1 logs.go:41] skydns: ready for queries on cluster.local. for tcp://0.0.0.0:10053 [rcache 0] I0201 13:29:23.615898 1 logs.go:41] skydns: ready for queries on cluster.local. for udp://0.0.0.0:10053 [rcache 0] I0201 13:29:24.116020 1 dns.go:170] Initialized services and endpoints from apiserver I0201 13:29:24.116075 1 server.go:128] Setting up Healthz Handler (/readiness) I0201 13:29:24.116108 1 server.go:133] Setting up cache handler (/cache) I0201 13:29:24.116121 1 server.go:119] Status HTTP port 8081 I0201 13:31:02.019718 1 server.go:153] Ignoring signal terminated (can only be terminated by SIGKILL) # kube-dns pods has crashed 39 times in 1 hour root@ccosvms0070:~/admin# kubectl get pods --namespace kube-system NAME READY STATUS RESTARTS AGE etcd-ccqservkm1 1/1 Running 0 1h kube-apiserver-ccqservkm1 1/1 Running 0 1h kube-controller-manager-ccqservkm1 1/1 Running 0 1h kube-dns-6f4fd4bdf-rb4x8 0/3 CrashLoopBackOff 39 1h kube-proxy-28ssp 1/1 Running 0 1h kube-proxy-4kb9z 1/1 Running 0 1h kube-proxy-67vk2 1/1 Running 0 1h ######################################################################## Use REPLY-ALL to reply to list To unsubscribe from the QSERV-L list, click the following link: https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=QSERV-L&A=1